Alternatives to Sarvam 105B
Compare Sarvam 105B alternatives for your business or organization using the curated list below. SourceForge ranks the best alternatives to Sarvam 105B in 2026. Compare features, ratings, user reviews, pricing, and more from Sarvam 105B competitors and alternatives in order to make an informed decision for your business.
-
1
Sarvam-M
Sarvam
Sarvam-M is a multilingual, hybrid-reasoning large language model designed to deliver strong performance across Indian languages, mathematical reasoning, and programming tasks within a single, efficient system. Built on top of Mistral-Small, it is a 24-billion-parameter text-only model that has been enhanced through supervised fine-tuning, reinforcement learning with verifiable rewards, and inference optimizations to improve both accuracy and efficiency. The model is specifically trained to handle more than ten major Indic languages, supporting native scripts, romanized text, and code-mixed inputs, enabling seamless multilingual communication across diverse linguistic contexts. Sarvam-M introduces a hybrid reasoning approach that allows it to switch between “thinking” mode for complex tasks like math, logic, and coding, and faster response mode for everyday interactions, balancing performance and speed. -
2
Sarvam 30B
Sarvam
Sarvam-30B is an open source, next-generation large language model designed as a unified system for both real-time conversational AI and deep reasoning workloads, built with a strong focus on multilingual intelligence and practical deployment. The 30B model is optimized for speed and efficiency, using a Mixture-of-Experts (MoE) architecture that activates only a subset of parameters per request, enabling high throughput, low latency, and deployment even in resource-constrained environments such as local machines or edge systems. It delivers strong performance in conversational tasks, coding, and reasoning while achieving state-of-the-art results across more than 20 Indian languages, making it highly effective for multilingual applications and voice-based systems. It represents a dual-tier architecture, a fast, deployable “conversational workhorse”, leveraging MoE designs to reduce compute cost while maintaining high performance.Starting Price: Free -
3
Sarvam Indus
Sarvam
Indus is Sarvam’s official conversational AI interface designed to give users direct access to its flagship sovereign language models through a simple, real-time chat experience. Introduced in February 2026 as a limited beta product, it serves as the primary interface for interacting with Sarvam’s 105-billion-parameter model, bringing advanced reasoning, multilingual understanding, and conversational capabilities into a single application. It is built to deliver an AI experience tailored specifically to Indian users, supporting more than 22 Indian languages, including native scripts and code-mixed inputs, while maintaining contextual understanding aligned with local culture and communication patterns. It enables both text and voice interactions, allowing users to speak naturally and receive responses in text or synthesized speech, creating a voice-first, accessible interface for diverse use cases. -
4
Nemotron 3 Ultra
NVIDIA
Nemotron 3 Nano is a compact, open large language model in NVIDIA’s Nemotron 3 family, designed for efficient agentic reasoning, conversational AI, and coding tasks. It uses a hybrid Mixture-of-Experts Mamba-Transformer architecture that activates only a small subset of parameters per token, enabling low-latency inference while maintaining strong accuracy and reasoning performance. It has approximately 31.6 billion total parameters with around 3.2 billion active (3.6 billion including embeddings), allowing it to achieve higher accuracy than previous Nemotron 2 Nano while using less computation per forward pass. Nemotron 3 Nano supports long-context processing of up to one million tokens, enabling it to handle large documents, multi-step workflows, and extended reasoning chains in a single pass. It is designed for high-throughput, real-time execution, excelling in multi-turn conversations, tool calling, and agent-based workflows where tasks require planning, reasoning, and more. -
5
Sarvam AI
Sarvam AI
Sarvam AI is a sovereign AI platform designed to build and deploy artificial intelligence solutions tailored for India. It offers a full-stack ecosystem that includes advanced models, infrastructure, and tools for enterprise, government, and developer use. The platform is built on sovereign compute, ensuring data control and compliance within India. Sarvam AI provides state-of-the-art models optimized for Indian languages, culture, and real-world use cases. It supports applications such as conversational agents, speech-to-text, text-to-speech, and vision-based solutions. The platform also includes scalable infrastructure that simplifies AI deployment and model serving. With flexible deployment options like cloud, private cloud, and on-premises environments, it adapts to various business needs. Overall, Sarvam AI enables organizations to build AI-driven solutions with greater control, localization, and scalability. -
6
Qwen3.6-35B-A3B
Alibaba
Qwen3.5-35B-A3B is part of the Qwen3.5 “Medium” model series, designed as a highly efficient, multimodal foundation model that balances strong reasoning ability with practical deployment requirements. It uses a Mixture-of-Experts (MoE) architecture with 35 billion total parameters but activates only about 3 billion per token, allowing it to deliver performance comparable to much larger models while significantly reducing computational cost. The model integrates a hybrid attention mechanism that combines linear attention with standard attention layers, enabling efficient long-context processing and improved scalability for complex tasks. As a native vision-language model, it can process both text and visual inputs, supporting use cases such as multimodal reasoning, coding, and agent-based workflows. It is designed to function as a general-purpose “AI agent,” capable of planning, tool use, and structured problem solving rather than just conversational responses.Starting Price: Free -
7
Nemotron 3 Super
NVIDIA
Nemotron-3 Super is part of NVIDIA’s Nemotron 3 family of open models designed to enable advanced agentic AI systems that can reason, plan, and execute multi-step workflows across complex environments. The model introduces a hybrid Mamba-Transformer Mixture-of-Experts architecture that combines the efficiency of state-space Mamba layers with the contextual understanding of transformer attention, allowing it to process long sequences and complex reasoning tasks with high accuracy and throughput. This architecture activates only a subset of model parameters for each token, improving computational efficiency while maintaining strong reasoning capabilities and enabling scalable inference for large workloads. Nemotron-3 Super contains roughly 120 billion parameters with around 12 billion active during inference, accelerating multi-step reasoning and collaborative agent interactions across large contexts. -
8
Qwen3.5
Alibaba
Qwen3.5 is a next-generation open-weight multimodal large language model designed to power native vision-language agents. The flagship release, Qwen3.5-397B-A17B, combines a hybrid linear attention architecture with sparse mixture-of-experts, activating only 17 billion parameters per forward pass out of 397 billion total to maximize efficiency. It delivers strong benchmark performance across reasoning, coding, multilingual understanding, visual reasoning, and agent-based tasks. The model expands language support from 119 to 201 languages and dialects while introducing a 1M-token context window in its hosted version, Qwen3.5-Plus. Built for multimodal tasks, it processes text, images, and video with advanced spatial reasoning and tool integration. Qwen3.5 also incorporates scalable reinforcement learning environments to improve general agent capabilities. Designed for developers and enterprises, it enables efficient, tool-augmented, multimodal AI workflows.Starting Price: Free -
9
DeepSeek-V4-Pro
DeepSeek
DeepSeek-V4-Pro is a large-scale Mixture-of-Experts (MoE) language model designed for advanced reasoning, coding, and long-context understanding. It features 1.6 trillion total parameters with 49 billion activated parameters, enabling high performance while maintaining efficiency. The model supports an exceptionally large context window of up to one million tokens, allowing it to process extensive documents and workflows. It uses a hybrid attention architecture to optimize long-context performance and reduce computational cost. DeepSeek-V4-Pro is trained on over 32 trillion tokens, improving its knowledge and reasoning capabilities. It also includes advanced optimization techniques for stability and faster convergence during training. The model supports multiple reasoning modes, allowing users to balance speed and accuracy based on their needs. Overall, it provides a powerful open-source solution for complex AI tasks and large-scale applications.Starting Price: Free -
10
DeepSeek-V4-Flash
DeepSeek
DeepSeek-V4-Flash is a high-efficiency Mixture-of-Experts (MoE) language model designed for fast, scalable reasoning and text generation. It features 284 billion total parameters with 13 billion activated parameters, delivering strong performance while optimizing computational cost. The model supports an extensive context window of up to one million tokens, enabling it to process large documents and complex workflows with ease. Its hybrid attention architecture enhances long-context efficiency by reducing memory and compute requirements. Trained on over 32 trillion tokens, DeepSeek-V4-Flash demonstrates solid capabilities across knowledge, reasoning, and coding tasks. It is designed for scenarios where speed and efficiency are critical, offering a balance between performance and resource usage. The model also supports multiple reasoning modes, allowing users to adjust between faster outputs and deeper analysis.Starting Price: Free -
11
Sarvam Samvaad
Sarvam
Sarvam Conversational Agents (Sarvam Samvaad) is an enterprise-grade conversational AI platform designed to help organizations build, deploy, and scale intelligent, human-like agents across multiple communication channels. It enables businesses to run voice calls, WhatsApp conversations, in-app chat, and web interactions from a single unified system, maintaining the same agent context and memory regardless of where the interaction occurs. It integrates deeply with enterprise systems such as CRM, core banking, and payment infrastructure, allowing agents to pull real-time customer data, execute workflows, and push outcomes back into business systems automatically. It supports multilingual communication, particularly across Indian languages, enabling agents to understand complex phrases, colloquial speech, alphanumeric inputs, and proper nouns with high accuracy. It is built for production environments, allowing enterprises to quickly move from pilot to full deployment. -
12
Kimi K2 Thinking
Moonshot AI
Kimi K2 Thinking is an advanced open source reasoning model developed by Moonshot AI, designed specifically for long-horizon, multi-step workflows where the system interleaves chain-of-thought processes with tool invocation across hundreds of sequential tasks. The model uses a mixture-of-experts architecture with a total of 1 trillion parameters, yet only about 32 billion parameters are activated per inference pass, optimizing efficiency while maintaining vast capacity. It supports a context window of up to 256,000 tokens, enabling the handling of extremely long inputs and reasoning chains without losing coherence. Native INT4 quantization is built in, which reduces inference latency and memory usage without performance degradation. Kimi K2 Thinking is explicitly built for agentic workflows; it can autonomously call external tools, manage sequential logic steps (up to and typically between 200-300 tool calls in a single chain), and maintain consistent reasoning.Starting Price: Free -
13
DeepSeek-V4
DeepSeek
DeepSeek-V4 is a next-generation open-source language model designed for high-performance reasoning, coding, and long-context intelligence. It introduces a powerful architecture with up to one million token context length, enabling seamless handling of large datasets and complex multi-step workflows. The model comes in two variants: DeepSeek-V4-Pro for maximum performance and DeepSeek-V4-Flash for efficiency and speed. DeepSeek-V4-Pro features 1.6 trillion total parameters with 49 billion activated, delivering near state-of-the-art performance comparable to leading closed-source models. It excels in agentic coding, mathematical reasoning, and world knowledge tasks. The model integrates advanced attention mechanisms, including token-wise compression and sparse attention, significantly reducing compute and memory costs. It is also optimized for AI agents, supporting tool use and multi-step workflows.Starting Price: Free -
14
MiMo-V2-Flash
Xiaomi Technology
MiMo-V2-Flash is an open weight large language model developed by Xiaomi based on a Mixture-of-Experts (MoE) architecture that blends high performance with inference efficiency. It has 309 billion total parameters but activates only 15 billion active parameters per inference, letting it balance reasoning quality and computational efficiency while supporting extremely long context handling, for tasks like long-document understanding, code generation, and multi-step agent workflows. It incorporates a hybrid attention mechanism that interleaves sliding-window and global attention layers to reduce memory usage and maintain long-range comprehension, and it uses a Multi-Token Prediction (MTP) design that accelerates inference by processing batches of tokens in parallel. MiMo-V2-Flash delivers very fast generation speeds (up to ~150 tokens/second) and is optimized for agentic applications requiring sustained reasoning and multi-turn interactions.Starting Price: Free -
15
Trinity-Large-Thinking
Arcee AI
Trinity Large Thinking is a frontier open source reasoning model developed by Arcee AI, designed specifically for complex, multi-step problem solving and autonomous agent workflows that require long-horizon planning and tool use. Built on a sparse Mixture-of-Experts architecture with roughly 400 billion total parameters but only about 13 billion active per token, the model achieves high efficiency while maintaining strong reasoning performance across tasks such as mathematical problem solving, code generation, and multi-step analysis. It introduces extended chain-of-thought reasoning capabilities, allowing the model to generate intermediate “thinking traces” before producing final answers, which improves accuracy and reliability in complex scenarios. Trinity Large Thinking supports a very large context window of up to 262K tokens, enabling it to process long documents, maintain state across extended interactions, and operate effectively in continuous agent loops.Starting Price: Free -
16
GLM-4.5
Z.ai
GLM‑4.5 is Z.ai’s latest flagship model in the GLM family, engineered with 355 billion total parameters (32 billion active) and a companion GLM‑4.5‑Air variant (106 billion total, 12 billion active) to unify advanced reasoning, coding, and agentic capabilities in one architecture. It operates in a “thinking” mode for complex, multi‑step reasoning and tool use, and a “non‑thinking” mode for instant responses, supporting up to 128 K token context length and native function calling. Available via the Z.ai chat platform and API, with open weights on HuggingFace and ModelScope, GLM‑4.5 ingests diverse inputs to solve general problem‑solving, common‑sense reasoning, coding from scratch or within existing projects, and end‑to‑end agent workflows such as web browsing and slide generation. Built on a Mixture‑of‑Experts design with loss‑free balance routing, grouped‑query attention, and an MTP layer for speculative decoding, it delivers enterprise‑grade performance. -
17
Kimi K2
Moonshot AI
Kimi K2 is a state-of-the-art open source large language model series built on a mixture-of-experts (MoE) architecture, featuring 1 trillion total parameters and 32 billion activated parameters for task-specific efficiency. Trained with the Muon optimizer on over 15.5 trillion tokens and stabilized by MuonClip’s attention-logit clamping, it delivers exceptional performance in frontier knowledge, reasoning, mathematics, coding, and general agentic workflows. Moonshot AI provides two variants, Kimi-K2-Base for research-level fine-tuning and Kimi-K2-Instruct pre-trained for immediate chat and tool-driven interactions, enabling both custom development and drop-in agentic capabilities. Benchmarks show it outperforms leading open source peers and rivals top proprietary models in coding tasks and complex task breakdowns, while its 128 K-token context length, tool-calling API compatibility, and support for industry-standard inference engines.Starting Price: Free -
18
Qwen2
Alibaba
Qwen2 is the large language model series developed by Qwen team, Alibaba Cloud. Qwen2 is a series of large language models developed by the Qwen team at Alibaba Cloud. It includes both base language models and instruction-tuned models, ranging from 0.5 billion to 72 billion parameters, and features both dense models and a Mixture-of-Experts model. The Qwen2 series is designed to surpass most previous open-weight models, including its predecessor Qwen1.5, and to compete with proprietary models across a broad spectrum of benchmarks in language understanding, generation, multilingual capabilities, coding, mathematics, and reasoning.Starting Price: Free -
19
GigaChat 3 Ultra
Sberbank
GigaChat 3 Ultra is a 702-billion-parameter Mixture-of-Experts model built from scratch to deliver frontier-level reasoning, multilingual capability, and deep Russian-language fluency. It activates just 36 billion parameters per token, enabling massive scale with practical inference speeds. The model was trained on a 14-trillion-token corpus combining natural, multilingual, and high-quality synthetic data to strengthen reasoning, math, coding, and linguistic performance. Unlike modified foreign checkpoints, GigaChat 3 Ultra is entirely original—giving developers full control, modern alignment, and a dataset free of inherited limitations. Its architecture leverages MoE, MTP, and MLA to match open-source ecosystems and integrate easily with popular inference and fine-tuning tools. With leading results on Russian benchmarks and competitive performance on global tasks, GigaChat 3 Ultra represents one of the largest and most capable open-source LLMs in the world.Starting Price: Free -
20
MiMo-V2.5
Xiaomi Technology
Xiaomi MiMo-V2.5 is an advanced open-source AI model designed to combine strong agentic capabilities with native multimodal understanding. It can process and reason across text, images, and audio within a single unified system. The model uses a sparse Mixture-of-Experts architecture with hundreds of billions of parameters for efficient performance. It supports an extended context window of up to one million tokens, enabling long and complex workflows. MiMo-V2.5 is built to handle tasks such as coding, reasoning, and multimodal analysis with high accuracy. It incorporates dedicated visual and audio encoders to enhance perception and cross-modal reasoning. The model demonstrates strong benchmark performance across coding, reasoning, and multimodal tasks. By combining multimodality, efficiency, and agentic intelligence, MiMo-V2.5 advances the capabilities of open-source AI systems. -
21
Qwen3-Max
Alibaba
Qwen3-Max is Alibaba’s latest trillion-parameter large language model, designed to push performance in agentic tasks, coding, reasoning, and long-context processing. It is built atop the Qwen3 family and benefits from the architectural, training, and inference advances introduced there; mixing thinker and non-thinker modes, a “thinking budget” mechanism, and support for dynamic mode switching based on complexity. The model reportedly processes extremely long inputs (hundreds of thousands of tokens), supports tool invocation, and exhibits strong performance on benchmarks in coding, multi-step reasoning, and agent benchmarks (e.g., Tau2-Bench). While its initial variant emphasizes instruction following (non-thinking mode), Alibaba plans to bring reasoning capabilities online to enable autonomous agent behavior. Qwen3-Max inherits multilingual support and extensive pretraining on trillions of tokens, and it is delivered via API interfaces compatible with OpenAI-style functions.Starting Price: Free -
22
MiMo-V2.5-Pro
Xiaomi Technology
Xiaomi MiMo-V2.5-Pro is an advanced open-source AI model designed to handle complex, long-horizon tasks with strong agentic capabilities. It features a Mixture-of-Experts architecture with over one trillion parameters and a large context window of up to one million tokens. The model is built to perform sophisticated reasoning, coding, and problem-solving across extended workflows. It demonstrates high performance on benchmark tests related to software engineering, reasoning, and general intelligence. MiMo-V2.5-Pro can autonomously complete complex projects, such as building full software systems or optimizing engineering designs. It uses hybrid attention mechanisms to balance efficiency and performance across long contexts. The model is also optimized for token efficiency, reducing computational cost while maintaining strong results. By combining scalability, efficiency, and advanced reasoning, MiMo-V2.5-Pro represents a major step forward in open-source AI models. -
23
DeepSeek R1
DeepSeek
DeepSeek-R1 is an advanced open-source reasoning model developed by DeepSeek, designed to rival OpenAI's Model o1. Accessible via web, app, and API, it excels in complex tasks such as mathematics and coding, demonstrating superior performance on benchmarks like the American Invitational Mathematics Examination (AIME) and MATH. DeepSeek-R1 employs a mixture of experts (MoE) architecture with 671 billion total parameters, activating 37 billion parameters per token, enabling efficient and accurate reasoning capabilities. This model is part of DeepSeek's commitment to advancing artificial general intelligence (AGI) through open-source innovation.Starting Price: Free -
24
Ministral 3
Mistral AI
Mistral 3 is the latest generation of open-weight AI models from Mistral AI, offering a full family of models, from small, edge-optimized versions to a flagship, large-scale multimodal model. The lineup includes three compact “Ministral 3” models (3B, 8B, and 14B parameters) designed for efficiency and deployment on constrained hardware (even laptops, drones, or edge devices), plus the powerful “Mistral Large 3,” a sparse mixture-of-experts model with 675 billion total parameters (41 billion active). The models support multimodal and multilingual tasks, not only text, but also image understanding, and have demonstrated best-in-class performance on general prompts, multilingual conversations, and multimodal inputs. The base and instruction-fine-tuned versions are released under the Apache 2.0 license, enabling broad customization and integration in enterprise and open source projects.Starting Price: Free -
25
Mistral Small 4
Mistral AI
Mistral Small 4 is an advanced open-source AI model developed by Mistral AI that combines reasoning, coding, and multimodal capabilities into a single system. It unifies the strengths of previous models such as Magistral for reasoning, Pixtral for multimodal processing, and Devstral for agentic coding tasks. The model can handle both text and image inputs, allowing it to perform tasks ranging from conversational chat to visual analysis and document understanding. Built with a mixture-of-experts architecture, Mistral Small 4 delivers efficient performance while scaling to complex workloads. It also features a configurable reasoning parameter that allows users to switch between fast responses and deeper analytical outputs. With a large context window and optimized inference performance, the model supports long-form interactions and complex workflows.Starting Price: Free -
26
Olmo 3
Ai2
Olmo 3 is a fully open model family spanning 7 billion and 32 billion parameter variants that delivers not only high-performing base, reasoning, instruction, and reinforcement-learning models, but also exposure of the entire model flow, including raw training data, intermediate checkpoints, training code, long-context support (65,536 token window), and provenance tooling. Starting with the Dolma 3 dataset (≈9 trillion tokens) and its disciplined mix of web text, scientific PDFs, code, and long-form documents, the pre-training, mid-training, and long-context phases shape the base models, which are then post-trained via supervised fine-tuning, direct preference optimisation, and RL with verifiable rewards to yield the Think and Instruct variants. The 32 B Think model is described as the strongest fully open reasoning model to date, competitively close to closed-weight peers in math, code, and complex reasoning.Starting Price: Free -
27
GPT-5.4 Pro
OpenAI
GPT-5.4 Pro is an advanced AI model developed by OpenAI to deliver high-performance capabilities for professional and complex tasks. It combines improvements in reasoning, coding, and agent-based workflows into a single unified system. The model is designed to work efficiently across professional tools such as spreadsheets, presentations, documents, and development environments. GPT-5.4 Pro also includes native computer-use capabilities, enabling AI agents to interact with software, websites, and operating systems to complete tasks. With support for up to one million tokens of context, it can manage long workflows and large datasets more effectively than previous models. The model also improves tool usage, allowing it to search for and select the right tools during multi-step processes. By delivering more accurate outputs with fewer tokens, GPT-5.4 Pro helps professionals complete complex work faster and more efficiently. -
28
GLM-4.7-Flash
Z.ai
GLM-4.7 Flash is a lightweight variant of GLM-4.7, Z.ai’s flagship large language model designed for advanced coding, reasoning, and multi-step task execution with strong agentic performance and a very large context window. It is an MoE-based model optimized for efficient inference that balances performance and resource use, enabling deployment on local machines with moderate memory requirements while maintaining deep reasoning, coding, and agentic task abilities. GLM-4.7 itself advances over earlier generations with enhanced programming capabilities, stable multi-step reasoning, context preservation across turns, and improved tool-calling workflows, and supports very long context lengths (up to ~200 K tokens) for complex tasks that span large inputs or outputs. The Flash variant retains many of these strengths in a smaller footprint, offering competitive benchmark performance in coding and reasoning tasks for models in its size class.Starting Price: Free -
29
Nemotron 3
NVIDIA
NVIDIA Nemotron 3 is a family of open large language models developed by NVIDIA to power advanced reasoning, conversational AI, and autonomous AI agents. The Nemotron 3 series includes three models designed for different scales of AI workloads while maintaining high efficiency and accuracy. These models focus on “agentic AI” capabilities, meaning they can perform multi-step reasoning, coordinate with tools, and operate as components within multi-agent systems used in automation, research, and enterprise applications. The architecture uses a hybrid mixture-of-experts (MoE) design combined with transformer-based techniques, allowing the model to activate only a subset of parameters for each task, which improves performance while reducing computational cost. Nemotron 3 models are built to deliver strong reasoning, conversational, and planning abilities while maintaining high throughput for large-scale deployment. -
30
Kimi K2.6
Moonshot AI
Kimi K2.6 is a next-generation agentic AI model developed by Moonshot AI, designed to push forward real-world execution, coding, and multi-step reasoning beyond earlier K2 and K2.5 versions. It builds on a Mixture-of-Experts architecture and the multimodal, agent-first foundation of the Kimi series, combining language understanding, coding, and tool use into a single system capable of planning and executing complex workflows. It introduces deeper reasoning capabilities and significantly improved agent planning, allowing it to break down tasks, coordinate tools, and handle multi-file or multi-step problems with greater accuracy and efficiency. It supports advanced tool calling with high reliability, enabling integration with external systems such as web search or APIs, and includes built-in validation mechanisms to ensure correct execution formats.Starting Price: Free -
31
SubQ
Subquadratic
SubQ is a large language model developed by Subquadratic, designed specifically for long-context reasoning tasks. It can process up to 12 million tokens in a single prompt, allowing it to analyze entire codebases, long histories, and complex datasets at once. The model uses a sub-quadratic sparse-attention architecture that improves efficiency by focusing only on the most relevant relationships in the data. This approach reduces computational overhead while maintaining strong performance on large-scale tasks. SubQ is optimized for use cases such as software engineering, coding agents, and long-context retrieval. It delivers fast processing speeds and operates at a lower cost compared to many traditional models. Developers can access SubQ through APIs or integrate it into coding tools for enhanced workflows. Its architecture enables scalable AI reasoning without the limitations of standard transformer models. -
32
Llama 4 Scout
Meta
Llama 4 Scout is a powerful 17 billion active parameter multimodal AI model that excels in both text and image processing. With an industry-leading context length of 10 million tokens, it outperforms its predecessors, including Llama 3, in tasks such as multi-document summarization and parsing large codebases. Llama 4 Scout is designed to handle complex reasoning tasks while maintaining high efficiency, making it perfect for use cases requiring long-context comprehension and image grounding. It offers cutting-edge performance in image-related tasks and is particularly well-suited for applications requiring both text and visual understanding.Starting Price: Free -
33
GPT-5.4
OpenAI
GPT-5.4 is an advanced artificial intelligence model developed by OpenAI to support complex professional and technical work. The model combines improvements in reasoning, coding, and agent-based workflows into a single system designed for real-world productivity tasks. GPT-5.4 can generate, analyze, and edit documents, spreadsheets, presentations, and other work outputs with greater accuracy and efficiency. It also features improved tool integration, enabling the model to interact with software environments and external tools to complete multi-step workflows. With enhanced context capabilities supporting up to one million tokens, GPT-5.4 can process and reason over very large amounts of information. The model also improves factual accuracy and reduces errors compared to earlier versions. By combining strong reasoning, coding ability, and tool use, GPT-5.4 helps users complete complex tasks faster and with fewer iterations. -
34
Qwen3.6-27B
Alibaba
Qwen3.6-27B is a dense, open source multimodal language model in the Qwen3.6 series, designed to deliver flagship-level performance in coding, reasoning, and agent-based workflows while maintaining a relatively efficient parameter size of 27 billion. It is positioned as a high-performance general model that “punches above its weight,” achieving results competitive with or superior to significantly larger models on key benchmarks, particularly in agentic coding tasks. It supports both thinking and non-thinking modes, allowing it to dynamically balance deep reasoning with fast responses depending on the task, and integrates capabilities across text and multimodal inputs such as images and video. Built as part of the Qwen3.6 family, the model emphasizes real-world usability, stability, and developer productivity, incorporating improvements driven by community feedback and practical deployment needs.Starting Price: Free -
35
GLM-5-Turbo
Z.ai
GLM-5-Turbo is a high-speed variant of Z.ai’s GLM-5 model, designed to deliver efficient and stable performance in agent-driven environments while maintaining strong reasoning and coding capabilities. It is optimized for high-throughput workloads, particularly long-chain agent tasks where multiple steps, tools, and decisions must be executed in sequence with reliability and low latency. It supports advanced agentic workflows, enabling systems to perform multi-step planning, tool calling, and task execution with improved responsiveness compared to larger flagship models. GLM-5-Turbo inherits core capabilities from the GLM-5 family, including strong reasoning, coding performance, and support for long-context processing, while focusing on optimization of core requirements such as speed, efficiency, and stability in production environments. It is designed to integrate with agent frameworks like OpenClaw, where it can coordinate actions, process inputs, and execute tasks.Starting Price: Free -
36
DeepSeek-V2
DeepSeek
DeepSeek-V2 is a state-of-the-art Mixture-of-Experts (MoE) language model introduced by DeepSeek-AI, characterized by its economical training and efficient inference capabilities. With a total of 236 billion parameters, of which only 21 billion are active per token, it supports a context length of up to 128K tokens. DeepSeek-V2 employs innovative architectures like Multi-head Latent Attention (MLA) for efficient inference by compressing the Key-Value (KV) cache and DeepSeekMoE for cost-effective training through sparse computation. This model significantly outperforms its predecessor, DeepSeek 67B, by saving 42.5% in training costs, reducing the KV cache by 93.3%, and enhancing generation throughput by 5.76 times. Pretrained on an 8.1 trillion token corpus, DeepSeek-V2 excels in language understanding, coding, and reasoning tasks, making it a top-tier performer among open-source models.Starting Price: Free -
37
Command R+
Cohere AI
Command R+ is Cohere's newest large language model, optimized for conversational interaction and long-context tasks. It aims at being extremely performant, enabling companies to move beyond proof of concept and into production. We recommend using Command R+ for those workflows that lean on complex RAG functionality and multi-step tool use (agents). Command R, on the other hand, is great for simpler retrieval augmented generation (RAG) and single-step tool use tasks, as well as applications where price is a major consideration.Starting Price: Free -
38
Step 3.5 Flash
StepFun
Step 3.5 Flash is an advanced open source foundation language model engineered for frontier reasoning and agentic capabilities with exceptional efficiency, built on a sparse Mixture of Experts (MoE) architecture that selectively activates only about 11 billion of its ~196 billion parameters per token to deliver high-density intelligence and real-time responsiveness. Its 3-way Multi-Token Prediction (MTP-3) enables generation throughput in the hundreds of tokens per second for complex multi-step reasoning chains and task execution, and it supports efficient long contexts with a hybrid sliding window attention approach that reduces computational overhead across large datasets or codebases. It demonstrates robust performance on benchmarks for reasoning, coding, and agentic tasks, rivaling or exceeding many larger proprietary models, and includes a scalable reinforcement learning framework for consistent self-improvement.Starting Price: Free -
39
MiniMax M1
MiniMax
MiniMax‑M1 is a large‑scale hybrid‑attention reasoning model released by MiniMax AI under the Apache 2.0 license. It supports an unprecedented 1 million‑token context window and up to 80,000-token outputs, enabling extended reasoning across long documents. Trained using large‑scale reinforcement learning with a novel CISPO algorithm, MiniMax‑M1 completed full training on 512 H800 GPUs in about three weeks. It achieves state‑of‑the‑art performance on benchmarks in mathematics, coding, software engineering, tool usage, and long‑context understanding, matching or outperforming leading models. Two model variants are available (40K and 80K thinking budgets), with weights and deployment scripts provided via GitHub and Hugging Face. -
40
Qwen3.5-Plus
Alibaba
Qwen3.5-Plus is a high-performance native vision-language model designed for efficient text generation, deep reasoning, and multimodal understanding. Built on a hybrid architecture that combines linear attention with a sparse mixture-of-experts design, it delivers strong performance while optimizing inference efficiency. The model supports text, image, and video inputs and produces text outputs, making it suitable for complex multimodal workflows. With a massive 1 million token context window and up to 64K output tokens, Qwen3.5-Plus enables long-form reasoning and large-scale document analysis. It includes advanced capabilities such as structured outputs, function calling, web search, and tool integration via the Responses API. The model supports prefix continuation, caching, batch processing, and fine-tuning for flexible deployment. Designed for developers and enterprises, Qwen3.5-Plus provides scalable, high-throughput AI performance with OpenAI-compatible API access.Starting Price: $0.4 per 1M tokens -
41
DeepSeek-Coder-V2
DeepSeek
DeepSeek-Coder-V2 is an open source code language model designed to excel in programming and mathematical reasoning tasks. It features a Mixture-of-Experts (MoE) architecture with 236 billion total parameters and 21 billion activated parameters per token, enabling efficient processing and high performance. The model was trained on an extensive dataset of 6 trillion tokens, enhancing its capabilities in code generation and mathematical problem-solving. DeepSeek-Coder-V2 supports over 300 programming languages and has demonstrated superior performance on benchmarks such surpassing other models. It is available in multiple variants, including DeepSeek-Coder-V2-Instruct, optimized for instruction-based tasks; DeepSeek-Coder-V2-Base, suitable for general text generation; and lightweight versions like DeepSeek-Coder-V2-Lite-Base and DeepSeek-Coder-V2-Lite-Instruct, designed for environments with limited computational resources. -
42
DeepSeek R2
DeepSeek
DeepSeek R2 is the anticipated successor to DeepSeek R1, a groundbreaking AI reasoning model launched in January 2025 by the Chinese AI startup DeepSeek. Building on R1’s success, which disrupted the AI industry with its cost-effective performance rivaling top-tier models like OpenAI’s o1, R2 promises a quantum leap in capabilities. It is expected to deliver exceptional speed and human-like reasoning, excelling in complex tasks such as advanced coding and high-level mathematical problem-solving. Leveraging DeepSeek’s innovative Mixture-of-Experts architecture and efficient training methods, R2 aims to outperform its predecessor while maintaining a low computational footprint, potentially expanding its reasoning abilities to languages beyond English.Starting Price: Free -
43
Xiaomi MiMo
Xiaomi Technology
The Xiaomi MiMo API open platform is a developer-oriented interface for accessing and integrating Xiaomi’s MiMo family of AI models, including reasoning and language models such as MiMo-V2-Flash, into applications and services through standardized APIs and cloud endpoints, enabling developers to build AI-enabled features like conversational agents, reasoning workflows, code assistance, and search-augmented tasks without managing model infrastructure themselves. It offers REST-style API access with authentication, request signing, and structured responses so software can send prompts and receive generated text or processed outputs programmatically, and it supports common operations like text generation, prompt handling, and inference over MiMo models. By providing documentation and onboarding tools, the open platform lets teams integrate Xiaomi’s latest open source large language models, which leverage Mixture-of-Experts (MoE) architectures.Starting Price: Free -
44
Command A Reasoning
Cohere AI
Command A Reasoning is Cohere’s most advanced enterprise-ready language model, engineered for high-stakes reasoning tasks and seamless integration into AI agent workflows. The model delivers exceptional reasoning performance, efficiency, and controllability, scaling across multi-GPU setups with support for up to 256,000-token context windows, ideal for handling long documents and multi-step agentic tasks. Organizations can fine-tune output precision and latency through a token budget, allowing a single model to flexibly serve both high-accuracy and high-throughput use cases. It powers Cohere’s North platform with leading benchmark performance and excels in multilingual contexts across 23 languages. Designed with enterprise safety in mind, it balances helpfulness with robust safeguards against harmful outputs. A lightweight deployment option allows running the model securely on a single H100 or A100 GPU, simplifying private, scalable use. -
45
Mistral Large 3
Mistral AI
Mistral Large 3 is a next-generation, open multimodal AI model built with a powerful sparse Mixture-of-Experts architecture featuring 41B active parameters out of 675B total. Designed from scratch on NVIDIA H200 GPUs, it delivers frontier-level reasoning, multilingual performance, and advanced image understanding while remaining fully open-weight under the Apache 2.0 license. The model achieves top-tier results on modern instruction benchmarks, positioning it among the strongest permissively licensed foundation models available today. With native support across vLLM, TensorRT-LLM, and major cloud providers, Mistral Large 3 offers exceptional accessibility and performance efficiency. Its design enables enterprise-grade customization, letting teams fine-tune or adapt the model for domain-specific workflows and proprietary applications. Mistral Large 3 represents a major advancement in open AI, offering frontier intelligence without sacrificing transparency or control.Starting Price: Free -
46
Qwen3.6
Alibaba
Qwen3.6 is a large language model developed by Alibaba as part of its Qwen AI model family, designed for real-world applications and advanced reasoning tasks. It focuses on improving stability, usability, and performance compared to earlier versions. The model supports multimodal capabilities, allowing it to process and reason across text, images, and other data types. Qwen3.6 is particularly strong in coding and developer workflows, offering improved accuracy for complex programming tasks. It uses a mixture-of-experts architecture, enabling efficient performance while maintaining large-scale model capabilities. The model is designed to be deployable in production environments, including enterprise and cloud-based systems. It can be integrated into applications or run locally using open-weight variants. Overall, Qwen3.6 delivers a powerful, efficient, and versatile AI solution for modern use cases.Starting Price: Free -
47
DeepSeek-V3.2-Speciale
DeepSeek
DeepSeek-V3.2-Speciale is a high-compute variant of the DeepSeek-V3.2 model, created specifically for deep reasoning and advanced problem-solving tasks. It builds on DeepSeek Sparse Attention (DSA), a custom long-context attention mechanism that reduces computational overhead while preserving high performance. Through a large-scale reinforcement learning framework and extensive post-training compute, the Speciale variant surpasses GPT-5 on reasoning benchmarks and matches the capabilities of Gemini-3.0-Pro. The model achieved gold-medal performance in the International Mathematical Olympiad (IMO) 2025 and International Olympiad in Informatics (IOI) 2025. DeepSeek-V3.2-Speciale does not support tool-calling, making it purely optimized for uninterrupted reasoning and analytical accuracy. Released under the MIT license, it provides researchers and developers an open, state-of-the-art model focused entirely on high-precision reasoning.Starting Price: Free -
48
QwQ-32B
Alibaba
QwQ-32B is an advanced reasoning model developed by Alibaba Cloud's Qwen team, designed to enhance AI's problem-solving capabilities. With 32 billion parameters, it achieves performance comparable to state-of-the-art models like DeepSeek's R1, which has 671 billion parameters. This efficiency is achieved through optimized parameter utilization, allowing QwQ-32B to perform complex tasks such as mathematical reasoning, coding, and general problem-solving with fewer resources. The model supports a context length of up to 32,000 tokens, enabling it to process extensive input data effectively. QwQ-32B is accessible via Alibaba's chatbot service, Qwen Chat, and is open sourced under the Apache 2.0 license, promoting collaboration and further development within the AI community.Starting Price: Free -
49
Mixtral 8x22B
Mistral AI
Mixtral 8x22B is our latest open model. It sets a new standard for performance and efficiency within the AI community. It is a sparse Mixture-of-Experts (SMoE) model that uses only 39B active parameters out of 141B, offering unparalleled cost efficiency for its size. It is fluent in English, French, Italian, German, and Spanish. It has strong mathematics and coding capabilities. It is natively capable of function calling; along with the constrained output mode implemented on la Plateforme, this enables application development and tech stack modernization at scale. Its 64K tokens context window allows precise information recall from large documents. We build models that offer unmatched cost efficiency for their respective sizes, delivering the best performance-to-cost ratio within models provided by the community. Mixtral 8x22B is a natural continuation of our open model family. Its sparse activation patterns make it faster than any dense 70B model.Starting Price: Free -
50
GPT‑5.4 Thinking
OpenAI
GPT-5.4 Thinking is an advanced reasoning-focused AI model available within ChatGPT, designed to help users complete complex professional tasks more effectively. It combines improvements in reasoning, coding, and agent-based workflows to provide more accurate and reliable outputs. The model can present an upfront outline of its reasoning process, allowing users to adjust instructions while it is generating a response. This capability helps produce results that better align with user goals without requiring multiple follow-up prompts. GPT-5.4 Thinking also improves deep web research, enabling it to locate and synthesize information from multiple sources more efficiently. With stronger context management, it can handle longer conversations and complex problem-solving tasks with greater coherence. These capabilities make GPT-5.4 Thinking well suited for professional knowledge work and advanced analytical tasks.