Alternatives to CodeQwen
Compare CodeQwen alternatives for your business or organization using the curated list below. SourceForge ranks the best alternatives to CodeQwen in 2026. Compare features, ratings, user reviews, pricing, and more from CodeQwen competitors and alternatives in order to make an informed decision for your business.
-
1
Gemini Code Assist
Google
Increase software development and delivery velocity using generative AI assistance, with enterprise security and privacy protection. Gemini Code Assist completes your code as you write, and generates whole code blocks or functions on demand. Code assistance is available in many popular IDEs, such as Visual Studio Code, JetBrains IDEs (IntelliJ, PyCharm, GoLand, WebStorm, and more), Cloud Workstations, Cloud Shell Editor, and supports 20+ programming languages, including Java, JavaScript, Python, C, C++, Go, PHP, and SQL. Through a natural language chat interface, you can quickly chat with Gemini Code Assist to get answers to your coding questions, or receive guidance on coding best practices. Chat is available in all supported IDEs. Enterprises can customize Gemini Code Assist using their organization’s private codebases and knowledge sources so that Gemini Code Assist can offer more tailored assistance. Gemini Code Assist enables large-scale changes to entire codebases.Starting Price: Free -
2
Cody
Sourcegraph
Cody, Sourcegraph’s AI code assistant goes beyond individual dev productivity, helping enterprises achieve consistency and quality at scale with AI. Unlike traditional coding assistants, Cody understands the entire codebase, enabling deeper contextual awareness for smarter autocompletions, refactoring, and AI-driven code suggestions. It integrates with IDEs like VS Code, Visual Studio, Eclipse, and JetBrains, providing inline editing and chat without disrupting workflows. Cody also connects with tools like Notion, Linear, and Prometheus to enhance development context. Powered by advanced LLMs like Claude Sonnet 4 and GPT-4o, it optimizes speed and performance based on enterprise needs, and is always adding the latest AI models. Developers report significant efficiency gains, with some saving up to six hours per week and doubling their coding speed.Starting Price: $59 -
3
BLACKBOX AI
BLACKBOX AI
BLACKBOX AI is an advanced AI-powered platform designed to accelerate coding, app development, and deep research tasks. It features an AI Coding Agent that supports real-time voice interaction, GPU acceleration, and remote parallel task execution. Users can convert Figma designs into functional code and transform images into web applications with minimal coding effort. The platform enables screen sharing within IDEs like VSCode and offers mobile access to coding agents. BLACKBOX AI also supports integration with GitHub repositories for streamlined remote workflows. Its capabilities extend to website design, app building with PDF context, and image generation and editing.Starting Price: Free -
4
Amazon CodeWhisperer
Amazon
Build apps faster with ML-powered coding companion. Accelerate application development with automatic code recommendations based on the code and comments in your IDE. Empower developers to use artificial intelligence (AI) responsibly to create syntactically correct and secure applications. Generate entire functions and logical code blocks without having to search and customize code snippets from the web. Stay focused and never leave the IDE, with real-time customized code recommendations for all your Java, Python, and JavaScript projects. Amazon CodeWhisperer is a machine learning (ML)–powered service that helps improve developer productivity by generating code recommendations based on their comments in natural language and code in the integrated development environment (IDE). Accelerate frontend and backend development by empowering developers with automatic code recommendations. Save time and effort by using CodeWhisperer to generate code to build and train your ML models. -
5
CodeGemma
Google
CodeGemma is a collection of powerful, lightweight models that can perform a variety of coding tasks like fill-in-the-middle code completion, code generation, natural language understanding, mathematical reasoning, and instruction following. CodeGemma has 3 model variants, a 7B pre-trained variant that specializes in code completion and generation from code prefixes and/or suffixes, a 7B instruction-tuned variant for natural language-to-code chat and instruction following; and a state-of-the-art 2B pre-trained variant that provides up to 2x faster code completion. Complete lines, and functions, and even generate entire blocks of code, whether you're working locally or using Google Cloud resources. Trained on 500 billion tokens of primarily English language data from web documents, mathematics, and code, CodeGemma models generate code that's not only more syntactically correct but also semantically meaningful, reducing errors and debugging time. -
6
Qwen-7B
Alibaba
Qwen-7B is the 7B-parameter version of the large language model series, Qwen (abbr. Tongyi Qianwen), proposed by Alibaba Cloud. Qwen-7B is a Transformer-based large language model, which is pretrained on a large volume of data, including web texts, books, codes, etc. Additionally, based on the pretrained Qwen-7B, we release Qwen-7B-Chat, a large-model-based AI assistant, which is trained with alignment techniques. The features of the Qwen-7B series include: Trained with high-quality pretraining data. We have pretrained Qwen-7B on a self-constructed large-scale high-quality dataset of over 2.2 trillion tokens. The dataset includes plain texts and codes, and it covers a wide range of domains, including general domain data and professional domain data. Strong performance. In comparison with the models of the similar model size, we outperform the competitors on a series of benchmark datasets, which evaluates natural language understanding, mathematics, coding, etc. And more.Starting Price: Free -
7
Qwen2
Alibaba
Qwen2 is the large language model series developed by Qwen team, Alibaba Cloud. Qwen2 is a series of large language models developed by the Qwen team at Alibaba Cloud. It includes both base language models and instruction-tuned models, ranging from 0.5 billion to 72 billion parameters, and features both dense models and a Mixture-of-Experts model. The Qwen2 series is designed to surpass most previous open-weight models, including its predecessor Qwen1.5, and to compete with proprietary models across a broad spectrum of benchmarks in language understanding, generation, multilingual capabilities, coding, mathematics, and reasoning.Starting Price: Free -
8
Qwen2.5-Max
Alibaba
Qwen2.5-Max is a large-scale Mixture-of-Experts (MoE) model developed by the Qwen team, pretrained on over 20 trillion tokens and further refined through Supervised Fine-Tuning (SFT) and Reinforcement Learning from Human Feedback (RLHF). In evaluations, it outperforms models like DeepSeek V3 in benchmarks such as Arena-Hard, LiveBench, LiveCodeBench, and GPQA-Diamond, while also demonstrating competitive results in other assessments, including MMLU-Pro. Qwen2.5-Max is accessible via API through Alibaba Cloud and can be explored interactively on Qwen Chat.Starting Price: Free -
9
Qwen3-Max
Alibaba
Qwen3-Max is Alibaba’s latest trillion-parameter large language model, designed to push performance in agentic tasks, coding, reasoning, and long-context processing. It is built atop the Qwen3 family and benefits from the architectural, training, and inference advances introduced there; mixing thinker and non-thinker modes, a “thinking budget” mechanism, and support for dynamic mode switching based on complexity. The model reportedly processes extremely long inputs (hundreds of thousands of tokens), supports tool invocation, and exhibits strong performance on benchmarks in coding, multi-step reasoning, and agent benchmarks (e.g., Tau2-Bench). While its initial variant emphasizes instruction following (non-thinking mode), Alibaba plans to bring reasoning capabilities online to enable autonomous agent behavior. Qwen3-Max inherits multilingual support and extensive pretraining on trillions of tokens, and it is delivered via API interfaces compatible with OpenAI-style functions.Starting Price: Free -
10
Qwen3
Alibaba
Qwen3, the latest iteration of the Qwen family of large language models, introduces groundbreaking features that enhance performance across coding, math, and general capabilities. With models like the Qwen3-235B-A22B and Qwen3-30B-A3B, Qwen3 achieves impressive results compared to top-tier models, thanks to its hybrid thinking modes that allow users to control the balance between deep reasoning and quick responses. The platform supports 119 languages and dialects, making it an ideal choice for global applications. Its pre-training process, which uses 36 trillion tokens, enables robust performance, and advanced reinforcement learning (RL) techniques continue to refine its capabilities. Available on platforms like Hugging Face and ModelScope, Qwen3 offers a powerful tool for developers and researchers working in diverse fields.Starting Price: Free -
11
Qwen
Alibaba
Qwen is a powerful, free AI assistant built on the advanced Qwen model series, designed to help anyone with creativity, research, problem-solving, and everyday tasks. While Qwen Chat is the main interface for most users, Qwen itself powers a broad range of intelligent capabilities including image generation, deep research, website creation, advanced reasoning, and context-aware search. Its multimodal intelligence enables Qwen to understand and process text, images, audio, and video simultaneously for richer insights. Qwen is available on web, desktop, and mobile, ensuring seamless access across all devices. For developers, the Qwen API provides OpenAI-compatible endpoints, making integration simple and allowing Qwen’s intelligence to power apps, services, and automation. Whether you're chatting through Qwen Chat or building with the Qwen API, Qwen delivers fast, flexible, and highly capable AI support.Starting Price: Free -
12
QwQ-32B
Alibaba
QwQ-32B is an advanced reasoning model developed by Alibaba Cloud's Qwen team, designed to enhance AI's problem-solving capabilities. With 32 billion parameters, it achieves performance comparable to state-of-the-art models like DeepSeek's R1, which has 671 billion parameters. This efficiency is achieved through optimized parameter utilization, allowing QwQ-32B to perform complex tasks such as mathematical reasoning, coding, and general problem-solving with fewer resources. The model supports a context length of up to 32,000 tokens, enabling it to process extensive input data effectively. QwQ-32B is accessible via Alibaba's chatbot service, Qwen Chat, and is open sourced under the Apache 2.0 license, promoting collaboration and further development within the AI community.Starting Price: Free -
13
Qwen2.5-1M
Alibaba
Qwen2.5-1M is an open-source language model developed by the Qwen team, designed to handle context lengths of up to one million tokens. This release includes two model variants, Qwen2.5-7B-Instruct-1M and Qwen2.5-14B-Instruct-1M, marking the first time Qwen models have been upgraded to support such extensive context lengths. To facilitate efficient deployment, the team has also open-sourced an inference framework based on vLLM, integrated with sparse attention methods, enabling processing of 1M-token inputs with a 3x to 7x speed improvement. Comprehensive technical details, including design insights and ablation experiments, are available in the accompanying technical report.Starting Price: Free -
14
Qwen3-Coder
Qwen
Qwen3‑Coder is an agentic code model available in multiple sizes, led by the 480B‑parameter Mixture‑of‑Experts variant (35B active) that natively supports 256K‑token contexts (extendable to 1M) and achieves state‑of‑the‑art results comparable to Claude Sonnet 4. Pre‑training on 7.5T tokens (70 % code) and synthetic data cleaned via Qwen2.5‑Coder optimized both coding proficiency and general abilities, while post‑training employs large‑scale, execution‑driven reinforcement learning, scaling test‑case generation for diverse coding challenges, and long‑horizon RL across 20,000 parallel environments to excel on multi‑turn software‑engineering benchmarks like SWE‑Bench Verified without test‑time scaling. Alongside the model, the open source Qwen Code CLI (forked from Gemini Code) unleashes Qwen3‑Coder in agentic workflows with customized prompts, function calling protocols, and seamless integration with Node.js, OpenAI SDKs, and environment variables.Starting Price: Free -
15
Qwen Code
Qwen
Qwen3‑Coder is an agentic code model available in multiple sizes, led by the 480B‑parameter Mixture‑of‑Experts variant (35B active) that natively supports 256K‑token contexts (extendable to 1M) and achieves state‑of‑the‑art results on Agentic Coding, Browser‑Use, and Tool‑Use tasks comparable to Claude Sonnet 4. Pre‑training on 7.5T tokens (70 % code) and synthetic data cleaned via Qwen2.5‑Coder optimized both coding proficiency and general abilities, while post‑training employs large‑scale, execution‑driven reinforcement learning and long‑horizon RL across 20,000 parallel environments to excel on multi‑turn software‑engineering benchmarks like SWE‑Bench Verified without test‑time scaling. Alongside the model, the open source Qwen Code CLI (forked from Gemini Code) unleashes Qwen3‑Coder in agentic workflows with customized prompts, function calling protocols, and seamless integration with Node.js, OpenAI SDKs, and more.Starting Price: Free -
16
Qwen Chat
Alibaba
Qwen Chat is a versatile and powerful AI platform developed by Alibaba, offering an array of functionalities through a user-friendly web interface. It integrates multiple advanced Qwen AI models, allowing users to engage in text-based conversations, generate images and videos, perform web searches, and utilize various tools for enhanced productivity. With features like document and image processing, HTML preview for coding tasks, and the ability to create and test artifacts directly within the chat, Qwen Chat caters to developers, researchers, and AI enthusiasts. Users can switch between models seamlessly to fit different needs, from general conversation to specialized coding or vision tasks. The platform promises future updates including voice interaction, making it an evolving tool for diverse AI applications.Starting Price: Free -
17
Alibaba Cloud Model Studio
Alibaba
Model Studio is Alibaba Cloud’s one-stop generative AI platform that lets developers build intelligent, business-aware applications using industry-leading foundation models like Qwen-Max, Qwen-Plus, Qwen-Turbo, the Qwen-2/3 series, visual-language models (Qwen-VL/Omni), and the video-focused Wan series. Users can access these powerful GenAI models through familiar OpenAI-compatible APIs or purpose-built SDKs, no infrastructure setup required. It supports a full development workflow, experiment with models in the playground, perform real-time and batch inferences, fine-tune with tools like SFT or LoRA, then evaluate, compress, accelerate deployment, and monitor performance, all within an isolated Virtual Private Cloud (VPC) for enterprise-grade security. Customization is simplified via one-click Retrieval-Augmented Generation (RAG), enabling integration of business data into model outputs. Visual, template-driven interfaces facilitate prompt engineering and application design. -
18
Qwen2-VL
Alibaba
Qwen2-VL is the latest version of the vision language models based on Qwen2 in the Qwen model familities. Compared with Qwen-VL, Qwen2-VL has the capabilities of: SoTA understanding of images of various resolution & ratio: Qwen2-VL achieves state-of-the-art performance on visual understanding benchmarks, including MathVista, DocVQA, RealWorldQA, MTVQA, etc. Understanding videos of 20 min+: Qwen2-VL can understand videos over 20 minutes for high-quality video-based question answering, dialog, content creation, etc. Agent that can operate your mobiles, robots, etc.: with the abilities of complex reasoning and decision making, Qwen2-VL can be integrated with devices like mobile phones, robots, etc., for automatic operation based on visual environment and text instructions. Multilingual Support: to serve global users, besides English and Chinese, Qwen2-VL now supports the understanding of texts in different languages inside imagesStarting Price: Free -
19
Qwen2.5-VL
Alibaba
Qwen2.5-VL is the latest vision-language model from the Qwen series, representing a significant advancement over its predecessor, Qwen2-VL. This model excels in visual understanding, capable of recognizing a wide array of objects, including text, charts, icons, graphics, and layouts within images. It functions as a visual agent, capable of reasoning and dynamically directing tools, enabling applications such as computer and phone usage. Qwen2.5-VL can comprehend videos exceeding one hour in length and can pinpoint relevant segments within them. Additionally, it accurately localizes objects in images by generating bounding boxes or points and provides stable JSON outputs for coordinates and attributes. The model also supports structured outputs for data like scanned invoices, forms, and tables, benefiting sectors such as finance and commerce. Available in base and instruct versions across 3B, 7B, and 72B sizes, Qwen2.5-VL is accessible through platforms like Hugging Face and ModelScope.Starting Price: Free -
20
Qwen2.5-Coder
Alibaba
Qwen2.5-Coder-32B-Instruct has become the current SOTA open source code model, matching the coding capabilities of GPT-4o. While demonstrating strong and comprehensive coding abilities, it also possesses good general and mathematical skills. As of now, Qwen2.5-Coder has covered six mainstream model sizes to meet the needs of different developers. We explore the practicality of Qwen2.5-Coder in two scenarios, including code assistants and artifacts, with some examples showcasing the potential applications of Qwen2.5-Coder in real-world scenarios. Qwen2.5-Coder-32B-Instruct, as the flagship model of this open source release, has achieved the best performance among open source models on multiple popular code generation benchmarks and has competitive performance with GPT-4o. Code repair is an important programming skill. Qwen2.5-Coder-32B-Instruct can help users fix errors in their code, making programming more efficient.Starting Price: Free -
21
Qwen3-VL
Alibaba
Qwen3-VL is the newest vision-language model in the Qwen family (by Alibaba Cloud), designed to fuse powerful text understanding/generation with advanced visual and video comprehension into one unified multimodal model. It accepts inputs in mixed modalities, text, images, and video, and handles long, interleaved contexts natively (up to 256 K tokens, with extensibility beyond). Qwen3-VL delivers major advances in spatial reasoning, visual perception, and multimodal reasoning; the model architecture incorporates several innovations such as Interleaved-MRoPE (for robust spatio-temporal positional encoding), DeepStack (to leverage multi-level features from its Vision Transformer backbone for refined image-text alignment), and text–timestamp alignment (for precise reasoning over video content and temporal events). These upgrades enable Qwen3-VL to interpret complex scenes, follow dynamic video sequences, read and reason about visual layouts.Starting Price: Free -
22
QwQ-Max-Preview
Alibaba
QwQ-Max-Preview is an advanced AI model built on the Qwen2.5-Max architecture, designed to excel in deep reasoning, mathematical problem-solving, coding, and agent-related tasks. This preview version offers a sneak peek at its capabilities, which include improved performance in a wide range of general-domain tasks and the ability to handle complex workflows. QwQ-Max-Preview is slated for an official open-source release under the Apache 2.0 license, offering further advancements and refinements in its full version. It also paves the way for a more accessible AI ecosystem, with the upcoming launch of the Qwen Chat app and smaller variants of the model like QwQ-32B, aimed at developers seeking local deployment options.Starting Price: Free -
23
Qwen2.5
Alibaba
Qwen2.5 is an advanced multimodal AI model designed to provide highly accurate and context-aware responses across a wide range of applications. It builds on the capabilities of its predecessors, integrating cutting-edge natural language understanding with enhanced reasoning, creativity, and multimodal processing. Qwen2.5 can seamlessly analyze and generate text, interpret images, and interact with complex data to deliver precise solutions in real time. Optimized for adaptability, it excels in personalized assistance, data analysis, creative content generation, and academic research, making it a versatile tool for professionals and everyday users alike. Its user-centric design emphasizes transparency, efficiency, and alignment with ethical AI practices.Starting Price: Free -
24
Smaug-72B
Abacus
Smaug-72B is a powerful open-source large language model (LLM) known for several key features: High Performance: It currently holds the top spot on the Hugging Face Open LLM leaderboard, surpassing models like GPT-3.5 in various benchmarks. This means it excels at tasks like understanding, responding to, and generating human-like text. Open Source: Unlike many other advanced LLMs, Smaug-72B is freely available for anyone to use and modify, fostering collaboration and innovation in the AI community. Focus on Reasoning and Math: It specifically shines in handling reasoning and mathematical tasks, attributing this strength to unique fine-tuning techniques developed by Abacus AI, the creators of Smaug-72B. Based on Qwen-72B: It's technically a fine-tuned version of another powerful LLM called Qwen-72B, released by Alibaba, further improving upon its capabilities. Overall, Smaug-72B represents a significant step forward in open-source AI.Starting Price: Free -
25
CodeGeeX
AMiner
We introduce CodeGeeX, a large-scale multilingual code generation model with 13 billion parameters, pre-trained on a large code corpus of more than 20 programming languages. Based on CodeGeeX, we develop a VS Code extension (search 'CodeGeeX' in the Extension Marketplace) that assists the programming of different programming languages. Besides the multilingual code generation/translation abilities, we turn CodeGeeX into a custom programming assistant using its few-shot ability. It means that when a few examples are provided as extra prompts in the input, CodeGeeX will imitate what are done by these examples and generate codes accordingly. Some cool features can be implemented using this ability, like code explanation, summarization, generation with specific coding style, and more. For example, one can add code snippets with his/her own coding style, and CodeGeeX will generate codes in a similar way. You can also try prompts with specific formats to inspire CodeGeeX for new skills.Starting Price: Free -
26
Code Llama
Meta
Code Llama is a large language model (LLM) that can use text prompts to generate code. Code Llama is state-of-the-art for publicly available LLMs on code tasks, and has the potential to make workflows faster and more efficient for current developers and lower the barrier to entry for people who are learning to code. Code Llama has the potential to be used as a productivity and educational tool to help programmers write more robust, well-documented software. Code Llama is a state-of-the-art LLM capable of generating code, and natural language about code, from both code and natural language prompts. Code Llama is free for research and commercial use. Code Llama is built on top of Llama 2 and is available in three models: Code Llama, the foundational code model; Codel Llama - Python specialized for Python; and Code Llama - Instruct, which is fine-tuned for understanding natural language instructions.Starting Price: Free -
27
Llama 2
Meta
The next generation of our open source large language model. This release includes model weights and starting code for pretrained and fine-tuned Llama language models — ranging from 7B to 70B parameters. Llama 2 pretrained models are trained on 2 trillion tokens, and have double the context length than Llama 1. Its fine-tuned models have been trained on over 1 million human annotations. Llama 2 outperforms other open source language models on many external benchmarks, including reasoning, coding, proficiency, and knowledge tests. Llama 2 was pretrained on publicly available online data sources. The fine-tuned model, Llama-2-chat, leverages publicly available instruction datasets and over 1 million human annotations. We have a broad range of supporters around the world who believe in our open approach to today’s AI — companies that have given early feedback and are excited to build with Llama 2.Starting Price: Free -
28
Tülu 3
Ai2
Tülu 3 is an advanced instruction-following language model developed by the Allen Institute for AI (Ai2), designed to enhance capabilities in areas such as knowledge, reasoning, mathematics, coding, and safety. Built upon the Llama 3 Base, Tülu 3 employs a comprehensive four-stage post-training process: meticulous prompt curation and synthesis, supervised fine-tuning on a diverse set of prompts and completions, preference tuning using both off- and on-policy data, and a novel reinforcement learning approach to bolster specific skills with verifiable rewards. This open-source model distinguishes itself by providing full transparency, including access to training data, code, and evaluation tools, thereby closing the performance gap between open and proprietary fine-tuning methods. Evaluations indicate that Tülu 3 outperforms other open-weight models of similar size, such as Llama 3.1-Instruct and Qwen2.5-Instruct, across various benchmarks.Starting Price: Free -
29
Codestral
Mistral AI
We introduce Codestral, our first-ever code model. Codestral is an open-weight generative AI model explicitly designed for code generation tasks. It helps developers write and interact with code through a shared instruction and completion API endpoint. As it masters code and English, it can be used to design advanced AI applications for software developers. Codestral is trained on a diverse dataset of 80+ programming languages, including the most popular ones, such as Python, Java, C, C++, JavaScript, and Bash. It also performs well on more specific ones like Swift and Fortran. This broad language base ensures Codestral can assist developers in various coding environments and projects.Starting Price: Free -
30
Code Snippets AI
Code Snippets AI
Turn your questions into code. Easily store and fetch your snippets. Collaborate with your team. Powered by ChatGPT & our fine-tuned GPT3 model. Gain a deeper understanding of your code to further your knowledge. Increase the quality of your code with our refactor and debug features. Securely share code snippets with your team, without losing formatting. We use ChatGPT & our fine-tuned GPT3 Model, which provides faster and more accurate responses to your questions, compared to Codex apps. Create documentation, refactor, debug, and generate code with the click of a button. We use a fine-tuned AI model trained on GPT3, which provides faster and more accurate responses to your questions, compared to Codex apps. Save your code from your IDE straight into your library with our VSCode extension. Search snippets by language, name, or folder. Create your own folder structure to suit your needs. We use ChatGPT & our fine-tuned GPT3 Model, which provides faster and more accurate responses.Starting Price: $2 per month -
31
Qwen2.5-VL-32B
Alibaba
Qwen2.5-VL-32B is a state-of-the-art AI model designed for multimodal tasks, offering advanced capabilities in both text and image reasoning. It builds upon the earlier Qwen2.5-VL series, improving response quality with more human-like, formatted answers. The model excels in mathematical reasoning, fine-grained image understanding, and complex, multi-step reasoning tasks, such as those found in MathVista and MMMU benchmarks. Its superior performance has been demonstrated in comparison to other models, outperforming the larger Qwen2-VL-72B in certain areas. With improved image parsing and visual logic deduction, Qwen2.5-VL-32B provides a detailed, accurate analysis of images and can generate responses based on complex visual inputs. It has been optimized for both text and image tasks, making it ideal for applications requiring sophisticated reasoning and understanding across different media. -
32
Kimi K2
Moonshot AI
Kimi K2 is a state-of-the-art open source large language model series built on a mixture-of-experts (MoE) architecture, featuring 1 trillion total parameters and 32 billion activated parameters for task-specific efficiency. Trained with the Muon optimizer on over 15.5 trillion tokens and stabilized by MuonClip’s attention-logit clamping, it delivers exceptional performance in frontier knowledge, reasoning, mathematics, coding, and general agentic workflows. Moonshot AI provides two variants, Kimi-K2-Base for research-level fine-tuning and Kimi-K2-Instruct pre-trained for immediate chat and tool-driven interactions, enabling both custom development and drop-in agentic capabilities. Benchmarks show it outperforms leading open source peers and rivals top proprietary models in coding tasks and complex task breakdowns, while its 128 K-token context length, tool-calling API compatibility, and support for industry-standard inference engines.Starting Price: Free -
33
MonoQwen-Vision
LightOn
MonoQwen2-VL-v0.1 is the first visual document reranker designed to enhance the quality of retrieved visual documents in Retrieval-Augmented Generation (RAG) pipelines. Traditional RAG approaches rely on converting documents into text using Optical Character Recognition (OCR), which can be time-consuming and may result in loss of information, especially for non-textual elements like graphs and tables. MonoQwen2-VL-v0.1 addresses these limitations by leveraging Visual Language Models (VLMs) that process images directly, eliminating the need for OCR and preserving the integrity of visual content. This reranker operates in a two-stage pipeline, initially, it uses separate encoding to generate a pool of candidate documents, followed by a cross-encoding model that reranks these candidates based on their relevance to the query. By training a Low-Rank Adaptation (LoRA) on top of the Qwen2-VL-2B-Instruct model, MonoQwen2-VL-v0.1 achieves high performance without significant memory overhead. -
34
StarCoder
BigCode
StarCoder and StarCoderBase are Large Language Models for Code (Code LLMs) trained on permissively licensed data from GitHub, including from 80+ programming languages, Git commits, GitHub issues, and Jupyter notebooks. Similar to LLaMA, we trained a ~15B parameter model for 1 trillion tokens. We fine-tuned StarCoderBase model for 35B Python tokens, resulting in a new model that we call StarCoder. We found that StarCoderBase outperforms existing open Code LLMs on popular programming benchmarks and matches or surpasses closed models such as code-cushman-001 from OpenAI (the original Codex model that powered early versions of GitHub Copilot). With a context length of over 8,000 tokens, the StarCoder models can process more input than any other open LLM, enabling a wide range of interesting applications. For example, by prompting the StarCoder models with a series of dialogues, we enabled them to act as a technical assistant.Starting Price: Free -
35
Sky-T1
NovaSky
Sky-T1-32B-Preview is an open source reasoning model developed by the NovaSky team at UC Berkeley's Sky Computing Lab. It matches the performance of proprietary models like o1-preview on reasoning and coding benchmarks, yet was trained for under $450, showcasing the feasibility of cost-effective, high-level reasoning capabilities. The model was fine-tuned from Qwen2.5-32B-Instruct using a curated dataset of 17,000 examples across diverse domains, including math and coding. The training was completed in 19 hours on eight H100 GPUs with DeepSpeed Zero-3 offloading. All aspects of the project, including data, code, and model weights, are fully open-source, empowering the academic and open-source communities to replicate and enhance the model's performance.Starting Price: Free -
36
JetBrains AI Assistant
JetBrains
JetBrains AI Assistant connects your IDE to third-party and proprietary large language models (LLMs) optimized for specific tasks. This enables AI Assistant to provide precise and context-aware assistance, whether it's generating code snippets, completing complex code structures, or creating automated tests. We take data and code security seriously! JetBrains AI Assistant does not store or disclose your code snippets, regardless of whether the data originates from public repositories, private repositories, or local files. We strictly follow the JetBrains AI Terms of Service, ensuring that neither we nor our service providers use your data or code for training any generative AI models.Starting Price: $8.33 per month -
37
ChatGPT
OpenAI
ChatGPT is an AI-powered conversational assistant developed by OpenAI that helps users with writing, learning, brainstorming, coding, and more. It is free to use with easy access via web and apps on multiple devices. Users can interact through typing or voice to get answers, generate creative content, summarize information, and automate tasks. The platform supports various use cases, from casual questions to complex research and coding help. ChatGPT offers multiple subscription plans, including Free, Plus, and Pro, with increasing access to advanced AI models and features. It is designed to boost productivity and creativity for individuals, students, professionals, and developers alike.Starting Price: Free -
38
DeepScaleR
Agentica Project
DeepScaleR is a 1.5-billion-parameter language model fine-tuned from DeepSeek-R1-Distilled-Qwen-1.5B using distributed reinforcement learning and a novel iterative context-lengthening strategy that gradually increases its context window from 8K to 24K tokens during training. It was trained on ~40,000 carefully curated mathematical problems drawn from competition-level datasets like AIME (1984–2023), AMC (pre-2023), Omni-MATH, and STILL. DeepScaleR achieves 43.1% accuracy on AIME 2024, a roughly 14.3 percentage point boost over the base model, and surpasses the performance of the proprietary O1-Preview model despite its much smaller size. It also posts strong results on a suite of math benchmarks (e.g., MATH-500, AMC 2023, Minerva Math, OlympiadBench), demonstrating that small, efficient models tuned with RL can match or exceed larger baselines on reasoning tasks.Starting Price: Free -
39
Void Editor
Void Editor
Void is an open source AI code editor and Cursor alternative built as a fork of VS Code, enabling developers to write code with advanced AI assistance while retaining full control over their data. It supports seamless integration with any large language model, such as DeepSeek, Llama, Qwen, Gemini, Claude, and Grok, connecting directly without routing through a private backend. Core features include tab‑triggered autocomplete, inline quick edit, and a versatile AI chat interface offering normal chat, a restricted gather mode for read/search-only tasks, and an agent mode that automates file and folder operations, terminal commands, and MCP tool access. Void delivers high‑performance operations, including fast apply on files with thousands of lines, alongside checkpoint management for model updates, native tool execution, and lint error detection. Developers can transfer all themes, keybindings, and settings from VS Code in one click and host models locally or via the cloud.Starting Price: Free -
40
OPT
Meta
Large language models, which are often trained for hundreds of thousands of compute days, have shown remarkable capabilities for zero- and few-shot learning. Given their computational cost, these models are difficult to replicate without significant capital. For the few that are available through APIs, no access is granted to the full model weights, making them difficult to study. We present Open Pre-trained Transformers (OPT), a suite of decoder-only pre-trained transformers ranging from 125M to 175B parameters, which we aim to fully and responsibly share with interested researchers. We show that OPT-175B is comparable to GPT-3, while requiring only 1/7th the carbon footprint to develop. We are also releasing our logbook detailing the infrastructure challenges we faced, along with code for experimenting with all of the released models. -
41
Continue
Continue
The leading open-source AI code assistant. You can connect any models and any context to create custom autocomplete and chat experiences inside the IDE Remain in flow while coding by removing the barriers that block productivity when building software. Accelerate development with a plug-and-play system that makes it easy to get started and integrates with your entire stack. Become a leader in AI by setting up your code assistant to evolve as new capabilities emerge. Continue autocompletes single lines or entire sections of code in any programming language as you type. Attach code or other context to ask questions about functions, files, the entire codebase, and more. Highlight code sections and press a keyboard shortcut to rewrite code from natural language.Starting Price: $0/developer/month -
42
Granite Code
IBM
We introduce the Granite series of decoder-only code models for code generative tasks (e.g., fixing bugs, explaining code, documenting code), trained with code written in 116 programming languages. A comprehensive evaluation of the Granite Code model family on diverse tasks demonstrates that our models consistently reach state-of-the-art performance among available open source code LLMs. The key advantages of Granite Code models include: All-rounder Code LLM: Granite Code models achieve competitive or state-of-the-art performance on different kinds of code-related tasks, including code generation, explanation, fixing, editing, translation, and more. Demonstrating their ability to solve diverse coding tasks. Trustworthy Enterprise-Grade LLM: All our models are trained on license-permissible data collected following IBM's AI Ethics principles and guided by IBM’s Corporate Legal team for trustworthy enterprise usage.Starting Price: Free -
43
Mercury Coder
Inception Labs
Mercury, the latest innovation from Inception Labs, is the first commercial-scale diffusion large language model (dLLM), offering a 10x speed increase and significantly lower costs compared to traditional autoregressive models. Built for high-performance reasoning, coding, and structured text generation, Mercury processes over 1000 tokens per second on NVIDIA H100 GPUs, making it one of the fastest LLMs available. Unlike conventional models that generate text one token at a time, Mercury refines responses using a coarse-to-fine diffusion approach, improving accuracy and reducing hallucinations. With Mercury Coder, a specialized coding model, developers can experience cutting-edge AI-driven code generation with superior speed and efficiency.Starting Price: Free -
44
Featherless
Featherless
Featherless is an AI model provider that offers our subscribers access to a continually expanding library of Hugging Face models. With hundreds of new models daily, you need dedicated tools to keep up with the hype. No matter your use case, find and use the state-of-the-art AI model with Featherless. At present, we support LLaMA-3-based models, including LLaMA-3 and QWEN-2. Note that QWEN-2 models are only supported up to 16,000 context length. We plan to add more architectures to our supported list soon. We continuously onboard new models as they become available on Hugging Face. As we grow, we aim to automate this process to encompass all publicly available Hugging Face models with compatible architecture. To ensure fair individual account use, concurrent requests are limited according to the plan you've selected. Output is delivered at a speed of 10-40 tokens per second, depending on the model and prompt size.Starting Price: $10 per month -
45
kluster.ai
kluster.ai
Kluster.ai is a developer-centric AI cloud platform designed to deploy, scale, and fine-tune large language models (LLMs) with speed and efficiency. Built for developers by developers, it offers Adaptive Inference, a flexible and scalable service that adjusts seamlessly to workload demands, ensuring high-performance processing and consistent turnaround times. Adaptive Inference provides three distinct processing options: real-time inference for ultra-low latency needs, asynchronous inference for cost-effective handling of flexible timing tasks, and batch inference for efficient processing of high-volume, bulk tasks. It supports a range of open-weight, cutting-edge multimodal models for chat, vision, code, and more, including Meta's Llama 4 Maverick and Scout, Qwen3-235B-A22B, DeepSeek-R1, and Gemma 3 . Kluster.ai's OpenAI-compatible API allows developers to integrate these models into their applications seamlessly.Starting Price: $0.15per input -
46
DeepCoder
Agentica Project
DeepCoder is a fully open source code-reasoning and generation model released by Agentica Project in collaboration with Together AI. It is fine-tuned from DeepSeek-R1-Distilled-Qwen-14B using distributed reinforcement learning, achieving a 60.6% accuracy on LiveCodeBench (representing an 8% improvement over the base), a performance level that matches that of proprietary models such as o3-mini (2025-01-031 Low) and o1 while using only 14 billion parameters. It was trained over 2.5 weeks on 32 H100 GPUs with a curated dataset of roughly 24,000 coding problems drawn from verified sources (including TACO-Verified, PrimeIntellect SYNTHETIC-1, and LiveCodeBench submissions), each problem requiring a verifiable solution and at least five unit tests to ensure reliability for RL training. To handle long-range context, DeepCoder employs techniques such as iterative context lengthening and overlong filtering.Starting Price: Free -
47
Supernovas AI LLM
Supernovas AI LLM
Supernovas AI is a unified, team‑focused AI workspace that provides seamless access to all leading LLMs—including GPT‑4.1/4.5 Turbo, Claude Haiku/Sonnet/Opus, Gemini 2.5 Pro/Pro, Azure OpenAI, AWS Bedrock, Mistral, Meta LLaMA, Deepseek, Qwen, and more—through a single, secure interface. It offers essential chat tools like model access, prompt templates, bookmarks, static artifacts, and integrated web search, along with advanced features such as Model Context Protocol (MCP), a talk-to-your data knowledge base, built-in image generation and editing, memory‑enabled agents, and code execution. Supernovas AI simplifies AI tool management by eliminating multiple subscriptions and API keys, enabling fast onboarding and enterprise-grade privacy and collaboration—all from one streamlined platform.Starting Price: $19/month -
48
Athene-V2
Nexusflow
Athene-V2 is Nexusflow's latest 72-billion-parameter model suite, fine-tuned from Qwen 2.5 72B, designed to compete with GPT-4o across key capabilities. This suite includes Athene-V2-Chat-72B, a state-of-the-art chat model that matches GPT-4o in multiple benchmarks, excelling in chat helpfulness (Arena-Hard), code completion (ranking #2 on bigcode-bench-hard), mathematics (MATH), and precise long log extraction. Additionally, Athene-V2-Agent-72B balances chat and agent functionalities, offering concise, directive responses and surpassing GPT-4o in Nexus-V2 function calling benchmarks focused on complex enterprise-level use cases. These advancements underscore the industry's shift from merely scaling model sizes to specialized customization, illustrating how targeted post-training processes can finely optimize models for distinct skills and applications. -
49
Baichuan-13B
Baichuan Intelligent Technology
Baichuan-13B is an open source and commercially available large-scale language model containing 13 billion parameters developed by Baichuan Intelligent following Baichuan -7B . It has achieved the best results of the same size on authoritative Chinese and English benchmarks. This release contains two versions of pre-training ( Baichuan-13B-Base ) and alignment ( Baichuan-13B-Chat ). Larger size, more data : Baichuan-13B further expands the number of parameters to 13 billion on the basis of Baichuan -7B , and trains 1.4 trillion tokens on high-quality corpus, which is 40% more than LLaMA-13B. It is currently open source The model with the largest amount of training data in the 13B size. Support Chinese and English bilingual, use ALiBi position code, context window length is 4096.Starting Price: Free -
50
Olmo 3
Ai2
Olmo 3 is a fully open model family spanning 7 billion and 32 billion parameter variants that delivers not only high-performing base, reasoning, instruction, and reinforcement-learning models, but also exposure of the entire model flow, including raw training data, intermediate checkpoints, training code, long-context support (65,536 token window), and provenance tooling. Starting with the Dolma 3 dataset (≈9 trillion tokens) and its disciplined mix of web text, scientific PDFs, code, and long-form documents, the pre-training, mid-training, and long-context phases shape the base models, which are then post-trained via supervised fine-tuning, direct preference optimisation, and RL with verifiable rewards to yield the Think and Instruct variants. The 32 B Think model is described as the strongest fully open reasoning model to date, competitively close to closed-weight peers in math, code, and complex reasoning.Starting Price: Free