MiniMax M2
MiniMax M2 is an open source foundation model built specifically for agentic applications and coding workflows, striking a new balance of performance, speed, and cost. It excels in end-to-end development scenarios, handling programming, tool-calling, and complex, long-chain workflows with capabilities such as Python integration, while delivering inference speeds of around 100 tokens per second and offering API pricing at just ~8% of the cost of comparable proprietary models. The model supports “Lightning Mode” for high-speed, lightweight agent tasks, and “Pro Mode” for in-depth full-stack development, report generation, and web-based tool orchestration; its weights are fully open source and available for local deployment with vLLM or SGLang. MiniMax M2 positions itself as a production-ready model that enables agents to complete independent tasks, such as data analysis, programming, tool orchestration, and large-scale multi-step logic at real organizational scale.
Learn more
Amp
Amp is a frontier coding agent built to give developers full access to the power of today’s leading AI models directly in their workflow. Available in the terminal and popular editors like VS Code, Cursor, Windsurf, JetBrains, and Neovim, Amp integrates seamlessly into existing development environments. It enables developers to delegate complex coding tasks, refactors, reviews, and explorations to intelligent agents that understand and operate across entire codebases. With support for advanced models such as Claude Opus, Gemini, and GPT-class models, Amp delivers fast, reliable, and highly agentic code generation. The platform is designed for real-world engineering work, handling multi-file changes, deep context, and iterative improvements. Amp helps developers move faster while maintaining confidence in code quality.
Learn more
Cody
Cody, Sourcegraph’s AI code assistant goes beyond individual dev productivity, helping enterprises achieve consistency and quality at scale with AI.
Unlike traditional coding assistants, Cody understands the entire codebase, enabling deeper contextual awareness for smarter autocompletions, refactoring, and AI-driven code suggestions. It integrates with IDEs like VS Code, Visual Studio, Eclipse, and JetBrains, providing inline editing and chat without disrupting workflows. Cody also connects with tools like Notion, Linear, and Prometheus to enhance development context. Powered by advanced LLMs like Claude Sonnet 4 and GPT-4o, it optimizes speed and performance based on enterprise needs, and is always adding the latest AI models. Developers report significant efficiency gains, with some saving up to six hours per week and doubling their coding speed.
Learn more
Mistral Large 3
Mistral Large 3 is a next-generation, open multimodal AI model built with a powerful sparse Mixture-of-Experts architecture featuring 41B active parameters out of 675B total. Designed from scratch on NVIDIA H200 GPUs, it delivers frontier-level reasoning, multilingual performance, and advanced image understanding while remaining fully open-weight under the Apache 2.0 license. The model achieves top-tier results on modern instruction benchmarks, positioning it among the strongest permissively licensed foundation models available today. With native support across vLLM, TensorRT-LLM, and major cloud providers, Mistral Large 3 offers exceptional accessibility and performance efficiency. Its design enables enterprise-grade customization, letting teams fine-tune or adapt the model for domain-specific workflows and proprietary applications. Mistral Large 3 represents a major advancement in open AI, offering frontier intelligence without sacrificing transparency or control.
Learn more