74 Integrations with Cline
View a list of Cline integrations and software that integrates with Cline below. Compare the best Cline integrations as well as features, ratings, user reviews, and pricing of software that integrates with Cline. Here are the current Cline integrations in 2026:
-
1
Google Cloud Platform
Google
Google Cloud is a cloud-based service that allows you to create anything from simple websites to complex applications for businesses of all sizes. New customers get $300 in free credits to run, test, and deploy workloads. All customers can use 25+ products for free, up to monthly usage limits. Use Google's core infrastructure, data analytics & machine learning. Secure and fully featured for all enterprises. Tap into big data to find answers faster and build better products. Grow from prototype to production to planet-scale, without having to think about capacity, reliability or performance. From virtual machines with proven price/performance advantages to a fully managed app development platform. Scalable, resilient, high performance object storage and databases for your applications. State-of-the-art software-defined networking products on Google’s private fiber network. Fully managed data warehousing, batch and stream processing, data exploration, Hadoop/Spark, and messaging.Starting Price: Free ($300 in free credits) -
2
OpenRouter
OpenRouter
OpenRouter is a unified interface for LLMs. OpenRouter scouts for the lowest prices and best latencies/throughputs across dozens of providers, and lets you choose how to prioritize them. No need to change your code when switching between models or providers. You can even let users choose and pay for their own. Evals are flawed; instead, compare models by how often they're used for different purposes. Chat with multiple at once in the chatroom. Model usage can be paid by users, developers, or both, and may shift in availability. You can also fetch models, prices, and limits via API. OpenRouter routes requests to the best available providers for your model, given your preferences. By default, requests are load-balanced across the top providers to maximize uptime, but you can customize how this works using the provider object in the request body. Prioritize providers that have not seen significant outages in the last 10 seconds.Starting Price: $2 one-time payment -
3
Visual Studio Code
Microsoft
Visual Studio Code (VS Code) is Microsoft’s open-source AI code editor designed to make coding faster, smarter, and more collaborative. It supports thousands of extensions and nearly every programming language, offering developers a lightweight yet powerful environment for writing, testing, and debugging code. With AI-powered features like GitHub Copilot, Next Edit Suggestions, and Agent Mode, VS Code helps you code with precision, automate complex tasks, and streamline development workflows. It integrates seamlessly with cloud services, remote repositories, and tools like Git, Docker, and Azure. The editor is fully customizable, allowing you to personalize your layout, color themes, and keyboard shortcuts. Whether coding locally or in the browser, VS Code delivers a complete development experience for individuals and teams alike.Starting Price: Free -
4
MATLAB
The MathWorks
MATLAB® combines a desktop environment tuned for iterative analysis and design processes with a programming language that expresses matrix and array mathematics directly. It includes the Live Editor for creating scripts that combine code, output, and formatted text in an executable notebook. MATLAB toolboxes are professionally developed, rigorously tested, and fully documented. MATLAB apps let you see how different algorithms work with your data. Iterate until you’ve got the results you want, then automatically generate a MATLAB program to reproduce or automate your work. Scale your analyses to run on clusters, GPUs, and clouds with only minor code changes. There’s no need to rewrite your code or learn big data programming and out-of-memory techniques. Automatically convert MATLAB algorithms to C/C++, HDL, and CUDA code to run on your embedded processor or FPGA/ASIC. MATLAB works with Simulink to support Model-Based Design. -
5
OpenAI
OpenAI
OpenAI’s mission is to ensure that artificial general intelligence (AGI)—by which we mean highly autonomous systems that outperform humans at most economically valuable work—benefits all of humanity. We will attempt to directly build safe and beneficial AGI, but will also consider our mission fulfilled if our work aids others to achieve this outcome. Apply our API to any language task — semantic search, summarization, sentiment analysis, content generation, translation, and more — with only a few examples or by specifying your task in English. One simple integration gives you access to our constantly-improving AI technology. Explore how you integrate with the API with these sample completions. -
6
Claude
Anthropic
Claude is a next-generation AI assistant developed by Anthropic to help individuals and teams solve complex problems with safety, accuracy, and reliability at its core. It is designed to support a wide range of tasks, including writing, editing, coding, data analysis, and research. Claude allows users to create and iterate on documents, websites, graphics, and code directly within chat using collaborative tools like Artifacts. The platform supports file uploads, image analysis, and data visualization to enhance productivity and understanding. Claude is available across web, iOS, and Android, making it accessible wherever work happens. With built-in web search and extended reasoning capabilities, Claude helps users find information and think through challenging problems more effectively. Anthropic emphasizes security, privacy, and responsible AI development to ensure Claude can be trusted in professional and personal workflows.Starting Price: Free -
7
Gemini
Google
Gemini is Google’s advanced AI assistant designed to help users think, create, learn, and complete tasks with a new level of intelligence. Powered by Google’s most capable models, including Gemini 3, it enables users to ask complex questions, generate content, analyze information, and explore ideas through natural conversation. Gemini can create images, videos, summaries, study plans, and first drafts while also providing feedback on uploaded files and written work. The platform is grounded in Google Search, allowing it to deliver accurate, up-to-date information and support deep follow-up questions. Gemini connects seamlessly with Google apps like Gmail, Docs, Calendar, Maps, YouTube, and Photos to help users complete tasks without switching tools. Features such as Gemini Live, Deep Research, and Gems enhance brainstorming, research, and personalized workflows. Available through flexible free and paid plans, Gemini supports everyday users, students, and professionals across devices.Starting Price: Free -
8
Gemini Advanced
Google
Gemini Advanced is a cutting-edge AI model designed for unparalleled performance in natural language understanding, generation, and problem-solving across diverse domains. Featuring a revolutionary neural architecture, it delivers exceptional accuracy, nuanced contextual comprehension, and deep reasoning capabilities. Gemini Advanced is engineered to handle complex, multifaceted tasks, from creating detailed technical content and writing code to conducting in-depth data analysis and providing strategic insights. Its adaptability and scalability make it a powerful solution for both individual users and enterprise-level applications. Gemini Advanced sets a new standard for intelligence, innovation, and reliability in AI-powered solutions. You'll also get access to Gemini in Gmail, Docs, and more, 2 TB storage, and other benefits from Google One. Gemini Advanced also offers access to Gemini with Deep Research. You can conduct in-depth and real-time research on almost any subject.Starting Price: $19.99 per month -
9
React
React
React makes it painless to create interactive UIs. Design simple views for each state in your application, and React will efficiently update and render just the right components when your data changes. Declarative views make your code more predictable and easier to debug. Build encapsulated components that manage their own state, then compose them to make complex UIs. Since component logic is written in JavaScript instead of templates, you can easily pass rich data through your app and keep state out of the DOM. We don’t make assumptions about the rest of your technology stack, so you can develop new features in React without rewriting existing code. React components implement a render() method that takes input data and returns what to display. This example uses an XML-like syntax called JSX. Input data that is passed into the component can be accessed by render() via this.props.Starting Price: Free -
10
Java
Oracle
The Java™ Programming Language is a general-purpose, concurrent, strongly typed, class-based object-oriented language. It is normally compiled to the bytecode instruction set and binary format defined in the Java Virtual Machine Specification. In the Java programming language, all source code is first written in plain text files ending with the .java extension. Those source files are then compiled into .class files by the javac compiler. A .class file does not contain code that is native to your processor; it instead contains bytecodes — the machine language of the Java Virtual Machine1 (Java VM). The java launcher tool then runs your application with an instance of the Java Virtual Machine.Starting Price: Free -
11
Claude Sonnet 3.5
Anthropic
Claude Sonnet 3.5 sets new industry benchmarks for graduate-level reasoning (GPQA), undergraduate-level knowledge (MMLU), and coding proficiency (HumanEval). It shows marked improvement in grasping nuance, humor, and complex instructions, and is exceptional at writing high-quality content with a natural, relatable tone. Claude Sonnet 3.5 operates at twice the speed of Claude Opus 3. This performance boost, combined with cost-effective pricing, makes Claude Sonnet 3.5 ideal for complex tasks such as context-sensitive customer support and orchestrating multi-step workflows.Starting Price: Free -
12
Claude Sonnet 3.7
Anthropic
Claude Sonnet 3.7, developed by Anthropic, is a cutting-edge AI model that combines rapid response with deep reflective reasoning. This innovative model allows users to toggle between quick, efficient responses and more thoughtful, reflective answers, making it ideal for complex problem-solving. By allowing Claude to self-reflect before answering, it excels at tasks that require high-level reasoning and nuanced understanding. With its ability to engage in deeper thought processes, Claude Sonnet 3.7 enhances tasks such as coding, natural language processing, and critical thinking applications. Available across various platforms, it offers a powerful tool for professionals and organizations seeking a high-performance, adaptable AI.Starting Price: Free -
13
GPT-4.1
OpenAI
GPT-4.1 is an advanced AI model from OpenAI, designed to enhance performance across key tasks such as coding, instruction following, and long-context comprehension. With a large context window of up to 1 million tokens, GPT-4.1 can process and understand extensive datasets, making it ideal for tasks like software development, document analysis, and AI agent workflows. Available through the API, GPT-4.1 offers significant improvements over previous models, excelling at real-world applications where efficiency and accuracy are crucial.Starting Price: $2 per 1M tokens (input) -
14
Python
Python
The core of extensible programming is defining functions. Python allows mandatory and optional arguments, keyword arguments, and even arbitrary argument lists. Whether you're new to programming or an experienced developer, it's easy to learn and use Python. Python can be easy to pick up whether you're a first-time programmer or you're experienced with other languages. The following pages are a useful first step to get on your way to writing programs with Python! The community hosts conferences and meetups to collaborate on code, and much more. Python's documentation will help you along the way, and the mailing lists will keep you in touch. The Python Package Index (PyPI) hosts thousands of third-party modules for Python. Both Python's standard library and the community-contributed modules allow for endless possibilities.Starting Price: Free -
15
Gemini 2.0
Google
Gemini 2.0 is an advanced AI-powered model developed by Google, designed to offer groundbreaking capabilities in natural language understanding, reasoning, and multimodal interactions. Building on the success of its predecessor, Gemini 2.0 integrates large language processing with enhanced problem-solving and decision-making abilities, enabling it to interpret and generate human-like responses with greater accuracy and nuance. Unlike traditional AI models, Gemini 2.0 is trained to handle multiple data types simultaneously, including text, images, and code, making it a versatile tool for research, business, education, and creative industries. Its core improvements include better contextual understanding, reduced bias, and a more efficient architecture that ensures faster, more reliable outputs. Gemini 2.0 is positioned as a major step forward in the evolution of AI, pushing the boundaries of human-computer interaction.Starting Price: Free -
16
Gemini Pro
Google
Gemini Pro is a powerful multimodal AI model developed by Google as part of the broader Gemini family of large language models. It is designed to handle a wide range of tasks, including text generation, reasoning, coding, and data analysis. The model can process multiple types of input such as text, images, audio, and video, making it highly versatile for real-world applications. Gemini Pro is optimized for delivering accurate, context-aware responses across complex workflows. It integrates seamlessly with Google products and cloud services, enabling scalable AI-powered applications. The model is commonly used for tasks like content creation, summarization, and conversational AI. It balances performance and efficiency, making it suitable for both developers and enterprise users. Overall, it serves as a robust foundation for building intelligent AI-driven solutions. -
17
Gemini 2.0 Flash
Google
The Gemini 2.0 Flash AI model represents the next generation of high-speed, intelligent computing, designed to set new benchmarks in real-time language processing and decision-making. Building on the robust foundation of its predecessor, it incorporates enhanced neural architecture and breakthrough advancements in optimization, enabling even faster and more accurate responses. Gemini 2.0 Flash is designed for applications requiring instantaneous processing and adaptability, such as live virtual assistants, automated trading systems, and real-time analytics. Its lightweight, efficient design ensures seamless deployment across cloud, edge, and hybrid environments, while its improved contextual understanding and multitasking capabilities make it a versatile tool for tackling complex, dynamic workflows with precision and speed. -
18
Gemini Nano
Google
Gemini Nano from Google is a lightweight, energy-efficient AI model designed for high performance in compact, resource-constrained environments. Tailored for edge computing and mobile applications, Gemini Nano combines Google's advanced AI architecture with cutting-edge optimization techniques to deliver seamless performance without compromising speed or accuracy. Despite its compact size, it excels in tasks like voice recognition, natural language processing, real-time translation, and personalized recommendations. With a focus on privacy and efficiency, Gemini Nano processes data locally, minimizing reliance on cloud infrastructure while maintaining robust security. Its adaptability and low power consumption make it an ideal choice for smart devices, IoT ecosystems, and on-the-go AI solutions. -
19
Gemini 1.5 Pro
Google
The Gemini 1.5 Pro AI model is a state-of-the-art language model designed to deliver highly accurate, context-aware, and human-like responses across a variety of applications. Built with cutting-edge neural architecture, it excels in natural language understanding, generation, and reasoning tasks. The model is fine-tuned for versatility, supporting tasks like content creation, code generation, data analysis, and complex problem-solving. Its advanced algorithms ensure nuanced comprehension, enabling it to adapt to different domains and conversational styles seamlessly. With a focus on scalability and efficiency, the Gemini 1.5 Pro is optimized for both small-scale implementations and enterprise-level integrations, making it a powerful tool for enhancing productivity and innovation. -
20
Gemini 1.5 Flash
Google
The Gemini 1.5 Flash AI model is an advanced, high-speed language model engineered for lightning-fast processing and real-time responsiveness. Designed to excel in dynamic and time-sensitive applications, it combines streamlined neural architecture with cutting-edge optimization techniques to deliver exceptional performance without compromising on accuracy. Gemini 1.5 Flash is tailored for scenarios requiring rapid data processing, instant decision-making, and seamless multitasking, making it ideal for chatbots, customer support systems, and interactive applications. Its lightweight yet powerful design ensures it can be deployed efficiently across a range of platforms, from cloud-based environments to edge devices, enabling businesses to scale their operations with unmatched agility. -
21
Cognee
Cognee
Cognee is an open source AI memory engine that transforms raw data into structured knowledge graphs, enhancing the accuracy and contextual understanding of AI agents. It supports various data types, including unstructured text, media files, PDFs, and tables, and integrates seamlessly with several data sources. Cognee employs modular ECL pipelines to process and organize data, enabling AI agents to retrieve relevant information efficiently. It is compatible with vector and graph databases and supports LLM frameworks like OpenAI, LlamaIndex, and LangChain. Key features include customizable storage options, RDF-based ontologies for smart data structuring, and the ability to run on-premises, ensuring data privacy and compliance. Cognee's distributed system is scalable, capable of handling large volumes of data, and is designed to reduce AI hallucinations by providing AI agents with a coherent and interconnected data landscape.Starting Price: $25 per month -
22
Grok Code Fast 1
xAI
Grok Code Fast 1 is a high-speed, economical reasoning model designed specifically for agentic coding workflows. Unlike traditional models that can feel slow in tool-based loops, it delivers near-instant responses, excelling in everyday software development tasks. Built from scratch with a programming-rich corpus and refined on real-world pull requests, it supports languages like TypeScript, Python, Java, Rust, C++, and Go. Developers can use it for everything from zero-to-one project building to precise bug fixes and codebase Q&A. With optimized inference and caching techniques, it achieves impressive responsiveness and a 90%+ cache hit rate when integrated with partners like GitHub Copilot, Cursor, and Cline. Offered at just $0.20 per million input tokens and $1.50 per million output tokens, Grok Code Fast 1 strikes a strong balance between speed, performance, and affordability.Starting Price: $0.20 per million input tokens -
23
GLM-5
Zhipu AI
GLM-5 is Z.ai’s latest large language model built for complex systems engineering and long-horizon agentic tasks. It scales significantly beyond GLM-4.5, increasing total parameters and training data while integrating DeepSeek Sparse Attention to reduce deployment costs without sacrificing long-context capacity. The model combines enhanced pre-training with a new asynchronous reinforcement learning infrastructure called slime, improving training efficiency and post-training refinement. GLM-5 achieves best-in-class performance among open-source models across reasoning, coding, and agent benchmarks, narrowing the gap with leading frontier models. It ranks highly on evaluations such as Vending Bench 2, demonstrating strong long-term planning and operational capabilities. The model is open-sourced under the MIT License.Starting Price: Free -
24
GLM-5.1
Zhipu AI
GLM-5.1 is the latest iteration of Z.ai’s GLM series, designed as a frontier-level, agent-oriented AI model optimized for coding, reasoning, and long-horizon workflows. It builds on the GLM-5 architecture, which uses a Mixture-of-Experts (MoE) design to deliver high performance while keeping inference costs efficient, and is part of a broader push toward open-weight, developer-accessible models. A core focus of GLM-5.1 is enabling agentic behavior, meaning it can plan, execute, and iterate across multi-step tasks rather than simply responding to single prompts. It is specifically designed to handle complex workflows such as debugging code, navigating repositories, and executing chained operations with sustained context. Compared to earlier models, GLM-5.1 improves reliability in long interactions, maintaining coherence across extended sessions and reducing breakdowns in multi-step reasoning.Starting Price: Free -
25
Butterbase
Butterbase
Butterbase is a backend-as-a-service (BaaS) platform purpose-built for developers and non-developers who build apps with AI coding tools. Connect Butterbase to your AI coding tool — Claude Code, Cursor, Windsurf, Codex, Copilot, or any MCP-compatible tool — and your entire backend is provisioned automatically. Describe what you need in natural language and get a production-ready Postgres database, authentication, REST APIs, real-time subscriptions, file storage, and edge functions in seconds. No SQL, no DevOps, no backend experience required.Starting Price: $0/month -
26
Node.js
Node.js
As an asynchronous event-driven JavaScript runtime, Node.js is designed to build scalable network applications. Upon each connection, the callback is fired, but if there is no work to be done, Node.js will sleep. This is in contrast to today's more common concurrency model, in which OS threads are employed. Thread-based networking is relatively inefficient and very difficult to use. Furthermore, users of Node.js are free from worries of dead-locking the process, since there are no locks. Almost no function in Node.js directly performs I/O, so the process never blocks except when the I/O is performed using synchronous methods of Node.js standard library. Because nothing blocks, scalable systems are very reasonable to develop in Node.js. Node.js is similar in design to, and influenced by, systems like Ruby's Event Machine and Python's Twisted. Node.js takes the event model a bit further. It presents an event loop as a runtime construct instead of as a library.Starting Price: Free -
27
CSS
CSS
CSS, short for Cascading Style Sheets, is a style sheet language used by web developers to structure the HTML and other elements of a website. CSS is one of the most widely used languages on the web. For style sheets to work, it is important that your markup be free of errors. A convenient way to automatically fix markup errors is to use the HTML Tidy utility. This also tidies the markup making it easier to read and easier to edit. I recommend you regularly run Tidy over any markup you are editing. Tidy is very effective at cleaning up markup created by authoring tools with sloppy habits. Each style property starts with the property's name, then a colon and lastly the value for this property. When there is more than one style property in the list, you need to use a semicolon between each of them to delimit one property from the next.Starting Price: Free -
28
Kotlin
Kotlin
Easy to pick up, so you can create powerful applications immediately. Compatible with the Java ecosystem. Use your favorite JVM frameworks and libraries. Share application logic between web, mobile, and desktop platforms while keeping an experience native to users. Save time and get the benefit of unlimited access to features specific to these platforms. Kotlin has great support and many contributors in its fast-growing global community. Enjoy the benefits of a rich ecosystem with a wide range of community libraries. Help is never far away — consult extensive community resources or ask the Kotlin team directly. Kotlin Multiplatform Mobile is an SDK for iOS and Android app development. It offers all the combined benefits of creating cross-platform and native apps. Maintain a single codebase for networking, data storage, analytics, and the other logic of your Android and iOS apps.Starting Price: Free -
29
PHP
PHP
Fast, flexible and pragmatic, PHP powers everything from your blog to the most popular websites in the world. The PHP development team announces the immediate availability of PHP 8.0.20. When using the PHP.net website, there is even no need to get to a search box to access the content you would like to see quickly. You can use short PHP.net URLs to access pages directly.Starting Price: Free -
30
Swift
Apple
Writing Swift code is interactive and fun, the syntax is concise yet expressive, and Swift includes modern features developers love. Swift code is safe by design and produces software that runs lightning-fast. Swift is the result of the latest research on programming languages, combined with decades of experience building Apple platforms. Named parameters are expressed in a clean syntax that makes APIs in Swift even easier to read and maintain. Even better, you don’t even need to type semi-colons. Inferred types make code cleaner and less prone to mistakes, while modules eliminate headers and provide namespaces. To best support international languages and emoji, Strings are Unicode-correct and use a UTF-8 based encoding to optimize performance for a wide-variety of use cases. You can even write concurrent code with simple, built-in keywords that define asynchronous behavior, making your code more readable and less error-prone.Starting Price: Free -
31
TypeScript
TypeScript
TypeScript adds additional syntax to JavaScript to support a tighter integration with your editor. Catch errors early in your editor. TypeScript code converts to JavaScript, which runs anywhere JavaScript runs: In a browser, on Node.js or Deno and in your apps. TypeScript understands JavaScript and uses type inference to give you great tooling without additional code. TypeScript was used by 78% of the 2020 State of JS respondents, with 93% saying they would use it again. The most common kinds of errors that programmers write can be described as type errors: a certain kind of value was used where a different kind of value was expected. This could be due to simple typos, a failure to understand the API surface of a library, incorrect assumptions about runtime behavior, or other errors.Starting Price: Free -
32
R
The R Foundation
R is a language and environment for statistical computing and graphics. It is a GNU project which is similar to the S language and environment which was developed at Bell Laboratories (formerly AT&T, now Lucent Technologies) by John Chambers and colleagues. R can be considered as a different implementation of S. There are some important differences, but much code written for S runs unaltered under R. R provides a wide variety of statistical (linear and nonlinear modelling, classical statistical tests, time-series analysis, classification, clustering, …) and graphical techniques, and is highly extensible. The S language is often the vehicle of choice for research in statistical methodology, and R provides an Open Source route to participation in that activity. One of R’s strengths is the ease with which well-designed publication-quality plots can be produced, including mathematical symbols and formulae where needed.Starting Price: Free -
33
Rust
Rust
Rust is blazingly fast and memory-efficient: with no runtime or garbage collector, it can power performance-critical services, run on embedded devices, and easily integrate with other languages. Rust’s rich type system and ownership model guarantee memory-safety and thread-safety — enabling you to eliminate many classes of bugs at compile-time. Rust has great documentation, a friendly compiler with useful error messages, and top-notch tooling — an integrated package manager and build tool, smart multi-editor support with auto-completion and type inspections, an auto-formatter, and more. Whip up a CLI tool quickly with Rust’s robust ecosystem. Rust helps you maintain your app with confidence and distribute it with ease. Use Rust to supercharge your JavaScript, one module at a time. Publish to npm, bundle with webpack, and you’re off to the races.Starting Price: Free -
34
Go
Golang
With a strong ecosystem of tools and APIs on major cloud providers, it is easier than ever to build services with Go. With popular open source packages and a robust standard library, use Go to create fast and elegant CLIs. With enhanced memory performance and support for several IDEs, Go powers fast and scalable web applications. With fast build times, lean syntax, an automatic formatter and doc generator, Go is built to support both DevOps and SRE. Everything there is to know about Go. Get started on a new project or brush up for your existing Go code. An interactive introduction to Go in three sections. Each section concludes with a few exercises so you can practice what you've learned. The Playground allows anyone with a web browser to write Go code that we immediately compile, link, and run on our servers.Starting Price: Free -
35
Markdown
Markdown
Markdown allows you to write using an easy-to-read, easy-to-write plain text format, then convert it to structurally valid XHTML (or HTML). Thus, “Markdown” is two things: (1) a plain text formatting syntax; and (2) a software tool, written in Perl, that converts the plain text formatting to HTML. See the Syntax page for details pertaining to Markdown’s formatting syntax. You can try it out, right now, using the online Dingus. The overriding design goal for Markdown’s formatting syntax is to make it as readable as possible. The idea is that a Markdown-formatted document should be publishable as-is, as plain text, without looking like it’s been marked up with tags or formatting instructions. While Markdown’s syntax has been influenced by several existing text-to-HTML filters, the single biggest source of inspiration for Markdown’s syntax is the format of plain text email.Starting Price: Free -
36
Gemini Enterprise Agent Platform Notebooks provide a unified environment for data science workflows, combining the flexibility of Colab Enterprise with the power of Agent Platform Workbench. These notebooks enable users to explore data, build models, and deploy solutions without switching between multiple tools. With seamless integration into Google Cloud services like BigQuery and Apache Spark, users can analyze large datasets directly within the notebook interface. The platform supports rapid prototyping and model development by offering scalable compute resources and AI-powered coding assistance. It allows teams to move from experimentation to production efficiently using end-to-end workflows. Fully managed infrastructure ensures scalability, cost optimization, and minimal operational overhead. Enterprise-grade security features such as single sign-on and access controls provide a safe environment for development.Starting Price: $10 per GB
-
37
Ollama
Ollama
Ollama is an innovative platform that focuses on providing AI-powered tools and services, designed to make it easier for users to interact with and build AI-driven applications. Run AI models locally. By offering a range of solutions, including natural language processing models and customizable AI features, Ollama empowers developers, businesses, and organizations to integrate advanced machine learning technologies into their workflows. With an emphasis on usability and accessibility, Ollama strives to simplify the process of working with AI, making it an appealing option for those looking to harness the potential of artificial intelligence in their projects.Starting Price: Free -
38
ByteRover
ByteRover
ByteRover is a self-improving memory layer for AI coding agents that unifies the creation, retrieval, and sharing of “vibe-coding” memories across projects and teams. Designed for dynamic AI-assisted development, it integrates into any AI IDE via the Memory Compatibility Protocol (MCP) extension, enabling agents to automatically save and recall context without altering existing workflows. It provides instant IDE integration, automated memory auto-save and recall, intuitive memory management (create, edit, delete, and prioritize memories), and team-wide intelligence sharing to enforce consistent coding standards. These capabilities let developer teams of all sizes maximize AI coding efficiency, eliminate repetitive training, and maintain a centralized, searchable memory store. Install ByteRover’s extension in your IDE to start capturing and leveraging agent memory across projects in seconds.Starting Price: $19.99 per month -
39
ToolSDK.ai
ToolSDK.ai
ToolSDK.ai is a free TypeScript SDK and marketplace that accelerates building agentic AI applications by providing instant access to over 5,300+ MCP (Model Context Protocol) servers and composable tools with one line of code, enabling developers to wire up real-world workflows combining language models with external systems. The platform exposes a unified client for loading packaged MCP servers (e.g., search, email, CRM, task management, storage, analytics) and converting them into OpenAI-compatible tools, handling authentication, invocation, and result orchestration so assistants can call, compare, and act on live data from services like Gmail, Salesforce, Google Drive, ClickUp, Notion, Slack, GitHub, analytics platforms, and custom web search or automation endpoints. It includes example quick-start integrations, supports metadata and conditional logic in multi-step orchestrations, and makes scaling to parallel agents and complex pipelines straightforward.Starting Price: Free -
40
Dash0
Dash0
Dash0 is an OpenTelemetry-native observability platform that unifies metrics, logs, traces, and resources into one intuitive interface, enabling fast and context-rich monitoring without vendor lock-in. It centralizes Prometheus and OpenTelemetry metrics, supports powerful filtering of high-cardinality attributes, and provides heatmap drilldowns and detailed trace views to pinpoint errors and bottlenecks in real time. Users benefit from fully customizable dashboards built on Perses, with support for code-based configuration and Grafana import, plus seamless integration with predefined alerts, checks, and PromQL queries. Dash0's AI-enhanced tools, such as Log AI for automated severity inference and pattern extraction, enrich telemetry data without requiring users to even notice that AI is working behind the scenes. These AI capabilities power features like log classification, grouping, inferred severity tagging, and streamlined triage workflows through the SIFT framework.Starting Price: $0.20 per month -
41
SimpliflowAI
SimpliflowAI
Simpliflow AI centers on Loop, a unified MCP (Model Context Protocol) gateway that consolidates integrations across AI agents and tools into one orchestrated layer. With Loop, you mount external MCP servers and integrate multiple apps just once; then, instead of embedding every tool schema into the language model context (which bloats prompts), Simpliflow dynamically retrieves and executes the needed tool only when a query triggers it. This keeps LLM contexts compact and avoids hitting tool limits. Loop offers a dashboard to manage all integrations and MCP connections, supports over 1,500 pre-built integrations via managed OAuth, and enables compatibility with any MCP-enabled AI app. It enforces schema validation and integrity checks for security and reliability, giving power users fine-grained control while providing a safer, unified environment for AI workflows.Starting Price: Free -
42
GLM-4.6
Zhipu AI
GLM-4.6 advances upon its predecessor with stronger reasoning, coding, and agentic capabilities: it demonstrates clear improvements in inferential performance, supports tool use during inference, and more effectively integrates into agent frameworks. In benchmark tests spanning reasoning, coding, and agents, GLM-4.6 outperforms GLM-4.5 and shows competitive strength against models such as DeepSeek-V3.2-Exp and Claude Sonnet 4, though it still trails Claude Sonnet 4.5 in pure coding performance. In real-world tests using an extended “CC-Bench” suite across front-end development, tool building, data analysis, and algorithmic tasks, GLM-4.6 beats GLM-4.5 and approaches parity with Claude Sonnet 4, winning ~48.6% of head-to-head comparisons, while also achieving ~15% better token efficiency. GLM-4.6 is available via the Z.ai API, and developers can integrate it as an LLM backend or agent core using the platform’s API.Starting Price: Free -
43
Gemini Enterprise
Google
Gemini Enterprise is a comprehensive AI platform built by Google Cloud designed to bring the full power of Google’s advanced AI models, agent-creation tools, and enterprise-grade data access into everyday workflows. The solution offers a unified chat interface that lets employees interact with internal documents, applications, data sources, and custom AI agents. At its core, Gemini Enterprise comprises six key components: the Gemini family of large multimodal models, an agent orchestration workbench (formerly Google Agentspace), pre-built starter agents, robust data-integration connectors to business systems, extensive security and governance controls, and a partner ecosystem for tailored integrations. It is engineered to scale across departments and enterprises, enabling users to build no-code or low-code agents that automate tasks, such as research synthesis, customer support response, code assist, contract analysis, and more, while operating within corporate compliance standards.Starting Price: $21 per month -
44
MiniMax M2
MiniMax
MiniMax M2 is an open source foundation model built specifically for agentic applications and coding workflows, striking a new balance of performance, speed, and cost. It excels in end-to-end development scenarios, handling programming, tool-calling, and complex, long-chain workflows with capabilities such as Python integration, while delivering inference speeds of around 100 tokens per second and offering API pricing at just ~8% of the cost of comparable proprietary models. The model supports “Lightning Mode” for high-speed, lightweight agent tasks, and “Pro Mode” for in-depth full-stack development, report generation, and web-based tool orchestration; its weights are fully open source and available for local deployment with vLLM or SGLang. MiniMax M2 positions itself as a production-ready model that enables agents to complete independent tasks, such as data analysis, programming, tool orchestration, and large-scale multi-step logic at real organizational scale.Starting Price: $0.30 per million input tokens -
45
Devstral 2
Mistral AI
Devstral 2 is a next-generation, open source agentic AI model tailored for software engineering: it doesn’t just suggest code snippets, it understands and acts across entire codebases, enabling multi-file edits, bug fixes, refactoring, dependency resolution, and context-aware code generation. The Devstral 2 family includes a large 123-billion-parameter model as well as a smaller 24-billion-parameter variant (“Devstral Small 2”), giving teams flexibility; the larger model excels in heavy-duty coding tasks requiring deep context, while the smaller one can run on more modest hardware. With a vast context window of up to 256 K tokens, Devstral 2 can reason across extensive repositories, track project history, and maintain a consistent understanding of lengthy files, an advantage for complex, real-world projects. The CLI tracks project metadata, Git statuses, and directory structure to give the model context, making “vibe-coding” more powerful.Starting Price: Free -
46
Devstral Small 2
Mistral AI
Devstral Small 2 is the compact, 24 billion-parameter variant of the new coding-focused model family from Mistral AI, released under the permissive Apache 2.0 license to enable both local deployment and API use. Alongside its larger sibling (Devstral 2), this model brings “agentic coding” capabilities to environments with modest compute: it supports a large 256K-token context window, enabling it to understand and make changes across entire codebases. On the standard code-generation benchmark (SWE-Bench Verified), Devstral Small 2 scores around 68.0%, placing it among open-weight models many times its size. Because of its reduced size and efficient design, Devstral Small 2 can run on a single GPU or even CPU-only setups, making it practical for developers, small teams, or hobbyists without access to data-center hardware. Despite its compact footprint, Devstral Small 2 retains key capabilities of larger models; it can reason across multiple files and track dependencies.Starting Price: Free -
47
GLM-4.6V
Zhipu AI
GLM-4.6V is a state-of-the-art open source multimodal vision-language model from the Z.ai (GLM-V) family designed for reasoning, perception, and action. It ships in two variants: a full-scale version (106B parameters) for cloud or high-performance clusters, and a lightweight “Flash” variant (9B) optimized for local deployment or low-latency use. GLM-4.6V supports a native context window of up to 128K tokens during training, enabling it to process very long documents or multimodal inputs. Crucially, it integrates native Function Calling, meaning the model can take images, screenshots, documents, or other visual media as input directly (without manual text conversion), reason about them, and trigger tool calls, bridging “visual perception” with “executable action.” This enables a wide spectrum of capabilities; interleaved image-and-text content generation (for example, combining document understanding with text summarization or generation of image-annotated responses).Starting Price: Free -
48
GLM-4.1V
Zhipu AI
GLM-4.1V is a vision-language model, providing a powerful, compact multimodal model designed for reasoning and perception across images, text, and documents. The 9-billion-parameter variant (GLM-4.1V-9B-Thinking) is built on the GLM-4-9B foundation and enhanced through a specialized training paradigm using Reinforcement Learning with Curriculum Sampling (RLCS). It supports a 64k-token context window and accepts high-resolution inputs (up to 4K images, any aspect ratio), enabling it to handle complex tasks such as optical character recognition, image captioning, chart and document parsing, video and scene understanding, GUI-agent workflows (e.g., interpreting screenshots, recognizing UI elements), and general vision-language reasoning. In benchmark evaluations at the 10 B-parameter scale, GLM-4.1V-9B-Thinking achieved top performance on 23 of 28 tasks.Starting Price: Free -
49
GLM-4.5V-Flash
Zhipu AI
GLM-4.5V-Flash is an open source vision-language model, designed to bring strong multimodal capabilities into a lightweight, deployable package. It supports image, video, document, and GUI inputs, enabling tasks such as scene understanding, chart and document parsing, screen reading, and multi-image analysis. Compared to larger models in the series, GLM-4.5V-Flash offers a compact footprint while retaining core VLM capabilities like visual reasoning, video understanding, GUI task handling, and complex document parsing. It can serve in “GUI agent” workflows, meaning it can interpret screenshots or desktop captures, recognize icons or UI elements, and assist with automated desktop or web-based tasks. Although it forgoes some of the largest-model performance gains, GLM-4.5V-Flash remains versatile for real-world multimodal tasks where efficiency, lower resource usage, and broad modality support are prioritized.Starting Price: Free -
50
GLM-4.5V
Zhipu AI
GLM-4.5V builds on the GLM-4.5-Air foundation, using a Mixture-of-Experts (MoE) architecture with 106 billion total parameters and 12 billion activation parameters. It achieves state-of-the-art performance among open-source VLMs of similar scale across 42 public benchmarks, excelling in image, video, document, and GUI-based tasks. It supports a broad range of multimodal capabilities, including image reasoning (scene understanding, spatial recognition, multi-image analysis), video understanding (segmentation, event recognition), complex chart and long-document parsing, GUI-agent workflows (screen reading, icon recognition, desktop automation), and precise visual grounding (e.g., locating objects and returning bounding boxes). GLM-4.5V also introduces a “Thinking Mode” switch, allowing users to choose between fast responses or deeper reasoning when needed.Starting Price: Free