Best Artificial Intelligence Software for Vertex AI - Page 2

Compare the Top Artificial Intelligence Software that integrates with Vertex AI as of November 2025 - Page 2

This a list of Artificial Intelligence software that integrates with Vertex AI. Use the filters on the left to add additional filters for products that have integrations with Vertex AI. View the products that work with Vertex AI in the table below.

  • 1
    Datasaur

    Datasaur

    Datasaur

    Welcome to the best tool for managing your labeling team, improving data quality, and working 70% faster—all in one place.
    Starting Price: $349/month
  • 2
    Arize AI

    Arize AI

    Arize AI

    Automatically discover issues, diagnose problems, and improve models with Arize’s machine learning observability platform. Machine learning systems address mission critical needs for businesses and their customers every day, yet often fail to perform in the real world. Arize is an end-to-end observability platform to accelerate detecting and resolving issues for your AI models at large. Seamlessly enable observability for any model, from any platform, in any environment. Lightweight SDKs to send training, validation, and production datasets. Link real-time or delayed ground truth to predictions. Gain foresight and confidence that your models will perform as expected once deployed. Proactively catch any performance degradation, data/prediction drift, and quality issues before they spiral. Reduce the time to resolution (MTTR) for even the most complex models with flexible, easy-to-use tools for root cause analysis.
    Starting Price: $50/month
  • 3
    GPTConsole

    GPTConsole

    GPTConsole

    GPTConsole is a revolutionary command-line interface (CLI) designed to empower developers with the benefits of artificial intelligence. It transcends traditional terminal functionalities, enabling you to execute complex tasks using prompts. To install GPTConsole, you can use either yarn or npm package managers: yarn global add gpt-console or npm i gpt-console -g After the installation, simply type gpt-console in any terminal to get started, AI agents are specialized modules built into GPTConsole to automate specific tasks. These agents add a layer of intelligence and convenience to your development workflow. Bird: Manages your Twitter activities such as tweets and replies. Pixie: Generates sophisticated ReactJS landing pages based on your prompts. Chip: An upcoming agent that will answer any code-related queries. GPTConsole isn't just another command-line tool. It's an intelligent, developer-friendly terminal that brings real-time collaboration to your fingertips.
    Starting Price: Freemium
  • 4
    NVIDIA Triton Inference Server
    NVIDIA Triton™ inference server delivers fast and scalable AI in production. Open-source inference serving software, Triton inference server streamlines AI inference by enabling teams deploy trained AI models from any framework (TensorFlow, NVIDIA TensorRT®, PyTorch, ONNX, XGBoost, Python, custom and more on any GPU- or CPU-based infrastructure (cloud, data center, or edge). Triton runs models concurrently on GPUs to maximize throughput and utilization, supports x86 and ARM CPU-based inferencing, and offers features like dynamic batching, model analyzer, model ensemble, and audio streaming. Triton helps developers deliver high-performance inference aTriton integrates with Kubernetes for orchestration and scaling, exports Prometheus metrics for monitoring, supports live model updates, and can be used in all major public cloud machine learning (ML) and managed Kubernetes platforms. Triton helps standardize model deployment in production.
    Starting Price: Free
  • 5
    Vertex AI Notebooks
    Vertex AI Notebooks is a fully managed, scalable solution from Google Cloud that accelerates machine learning (ML) development. It provides a seamless, interactive environment for data scientists and developers to explore data, prototype models, and collaborate in real-time. With integration into Google Cloud’s vast data and ML tools, Vertex AI Notebooks supports rapid prototyping, automated workflows, and deployment, making it easier to scale ML operations. The platform’s support for both Colab Enterprise and Vertex AI Workbench ensures a flexible and secure environment for diverse enterprise needs.
    Starting Price: $10 per GB
  • 6
    Vertex AI Vision
    Easily build, deploy, and manage computer vision applications with a fully managed, end-to-end application development environment that reduces the time to build computer vision applications from days to minutes at one-tenth the cost of current offerings. Quickly and conveniently ingest real-time video and image streams at a global scale. Easily build computer vision applications using a drag-and-drop interface. Store and search petabytes of data with built-in AI capabilities. Vertex AI Vision includes all the tools needed to manage the life cycle of computer vision applications, across ingestion, analysis, storage, and deployment. Easily connect application output to a data destination, like BigQuery for analytics, or live streaming to drive real-time business actions. Ingest thousands of video streams from across the globe. With a monthly pricing model, enjoy up to one-tenth lower costs than previous offerings.
    Starting Price: $0.0085 per GB
  • 7
    Athina AI

    Athina AI

    Athina AI

    Athina is a collaborative AI development platform that enables teams to build, test, and monitor AI applications efficiently. It offers features such as prompt management, evaluation tools, dataset handling, and observability, all designed to streamline the development of reliable AI systems. Athina supports integration with various models and services, including custom models, and ensures data privacy through fine-grained access controls and self-hosted deployment options. The platform is SOC-2 Type 2 compliant, providing a secure environment for AI development. Athina's user-friendly interface allows both technical and non-technical team members to collaborate effectively, accelerating the deployment of AI features.
    Starting Price: Free
  • 8
    OpenLIT

    OpenLIT

    OpenLIT

    OpenLIT is an OpenTelemetry-native application observability tool. It's designed to make the integration process of observability into AI projects with just a single line of code. Whether you're working with popular LLM libraries such as OpenAI and HuggingFace. OpenLIT's native support makes adding it to your projects feel effortless and intuitive. Analyze LLM and GPU performance, and costs to achieve maximum efficiency and scalability. Streams data to let you visualize your data and make quick decisions and modifications. Ensures that data is processed quickly without affecting the performance of your application. OpenLIT UI helps you explore LLM costs, token consumption, performance indicators, and user interactions in a straightforward interface. Connect to popular observability systems with ease, including Datadog and Grafana Cloud, to export data automatically. OpenLIT ensures your applications are monitored seamlessly.
    Starting Price: Free
  • 9
    Mistral Large

    Mistral Large

    Mistral AI

    Mistral Large is Mistral AI's flagship language model, designed for advanced text generation and complex multilingual reasoning tasks, including text comprehension, transformation, and code generation. It supports English, French, Spanish, German, and Italian, offering a nuanced understanding of grammar and cultural contexts. With a 32,000-token context window, it can accurately recall information from extensive documents. The model's precise instruction-following and native function-calling capabilities facilitate application development and tech stack modernization. Mistral Large is accessible through Mistral's platform, Azure AI Studio, and Azure Machine Learning, and can be self-deployed for sensitive use cases. Benchmark evaluations indicate that Mistral Large achieves strong results, making it the world's second-ranked model generally available through an API, next to GPT-4.
    Starting Price: Free
  • 10
    Aider

    Aider

    Aider AI

    Aider lets you pair program with LLMs, to edit code in your local git repository. Start a new project or work with an existing git repo. Aider works best with GPT-4o & Claude 3.5 Sonnet and can connect to almost any LLM. Aider has one of the top scores on SWE Bench. SWE Bench is a challenging software engineering benchmark where aider solved real GitHub issues from popular open source projects like django, scikitlearn, matplotlib, etc.
    Starting Price: Free
  • 11
    Imagen

    Imagen

    Google

    Imagen is a text-to-image generation model developed by Google Research. It uses advanced deep learning techniques, primarily leveraging large Transformer-based architectures, to generate high-quality, photorealistic images from natural language descriptions. Imagen's core innovation lies in combining the power of large language models (like those used in Google's NLP research) with the generative capabilities of diffusion models—a class of generative models known for creating images by progressively refining noise into detailed outputs. What sets Imagen apart is its ability to produce highly detailed and coherent images, often capturing fine-grained details and textures based on complex text prompts. It builds on the advancements in image generation made by models like DALL-E, but focuses heavily on semantic understanding and fine detail generation.
    Starting Price: Free
  • 12
    Restack

    Restack

    Restack

    A framework built specifically for the challenges of autonomous intelligence. Continue to write software using your language practices, libraries, APIs, data and models. Your proprietary autonomous product that adapts and scales with your development. Autonomous AI can automate video creation by generating, editing, and optimizing content, significantly reducing manual tasks in the production process. By integrating with AI tools like Luma AI or OpenAI for video generation, and scaling text-to-speech on Azure, your autonomous system can produce high-quality video content By integrating with platforms like YouTube your autonomous AI can continuously improve based on feedback and engagement metrics. We believe the most promising path to AGI is in the orchestration of millions of autonomous systems. We are a small group of passionate engineers and researchers dedicated to building autonomous artificial intelligence. If this sounds interesting to you, we would love to hear from you.
    Starting Price: $10 per month
  • 13
    Arize Phoenix
    Phoenix is an open-source observability library designed for experimentation, evaluation, and troubleshooting. It allows AI engineers and data scientists to quickly visualize their data, evaluate performance, track down issues, and export data to improve. Phoenix is built by Arize AI, the company behind the industry-leading AI observability platform, and a set of core contributors. Phoenix works with OpenTelemetry and OpenInference instrumentation. The main Phoenix package is arize-phoenix. We offer several helper packages for specific use cases. Our semantic layer is to add LLM telemetry to OpenTelemetry. Automatically instrumenting popular packages. Phoenix's open-source library supports tracing for AI applications, via manual instrumentation or through integrations with LlamaIndex, Langchain, OpenAI, and others. LLM tracing records the paths taken by requests as they propagate through multiple steps or components of an LLM application.
    Starting Price: Free
  • 14
    Lunary

    Lunary

    Lunary

    Lunary is an AI developer platform designed to help AI teams manage, improve, and protect Large Language Model (LLM) chatbots. It offers features such as conversation and feedback tracking, analytics on costs and performance, debugging tools, and a prompt directory for versioning and team collaboration. Lunary supports integration with various LLMs and frameworks, including OpenAI and LangChain, and provides SDKs for Python and JavaScript. Guardrails to deflect malicious prompts and sensitive data leaks. Deploy in your VPC with Kubernetes or Docker. Allow your team to judge responses from your LLMs. Understand what languages your users are speaking. Experiment with prompts and LLM models. Search and filter anything in milliseconds. Receive notifications when agents are not performing as expected. Lunary's core platform is 100% open-source. Self-host or in the cloud, get started in minutes.
    Starting Price: $20 per month
  • 15
    Dataplex Universal Catalog
    Dataplex Universal Catalog is Google Cloud’s intelligent governance platform for data and AI artifacts. It centralizes discovery, management, and monitoring across data lakes, warehouses, and databases, giving teams unified access to trusted data. With Vertex AI integration, users can instantly find datasets, models, features, and related assets in one search experience. It supports semantic search, data lineage, quality checks, and profiling to improve trust and compliance. Integrated with BigQuery and BigLake, it enables end-to-end governance for both proprietary and open lakehouse environments. Dataplex Universal Catalog helps organizations democratize data access, enforce governance, and accelerate analytics and AI initiatives.
    Starting Price: $0.060 per hour
  • 16
    MindMac

    MindMac

    MindMac

    MindMac is a native macOS application designed to enhance productivity by integrating seamlessly with ChatGPT and other AI models. It supports multiple AI providers, including OpenAI, Azure OpenAI, Google AI with Gemini, Google Cloud Vertex AI with Gemini, Anthropic Claude, OpenRouter, Mistral AI, Cohere, Perplexity, OctoAI, and local LLMs via LMStudio, LocalAI, GPT4All, Ollama, and llama.cpp. MindMac offers over 150 built-in prompt templates to facilitate user interaction and allows for extensive customization of OpenAI parameters, appearance, context modes, and keyboard shortcuts. The application features a powerful inline mode, enabling users to generate content or ask questions within any application without switching windows. MindMac ensures privacy by storing API keys securely in the Mac's Keychain and sending data directly to the AI provider without intermediary servers. The app is free to use with basic features, requiring no account for setup.
    Starting Price: $29 one-time payment
  • 17
    LiteLLM

    LiteLLM

    LiteLLM

    ​LiteLLM is a versatile platform designed to streamline interactions with over 100 Large Language Models (LLMs) through a unified interface. It offers both a Proxy Server (LLM Gateway) and a Python SDK, enabling developers to integrate various LLMs seamlessly into their applications. The Proxy Server facilitates centralized management, allowing for load balancing, cost tracking across projects, and consistent input/output formatting compatible with OpenAI standards. This setup supports multiple providers. It ensures robust observability by generating unique call IDs for each request, aiding in precise tracking and logging across systems. Developers can leverage pre-defined callbacks to log data using various tools. For enterprise users, LiteLLM offers advanced features like Single Sign-On (SSO), user management, and professional support through dedicated channels like Discord and Slack.
    Starting Price: Free
  • 18
    Agent Development Kit (ADK)
    The Agent Development Kit (ADK) is a flexible, open-source framework for building and deploying AI agents. It is tightly integrated with Google’s ecosystem, including Gemini models, and supports popular large language models (LLMs). ADK simplifies the development of both simple and complex AI agents, providing a structured environment for building dynamic workflows and multi-agent systems. With built-in tools for orchestration, deployment, and evaluation, ADK helps developers create scalable, modular AI solutions that can be easily deployed on platforms like Vertex AI or Cloud Run.
    Starting Price: Free
  • 19
    Mistral Medium 3
    Mistral Medium 3 is a powerful AI model designed to deliver state-of-the-art performance at a fraction of the cost compared to other models. It offers simpler deployment options, allowing for hybrid or on-premises configurations. Mistral Medium 3 excels in professional applications like coding and multimodal understanding, making it ideal for enterprise use. Its low-cost structure makes it highly accessible while maintaining top-tier performance, outperforming many larger models in specific domains.
    Starting Price: Free
  • 20
    Gemini CLI
    Gemini CLI is a free, open-source AI agent that integrates Gemini’s powerful AI capabilities directly into developers’ command line terminals. It offers fast, lightweight access to Gemini 2.5 Pro, enabling developers to generate code, solve problems, and manage tasks using natural language prompts. The CLI supports up to 60 model requests per minute and 1,000 requests per day at no cost, with additional paid options for professionals requiring higher usage. Gemini CLI includes advanced features like Google Search grounding for real-time web context, prompt customization, and automation within scripts. It is fully extensible and open source, welcoming community contributions via GitHub. Designed to enhance workflow efficiency, Gemini CLI brings AI-powered coding assistance to the terminal environment.
    Starting Price: Free
  • 21
    LLM Gateway

    LLM Gateway

    LLM Gateway

    LLM Gateway is a fully open source, unified API gateway that lets you route, manage, and analyze requests to any large language model provider, OpenAI, Anthropic, Google Vertex AI, and more, using a single, OpenAI-compatible endpoint. It offers multi-provider support with seamless migration and integration, dynamic model orchestration that routes each request to the optimal engine, and comprehensive usage analytics to track requests, token consumption, response times, and costs in real time. Built-in performance monitoring lets you compare models’ accuracy and cost-effectiveness, while secure key management centralizes API credentials under role-based controls. You can deploy LLM Gateway on your own infrastructure under the MIT license or use the hosted service as a progressive web app, and simple integration means you only need to change your API base URL, your existing code in any language or framework (cURL, Python, TypeScript, Go, etc.) continues to work without modification.
    Starting Price: $50 per month
  • 22
    TensorBlock

    TensorBlock

    TensorBlock

    TensorBlock is an open source AI infrastructure platform designed to democratize access to large language models through two complementary components. It has a self-hosted, privacy-first API gateway that unifies connections to any LLM provider under a single, OpenAI-compatible endpoint, with encrypted key management, dynamic model routing, usage analytics, and cost-optimized orchestration. TensorBlock Studio delivers a lightweight, developer-friendly multi-LLM interaction workspace featuring a plugin-based UI, extensible prompt workflows, real-time conversation history, and integrated natural-language APIs for seamless prompt engineering and model comparison. Built on a modular, scalable architecture and guided by principles of openness, composability, and fairness, TensorBlock enables organizations to experiment, deploy, and manage AI agents with full control and minimal infrastructure overhead.
    Starting Price: Free
  • 23
    Broxi AI

    Broxi AI

    Broxi AI

    Broxi AI is a no-code AI agent builder that lets you go from a simple text description to a fully functional AI agent in minutes using visual, drag-and-drop tools, requiring zero technical setup. Its Broxi Autopilot takes your natural‑language briefing, such as “an agent that answers FAQs from our PDF handbook”, then lets you define input types (PDFs, chats, websites, etc.) and outputs (emails, messages, API actions). With a single click, Broxi automatically constructs, tests in an interactive playground, and enables instant deployment of your agent via API, web widget, Slack, or embedded app. It supports integrations with dozens of tools and systems, offers real-time monitoring and centralized management, and maintains enterprise-grade security throughout, making it easy for non-technical teams to automate customer support, internal workflows, sales engagement, content generation, and data retrieval, without writing a line of code.
    Starting Price: $25 per month
  • 24
    Crush

    Crush

    Charm

    Crush is a glamorous AI coding agent that lives right in your terminal, seamlessly connecting your tools, code, and workflows with any Large Language Model (LLM) of your choice. It offers multi-model flexibility, letting you choose from a variety of LLMs or add your own using OpenAI or Anthropic-compatible APIs, and supports mid-session switching between them while preserving context. Crush is session-based, enabling multiple project-specific contexts to coexist. Powered by Language Server Protocol (LSP) enhancements, it incorporates coding-aware context just like a developer’s editor. It's highly extensible via Model Context Protocol (MCP) plugins using HTTP, stdio, or SSE for added capabilities. Crush runs anywhere, leveraging Charm’s sleek Bubble Tea-based TUI for a polished terminal user experience. Written in Go and MIT-licensed (with FSL-1.1 for trademarks), enabling developers to stay in their terminal while taking advantage of expressive AI coding assistance.
    Starting Price: Free
  • 25
    Gemini 2.5 Computer Use
    Introducing the Gemini 2.5 Computer Use model, a specialized agent model built on top of Gemini 2.5 Pro’s visual reasoning capabilities, designed to interact directly with user interfaces (UIs). It is exposed via a new computer-use tool in the Gemini API, with inputs that include the user’s request, a screenshot of the UI environment, and a history of recent actions. The model generates function calls corresponding to UI actions like clicking, typing, or selecting, and may request user confirmation for higher-risk tasks. After each action is executed, a new screenshot and URL are fed back into the model to continue the loop until the task completes or is halted. It is optimized primarily for web browser control and shows promise for mobile UI interaction, though it is not yet suited for desktop OS-level control. In benchmarks across web and mobile control tasks, Gemini 2.5 Computer Use outperforms leading alternatives, delivering high accuracy at lower latency.
    Starting Price: Free
  • 26
    Gemini Enterprise
    Gemini Enterprise is a comprehensive AI platform built by Google Cloud designed to bring the full power of Google’s advanced AI models, agent-creation tools, and enterprise-grade data access into everyday workflows. The solution offers a unified chat interface that lets employees interact with internal documents, applications, data sources, and custom AI agents. At its core, Gemini Enterprise comprises six key components: the Gemini family of large multimodal models, an agent orchestration workbench (formerly Google Agentspace), pre-built starter agents, robust data-integration connectors to business systems, extensive security and governance controls, and a partner ecosystem for tailored integrations. It is engineered to scale across departments and enterprises, enabling users to build no-code or low-code agents that automate tasks, such as research synthesis, customer support response, code assist, contract analysis, and more, while operating within corporate compliance standards.
    Starting Price: $21 per month
  • 27
    Claude Haiku 4.5
    Anthropic has launched Claude Haiku 4.5, its latest small-language model designed to deliver near-frontier performance at significantly lower cost. The model provides similar coding and reasoning quality as the company’s mid-tier Sonnet 4, yet it runs at roughly one-third of the cost and more than twice the speed. In benchmarks cited by Anthropic, Haiku 4.5 meets or exceeds Sonnet 4’s performance in key tasks such as code generation and multi-step “computer use” workflows. It is optimized for real-time, low-latency scenarios such as chat assistants, customer service agents, and pair-programming support. Haiku 4.5 is made available via the Claude API under the identifier “claude-haiku-4-5” and supports large-scale deployments where cost, responsiveness, and near-frontier intelligence matter. Claude Haiku 4.5 is available now on Claude Code and our apps. Its efficiency means you can accomplish more within your usage limits while maintaining premium model performance.
    Starting Price: $1 per million input tokens
  • 28
    Rebolt.ai

    Rebolt.ai

    Rebolt.ai

    Rebolt is an enterprise-grade AI platform that empowers companies to build custom applications and intelligent agents simply by speaking with AI. The platform integrates seamlessly with corporate tools such as OneDrive, SharePoint, Salesforce, Slack, and custom APIs, offering built-in infrastructure like databases, file storage, scheduling (cron jobs), audit logs, and distinct staging and production environments for deployment. Users can create apps and agents without writing API key code by simply describing what they need in natural language, while maintaining enterprise-class security, permissions mapping (e.g., via Azure groups), and role-based access. Rebolt is designed for building operational workflows, internal tools, and automation that connect to existing company data and services, enabling non-technical users (or low-code teams) to rapidly assemble solutions and replace spreadsheets, manual processes, and fragmented SaaS stacks.
    Starting Price: $25 per month
  • 29
    Google Cloud Text-to-Speech
    Convert text into natural-sounding speech using an API powered by Google’s AI technologies. Deploy Google’s groundbreaking technologies to generate speech with humanlike intonation. Built based on DeepMind’s speech synthesis expertise, the API delivers voices that are near human quality. Choose from a set of 220+ voices across 40+ languages and variants, including Mandarin, Hindi, Spanish, Arabic, Russian, and more. Pick the voice that works best for your user and application. Create a unique voice to represent your brand across all your customer touchpoints, instead of using a common voice shared with other organizations. Train a custom voice model using your own audio recordings to create a unique and more natural sounding voice for your organization. You can define and choose the voice profile that suits your organization and quickly adjust to changes in voice needs without needing to record new phrases.
  • 30
    Google Cloud AI Infrastructure
    Options for every business to train deep learning and machine learning models cost-effectively. AI accelerators for every use case, from low-cost inference to high-performance training. Simple to get started with a range of services for development and deployment. Tensor Processing Units (TPUs) are custom-built ASIC to train and execute deep neural networks. Train and run more powerful and accurate models cost-effectively with faster speed and scale. A range of NVIDIA GPUs to help with cost-effective inference or scale-up or scale-out training. Leverage RAPID and Spark with GPUs to execute deep learning. Run GPU workloads on Google Cloud where you have access to industry-leading storage, networking, and data analytics technologies. Access CPU platforms when you start a VM instance on Compute Engine. Compute Engine offers a range of both Intel and AMD processors for your VMs.