Best Artificial Intelligence Software for Vertex AI - Page 4

Compare the Top Artificial Intelligence Software that integrates with Vertex AI as of November 2025 - Page 4

This a list of Artificial Intelligence software that integrates with Vertex AI. Use the filters on the left to add additional filters for products that have integrations with Vertex AI. View the products that work with Vertex AI in the table below.

  • 1
    Veo 3

    Veo 3

    Google

    Veo 3 is Google’s latest state-of-the-art video generation model, designed to bring greater realism and creative control to filmmakers and storytellers. With the ability to generate videos in 4K resolution and enhanced with real-world physics and audio, Veo 3 allows creators to craft high-quality video content with unmatched precision. The model’s improved prompt adherence ensures more accurate and consistent responses to user instructions, making the video creation process more intuitive. It also introduces new features that give creators more control over characters, scenes, and transitions, enabling seamless integration of different elements to create dynamic, engaging videos.
  • 2
    Lyria 2

    Lyria 2

    Google

    Lyria 2 is an advanced AI music generation model developed by Google, designed to help musicians compose high-fidelity music across a wide variety of genres and styles. The model generates professional-grade 48kHz stereo audio, capturing intricate details and nuances in different instruments and playing styles. With granular creative control, musicians can use text prompts to shape compositions, adjusting elements like key, BPM, and other characteristics to match their artistic vision. Lyria 2 accelerates the creative process by providing new starting points, suggesting harmonies, and drafting longer arrangements, helping musicians overcome writer's block and explore new creative possibilities.
  • 3
    WeatherNext

    WeatherNext

    Google DeepMind

    WeatherNext is a family of AI models from Google DeepMind and Google Research that produces state-of-the-art weather forecasts. These models are faster and more efficient than traditional physics-based weather models and yield superior forecast reliability. The gains in forecast performance could enable better preparation to help save lives in the face of extreme weather events and enhance the reliability of sustainable energy and supply chains. WeatherNext Graph offers more accurate and efficient deterministic forecasts compared to the best deterministic systems in use today, providing a single weather forecast per time and location with a temporal resolution of 6 hours and a lead time of 10 days. WeatherNext Gen accurately generates an ensemble forecast, better than the current ensemble models most widely used today, helping decision-makers better understand weather uncertainties and risks of extreme conditions.
  • 4
    MedGemma

    MedGemma

    Google DeepMind

    MedGemma is a collection of Gemma 3 variants that are trained for performance on medical text and image comprehension. Developers can use MedGemma to accelerate building healthcare-based AI applications. MedGemma currently comes in two variants: a 4B multimodal version and a 27B text-only version. MedGemma 4B utilizes a SigLIP image encoder that has been specifically pre-trained on a variety of de-identified medical data, including chest X-rays, dermatology images, ophthalmology images, and histopathology slides. Its LLM component is trained on a diverse set of medical data, including radiology images, histopathology patches, ophthalmology images, and dermatology images. MedGemma 4B is available in both pre-trained (suffix: -pt) and instruction-tuned (suffix -it) versions. The instruction-tuned version is a better starting point for most applications.
  • 5
    Gemini 2.5 Flash-Lite
    Gemini 2.5 is Google DeepMind’s latest generation AI model family, designed to deliver advanced reasoning and native multimodality with a long context window. It improves performance and accuracy by reasoning through its thoughts before responding. The model offers different versions tailored for complex coding tasks, fast everyday performance, and cost-efficient high-volume workloads. Gemini 2.5 supports multiple data types including text, images, video, audio, and PDFs, enabling versatile AI applications. It features adaptive thinking budgets and fine-grained control for developers to balance cost and output quality. Available via Google AI Studio and Gemini API, Gemini 2.5 powers next-generation AI experiences.
  • 6
    Claude Opus 4.1
    Claude Opus 4.1 is an incremental upgrade to Claude Opus 4 that boosts coding, agentic reasoning, and data-analysis performance without changing deployment complexity. It raises coding accuracy to 74.5 percent on SWE-bench Verified and sharpens in-depth research and detailed tracking for agentic search tasks. GitHub reports notable gains in multi-file code refactoring, while Rakuten Group highlights its precision in pinpointing exact corrections within large codebases without introducing bugs. Independent benchmarks show about a one-standard-deviation improvement on junior developer tests compared to Opus 4, mirroring major leaps seen in prior Claude releases. Opus 4.1 is available now to paid Claude users, in Claude Code, and via the Anthropic API (model ID claude-opus-4-1-20250805), as well as through Amazon Bedrock and Google Cloud Vertex AI, and integrates seamlessly into existing workflows with no additional setup beyond selecting the new model.
  • 7
    Gemini 3.0 Pro
    Gemini 3.0 is Google’s upcoming next-generation AI model expected to launch in late 2025, promising unprecedented intelligence with the ability to think, plan, and act autonomously. It features chain-of-thought reasoning, a massive 1 million+ token context window, and built-in multimodal capabilities for text, images, audio, and video. Powered by Google’s TPU v5p hardware, Gemini 3.0 aims for lightning-fast, real-time AI responses with enhanced safety and alignment. While waiting for Gemini 3.0, users can access today’s top AI models like GPT-4o, Claude 4, and Gemini 2.5 Pro through the Fello AI Mac app. Fello AI offers native Mac integration, offline chat history, and seamless switching between multiple AI engines. This makes it a future-proof platform to build AI workflows and be ready for Gemini 3.0’s revolutionary capabilities.
    Starting Price: $19.99/month
  • 8
    Nano Banana
    Nano Banana is the internal codename for Google’s powerful AI image-editing model, officially known as Gemini 2.5 Flash Image. It enables users to transform and refine images using conversational prompts—whether that means colorizing old photos, changing outfits, or merging multiple images into one seamless composition. One of its standout features is preserving character consistency: people, pets, and objects remain recognizable even after multiple edits. It excels at multi-turn editing, meaning you can iteratively tweak the same image and expect coherent results throughout. Nano Banana is integrated into the Gemini app, making these capabilities accessible to both free and paid users. Despite its impressive performance, the images include visible and invisible (SynthID) watermarking to denote AI generation for transparency.
  • 9
    Claude Sonnet 4.5
    Claude Sonnet 4.5 is Anthropic’s latest frontier model, designed to excel in long-horizon coding, agentic workflows, and intensive computer use while maintaining safety and alignment. It achieves state-of-the-art performance on the SWE-bench Verified benchmark (for software engineering) and leads on OSWorld (a computer use benchmark), with the ability to sustain focus over 30 hours on complex, multi-step tasks. The model introduces improvements in tool handling, memory management, and context processing, enabling more sophisticated reasoning, better domain understanding (from finance and law to STEM), and deeper code comprehension. It supports context editing and memory tools to sustain long conversations or multi-agent tasks, and allows code execution and file creation within Claude apps. Sonnet 4.5 is deployed at AI Safety Level 3 (ASL-3), with classifiers protecting against inputs or outputs tied to risky domains, and includes mitigations against prompt injection.
  • 10
    Veo 3.1

    Veo 3.1

    Google

    Veo 3.1 builds on the capabilities of the previous model to enable longer and more versatile AI-generated videos. With this version, users can create multi-shot clips guided by multiple prompts, generate sequences from three reference images, and use frames in video workflows that transition between a start and end image, both with native, synchronized audio. The scene extension feature allows extension of a final second of a clip by up to a full minute of newly generated visuals and sound. Veo 3.1 supports editing of lighting and shadow parameters to improve realism and scene consistency, and offers advanced object removal that reconstructs backgrounds to remove unwanted items from generated footage. These enhancements make Veo 3.1 sharper in prompt-adherence, more cinematic in presentation, and broader in scale compared to shorter-clip models. Developers can access Veo 3.1 via the Gemini API or through the tool Flow, targeting professional video workflows.
  • 11
    SynthID

    SynthID

    Google

    We’re beta launching SynthID, a tool for watermarking and identifying AI-generated images. SynthID is being released to a limited number of Vertex AI customers using Imagen, one of our latest text-to-image models that uses input text to create photorealistic images. With this tool, users can embed an imperceptible digital watermark into their AI-generated images and identify if Imagen was used for generating the image, or even part of the image. Being able to identify AI-generated content is critical to promoting trust in information. While not a silver bullet for addressing the problem of misinformation, SynthID is an early and promising technical solution to this pressing AI safety issue. This technology was developed by Google DeepMind and refined in partnership with Google Research. SynthID could be expanded for use across other AI models and we plan to integrate it into more products in the near future.
  • 12
    Tune AI

    Tune AI

    NimbleBox

    Leverage the power of custom models to build your competitive advantage. With our enterprise Gen AI stack, go beyond your imagination and offload manual tasks to powerful assistants instantly – the sky is the limit. For enterprises where data security is paramount, fine-tune and deploy generative AI models on your own cloud, securely.
  • 13
    Imagen 3

    Imagen 3

    Google

    Imagen 3 is the next evolution of Google's cutting-edge text-to-image AI generation technology. Building on the strengths of its predecessors, Imagen 3 offers significant advancements in image fidelity, resolution, and semantic alignment with user prompts. By employing enhanced diffusion models and more sophisticated natural language understanding, it can produce hyper-realistic, high-resolution images with intricate textures, vivid colors, and precise object interactions. Imagen 3 also introduces better handling of complex prompts, including abstract concepts and multi-object scenes, while reducing artifacts and improving coherence. With its powerful capabilities, Imagen 3 is poised to revolutionize creative industries, from advertising and design to gaming and entertainment, by providing artists, developers, and creators with an intuitive tool for visual storytelling and ideation.
  • 14
    Chirp 3

    Chirp 3

    Google

    ​Google Cloud's Text-to-Speech API introduces Chirp 3, enabling users to create personalized voice models using their own high-quality audio recordings. This feature facilitates the rapid generation of custom voices, which can be utilized to synthesize audio through the Cloud Text-to-Speech API, supporting both streaming and long-form text. Access to this voice cloning capability is restricted to allow-listed users due to safety considerations; interested parties should contact the sales team to be added to the allowed list. Instant Custom Voice creation and synthesis are supported in various languages, including English (US), Spanish (US), and French (Canada), among others. It is available in multiple Google Cloud regions, and supported output formats include LINEAR16, OGG_OPUS, PCM, ALAW, MULAW, and MP3, depending on the API method used.
  • 15
    Lyria

    Lyria

    Google

    Lyria, introduced on Vertex AI, is a powerful text-to-music model designed to generate high-fidelity, custom soundtracks based on written descriptions. Ideal for businesses in marketing, content creation, and entertainment, Lyria enables users to quickly produce music that aligns with their brand identity, video content, or marketing campaigns. It offers a cost-effective and time-efficient solution for creating original, royalty-free music that captures the desired mood, tone, and narrative, accelerating production workflows and enhancing brand experiences.
  • 16
    Imagen 4

    Imagen 4

    Google

    Imagen 4 is Google's most advanced image generation model, designed for creativity and photorealism. With improved clarity, sharper image details, and better typography, it allows users to bring their ideas to life faster and more accurately than ever before. It supports photo-realistic generation of landscapes, animals, and people, and offers a diverse range of artistic styles, from abstract to illustration. The new features also include ultra-fast processing, enhanced color rendering, and a mode for up to 10x faster image creation. Imagen 4 can generate images at up to 2K resolution, providing exceptional clarity and detail, making it ideal for both artistic and practical applications.
  • 17
    Veo 3.1 Fast
    Veo 3.1 Fast is Google’s upgraded video-generation model, released in paid preview within the Gemini API alongside Veo 3.1. It enables developers to create cinematic, high-quality videos from text prompts or reference images at a much faster processing speed. The model introduces native audio generation with natural dialogue, ambient sound, and synchronized effects for lifelike storytelling. Veo 3.1 Fast also supports advanced controls such as “Ingredients to Video,” allowing up to three reference images, “Scene Extension” for longer sequences, and “First and Last Frame” transitions for seamless shot continuity. Built for efficiency and realism, it delivers improved image-to-video quality and character consistency across multiple scenes. With direct integration into Google AI Studio and Vertex AI, Veo 3.1 Fast empowers developers to bring creative video concepts to life in record time.