Business Software for Vertex AI Notebooks

Top Software that integrates with Vertex AI Notebooks as of July 2025

Compare business software, products, and services to find the best solution for your business or organization. Use the filters on the left to drill down by category, pricing, features, organization size, organization type, region, user reviews, integrations, and more. View and sort the products and solutions that match your needs in the results below.

  • 1
    Google Cloud Platform
    Google Cloud is a cloud-based service that allows you to create anything from simple websites to complex applications for businesses of all sizes. New customers get $300 in free credits to run, test, and deploy workloads. All customers can use 25+ products for free, up to monthly usage limits. Use Google's core infrastructure, data analytics & machine learning. Secure and fully featured for all enterprises. Tap into big data to find answers faster and build better products. Grow from prototype to production to planet-scale, without having to think about capacity, reliability or performance. From virtual machines with proven price/performance advantages to a fully managed app development platform. Scalable, resilient, high performance object storage and databases for your applications. State-of-the-art software-defined networking products on Google’s private fiber network. Fully managed data warehousing, batch and stream processing, data exploration, Hadoop/Spark, and messaging.
    Leader badge
    Starting Price: Free ($300 in free credits)
    View Software
    Visit Website
  • 2
    Vertex AI
    Build, deploy, and scale machine learning (ML) models faster, with fully managed ML tools for any use case. Through Vertex AI Workbench, Vertex AI is natively integrated with BigQuery, Dataproc, and Spark. You can use BigQuery ML to create and execute machine learning models in BigQuery using standard SQL queries on existing business intelligence tools and spreadsheets, or you can export datasets from BigQuery directly into Vertex AI Workbench and run your models from there. Use Vertex Data Labeling to generate highly accurate labels for your data collection. Vertex AI Agent Builder enables developers to create and deploy enterprise-grade generative AI applications. It offers both no-code and code-first approaches, allowing users to build AI agents using natural language instructions or by leveraging frameworks like LangChain and LlamaIndex.
    Starting Price: Free ($300 in free credits)
    View Software
    Visit Website
  • 3
    Google Cloud BigQuery
    BigQuery is a serverless, multicloud data warehouse that simplifies the process of working with all types of data so you can focus on getting valuable business insights quickly. At the core of Google’s data cloud, BigQuery allows you to simplify data integration, cost effectively and securely scale analytics, share rich data experiences with built-in business intelligence, and train and deploy ML models with a simple SQL interface, helping to make your organization’s operations more data-driven. Gemini in BigQuery offers AI-driven tools for assistance and collaboration, such as code suggestions, visual data preparation, and smart recommendations designed to boost efficiency and reduce costs. BigQuery delivers an integrated platform featuring SQL, a notebook, and a natural language-based canvas interface, catering to data professionals with varying coding expertise. This unified workspace streamlines the entire analytics process.
    Starting Price: Free ($300 in free credits)
    View Software
    Visit Website
  • 4
    TensorFlow

    TensorFlow

    TensorFlow

    An end-to-end open source machine learning platform. TensorFlow is an end-to-end open source platform for machine learning. It has a comprehensive, flexible ecosystem of tools, libraries and community resources that lets researchers push the state-of-the-art in ML and developers easily build and deploy ML powered applications. Build and train ML models easily using intuitive high-level APIs like Keras with eager execution, which makes for immediate model iteration and easy debugging. Easily train and deploy models in the cloud, on-prem, in the browser, or on-device no matter what language you use. A simple and flexible architecture to take new ideas from concept to code, to state-of-the-art models, and to publication faster. Build, deploy, and experiment easily with TensorFlow.
    Starting Price: Free
  • 5
    1min.AI

    1min.AI

    1min.AI

    💡 1min.AI is an all-in-one AI app that unlock all AI features. You pay only for what you use at 1min.AI, with no hidden costs or setup required elsewhere. 🔮 The unique features of 1min.AI is offering a variety of AI features powered by various AI models. You can see it clearly with the Chat with Many Assistants feature, it includes Gemini, GPT, Claude, Llama, MistralAI, ... 🪄 Other multi-media features like Content, Image, Audio, Video can also be used with different models to utilize their abilities and give out the best results. 💰 Lastly, we offer credit estimation and transparent usage history, so you know exact how does the feature cost before running and can track the usage easily. 🚀 Try for Free and get what you want within 1min
    Leader badge
    Starting Price: $5
  • 6
    Jupyter Notebook

    Jupyter Notebook

    Project Jupyter

    The Jupyter Notebook is an open-source web application that allows you to create and share documents that contain live code, equations, visualizations and narrative text. Uses include: data cleaning and transformation, numerical simulation, statistical modeling, data visualization, machine learning, and much more.
  • 7
    Gemini Flash
    Gemini Flash is an advanced large language model (LLM) from Google, specifically designed for high-speed, low-latency language processing tasks. Part of Google DeepMind’s Gemini series, Gemini Flash is tailored to provide real-time responses and handle large-scale applications, making it ideal for interactive AI-driven experiences such as customer support, virtual assistants, and live chat solutions. Despite its speed, Gemini Flash doesn’t compromise on quality; it’s built on sophisticated neural architectures that ensure responses remain contextually relevant, coherent, and precise. Google has incorporated rigorous ethical frameworks and responsible AI practices into Gemini Flash, equipping it with guardrails to manage and mitigate biased outputs, ensuring it aligns with Google’s standards for safe and inclusive AI. With Gemini Flash, Google empowers businesses and developers to deploy responsive, intelligent language tools that can meet the demands of fast-paced environments.
  • 8
    Gemini 2.5 Pro
    Gemini 2.5 Pro is an advanced AI model designed to handle complex tasks with enhanced reasoning and coding capabilities. Leading common benchmarks, it excels in math, science, and coding, demonstrating strong performance in tasks like web app creation and code transformation. Built on the Gemini 2.5 foundation, it features a 1 million token context window, enabling it to process vast datasets from various sources such as text, images, and code repositories. Available now in Google AI Studio, Gemini 2.5 Pro is optimized for more sophisticated applications and supports advanced users with improved performance for complex problem-solving.
    Starting Price: $19.99/month
  • 9
    PyTorch

    PyTorch

    PyTorch

    Transition seamlessly between eager and graph modes with TorchScript, and accelerate the path to production with TorchServe. Scalable distributed training and performance optimization in research and production is enabled by the torch-distributed backend. A rich ecosystem of tools and libraries extends PyTorch and supports development in computer vision, NLP and more. PyTorch is well supported on major cloud platforms, providing frictionless development and easy scaling. Select your preferences and run the install command. Stable represents the most currently tested and supported version of PyTorch. This should be suitable for many users. Preview is available if you want the latest, not fully tested and supported, 1.10 builds that are generated nightly. Please ensure that you have met the prerequisites (e.g., numpy), depending on your package manager. Anaconda is our recommended package manager since it installs all dependencies.
  • 10
    ZenML

    ZenML

    ZenML

    Simplify your MLOps pipelines. Manage, deploy, and scale on any infrastructure with ZenML. ZenML is completely free and open-source. See the magic with just two simple commands. Set up ZenML in a matter of minutes, and start with all the tools you already use. ZenML standard interfaces ensure that your tools work together seamlessly. Gradually scale up your MLOps stack by switching out components whenever your training or deployment requirements change. Keep up with the latest changes in the MLOps world and easily integrate any new developments. Define simple and clear ML workflows without wasting time on boilerplate tooling or infrastructure code. Write portable ML code and switch from experimentation to production in seconds. Manage all your favorite MLOps tools in one place with ZenML's plug-and-play integrations. Prevent vendor lock-in by writing extensible, tooling-agnostic, and infrastructure-agnostic code.
    Starting Price: Free
  • 11
    Cline

    Cline

    Cline AI Coding Agent

    Autonomous coding agent right in your IDE, capable of creating/editing files, running commands, using the browser, and more with your permission every step of the way. Cline can handle complex software development tasks step-by-step. With tools that let him create & edit files, explore large projects, use the browser, and execute terminal commands (after you grant permission), he can assist you in ways that go beyond code completion or tech support. While autonomous AI scripts traditionally run in sandboxed environments, this extension provides a human-in-the-loop GUI to approve every file change and terminal command, providing a safe and accessible way to explore the potential of agentic AI.
    Starting Price: Free
  • 12
    SmythOS

    SmythOS

    SmythOS

    Say goodbye to manual coding and build agents faster than ever. Describe what you need, and SmythOS builds it from your chat or image, using the best AI models and APIs for your task. Use any AI model or API. Integrate with OpenAI, Hugging Face, Amazon Bedrock, and hundreds of vendors without a line of code. A pre-built agent template library gives you agents that already work out of the box for dozens of use cases. Just hit the button and connect with your own API keys. Because your marketing team should not have access to agents that work with your code. We got you covered. Create a space for each client, team, and project with full user and permission management. Deploy on-prem or to AWS. Integrate with Bedrock, Vertex, Adobe, Salesforce, etc. Explainable AI with full control over data flows, audit logs, encryption, and auth. Chat with your agents, give them bulk work, inspect their work logs, assign them work schedules, and more.
    Starting Price: $30 per month
  • 13
    MindMac

    MindMac

    MindMac

    MindMac is a native macOS application designed to enhance productivity by integrating seamlessly with ChatGPT and other AI models. It supports multiple AI providers, including OpenAI, Azure OpenAI, Google AI with Gemini, Google Cloud Vertex AI with Gemini, Anthropic Claude, OpenRouter, Mistral AI, Cohere, Perplexity, OctoAI, and local LLMs via LMStudio, LocalAI, GPT4All, Ollama, and llama.cpp. MindMac offers over 150 built-in prompt templates to facilitate user interaction and allows for extensive customization of OpenAI parameters, appearance, context modes, and keyboard shortcuts. The application features a powerful inline mode, enabling users to generate content or ask questions within any application without switching windows. MindMac ensures privacy by storing API keys securely in the Mac's Keychain and sending data directly to the AI provider without intermediary servers. The app is free to use with basic features, requiring no account for setup.
    Starting Price: $29 one-time payment
  • 14
    Mistral Large 2
    Mistral AI has launched the Mistral Large 2, an advanced AI model designed to excel in code generation, multilingual capabilities, and complex reasoning tasks. The model features a 128k context window, supporting dozens of languages including English, French, Spanish, and Arabic, as well as over 80 programming languages. Mistral Large 2 is tailored for high-throughput single-node inference, making it ideal for large-context applications. Its improved performance on benchmarks like MMLU and its enhanced code generation and reasoning abilities ensure accuracy and efficiency. The model also incorporates better function calling and retrieval, supporting complex business applications.
    Starting Price: Free
  • 15
    Mistral Small 3.1
    ​Mistral Small 3.1 is a state-of-the-art, multimodal, and multilingual AI model released under the Apache 2.0 license. Building upon Mistral Small 3, this enhanced version offers improved text performance, and advanced multimodal understanding, and supports an expanded context window of up to 128,000 tokens. It outperforms comparable models like Gemma 3 and GPT-4o Mini, delivering inference speeds of 150 tokens per second. Designed for versatility, Mistral Small 3.1 excels in tasks such as instruction following, conversational assistance, image understanding, and function calling, making it suitable for both enterprise and consumer-grade AI applications. Its lightweight architecture allows it to run efficiently on a single RTX 4090 or a Mac with 32GB RAM, facilitating on-device deployments. It is available for download on Hugging Face, accessible via Mistral AI's developer playground, and integrated into platforms like Google Cloud Vertex AI, with availability on NVIDIA NIM and
    Starting Price: Free
  • 16
    Google Cloud Dataproc
    Dataproc makes open source data and analytics processing fast, easy, and more secure in the cloud. Build custom OSS clusters on custom machines faster. Whether you need extra memory for Presto or GPUs for Apache Spark machine learning, Dataproc can help accelerate your data and analytics processing by spinning up a purpose-built cluster in 90 seconds. Easy and affordable cluster management. With autoscaling, idle cluster deletion, per-second pricing, and more, Dataproc can help reduce the total cost of ownership of OSS so you can focus your time and resources elsewhere. Security built in by default. Encryption by default helps ensure no piece of data is unprotected. With JobsAPI and Component Gateway, you can define permissions for Cloud IAM clusters, without having to set up networking or gateway nodes.
  • 17
    Kong AI Gateway
    ​Kong AI Gateway is a semantic AI gateway designed to run and secure Large Language Model (LLM) traffic, enabling faster adoption of Generative AI (GenAI) through new semantic AI plugins for Kong Gateway. It allows users to easily integrate, secure, and monitor popular LLMs. The gateway enhances AI requests with semantic caching and security features, introducing advanced prompt engineering for compliance and governance. Developers can power existing AI applications written using SDKs or AI frameworks by simply changing one line of code, simplifying migration. Kong AI Gateway also offers no-code AI integrations, allowing users to transform, enrich, and augment API responses without writing code, using declarative configuration. It implements advanced prompt security by determining allowed behaviors and enables the creation of better prompts with AI templates compatible with the OpenAI interface.
  • 18
    Gemini 2.5 Pro Preview (I/O Edition)
    Gemini 2.5 Pro Preview (I/O Edition) by Google is an advanced AI model designed to streamline coding tasks and enhance web app development. This powerful tool allows developers to efficiently transform and edit code, reducing errors and improving function calling accuracy. With enhanced capabilities in video understanding and web app creation, Gemini 2.5 Pro Preview excels at building aesthetically pleasing and functional web applications. Available through Google’s Gemini API and AI platforms, this model provides a seamless solution for developers to create innovative applications with improved performance and reliability.
    Starting Price: $19.99/month
  • 19
    Apache Spark

    Apache Spark

    Apache Software Foundation

    Apache Spark™ is a unified analytics engine for large-scale data processing. Apache Spark achieves high performance for both batch and streaming data, using a state-of-the-art DAG scheduler, a query optimizer, and a physical execution engine. Spark offers over 80 high-level operators that make it easy to build parallel apps. And you can use it interactively from the Scala, Python, R, and SQL shells. Spark powers a stack of libraries including SQL and DataFrames, MLlib for machine learning, GraphX, and Spark Streaming. You can combine these libraries seamlessly in the same application. Spark runs on Hadoop, Apache Mesos, Kubernetes, standalone, or in the cloud. It can access diverse data sources. You can run Spark using its standalone cluster mode, on EC2, on Hadoop YARN, on Mesos, or on Kubernetes. Access data in HDFS, Alluxio, Apache Cassandra, Apache HBase, Apache Hive, and hundreds of other data sources.
  • 20
    Azure Data Lake
    Azure Data Lake includes all the capabilities required to make it easy for developers, data scientists, and analysts to store data of any size, shape, and speed, and do all types of processing and analytics across platforms and languages. It removes the complexities of ingesting and storing all of your data while making it faster to get up and running with batch, streaming, and interactive analytics. Azure Data Lake works with existing IT investments for identity, management, and security for simplified data management and governance. It also integrates seamlessly with operational stores and data warehouses so you can extend current data applications. We’ve drawn on the experience of working with enterprise customers and running some of the largest scale processing and analytics in the world for Microsoft businesses like Office 365, Xbox Live, Azure, Windows, Bing, and Skype. Azure Data Lake solves many of the productivity and scalability challenges that prevent you from maximizing the
  • 21
    Kubeflow

    Kubeflow

    Kubeflow

    The Kubeflow project is dedicated to making deployments of machine learning (ML) workflows on Kubernetes simple, portable and scalable. Our goal is not to recreate other services, but to provide a straightforward way to deploy best-of-breed open-source systems for ML to diverse infrastructures. Anywhere you are running Kubernetes, you should be able to run Kubeflow. Kubeflow provides a custom TensorFlow training job operator that you can use to train your ML model. In particular, Kubeflow's job operator can handle distributed TensorFlow training jobs. Configure the training controller to use CPUs or GPUs and to suit various cluster sizes. Kubeflow includes services to create and manage interactive Jupyter notebooks. You can customize your notebook deployment and your compute resources to suit your data science needs. Experiment with your workflows locally, then deploy them to a cloud when you're ready.
  • 22
    SQL

    SQL

    SQL

    SQL is a domain-specific programming language used for accessing, managing, and manipulating relational databases and relational database management systems.
  • Previous
  • You're on page 1
  • Next