Vertex AI
Build, deploy, and scale machine learning (ML) models faster, with fully managed ML tools for any use case.
Through Vertex AI Workbench, Vertex AI is natively integrated with BigQuery, Dataproc, and Spark. You can use BigQuery ML to create and execute machine learning models in BigQuery using standard SQL queries on existing business intelligence tools and spreadsheets, or you can export datasets from BigQuery directly into Vertex AI Workbench and run your models from there. Use Vertex Data Labeling to generate highly accurate labels for your data collection.
Vertex AI Agent Builder enables developers to create and deploy enterprise-grade generative AI applications. It offers both no-code and code-first approaches, allowing users to build AI agents using natural language instructions or by leveraging frameworks like LangChain and LlamaIndex.
Learn more
Ango Hub
Ango Hub is a quality-focused, enterprise-ready data annotation platform for AI teams, available on cloud and on-premise. It supports computer vision, medical imaging, NLP, audio, video, and 3D point cloud annotation, powering use cases from autonomous driving and robotics to healthcare AI.
Built for AI fine-tuning, RLHF, LLM evaluation, and human-in-the-loop workflows, Ango Hub boosts throughput with automation, model-assisted pre-labeling, and customizable QA while maintaining accuracy. Features include centralized instructions, review pipelines, issue tracking, and consensus across up to 30 annotators. With nearly twenty labeling tools—such as rotated bounding boxes, label relations, nested conditional questions, and table-based labeling—it supports both simple and complex projects. It also enables annotation pipelines for chain-of-thought reasoning and next-gen LLM training and enterprise-grade security with HIPAA compliance, SOC 2 certification, and role-based access controls.
Learn more
Olmo 3
Olmo 3 is a fully open model family spanning 7 billion and 32 billion parameter variants that delivers not only high-performing base, reasoning, instruction, and reinforcement-learning models, but also exposure of the entire model flow, including raw training data, intermediate checkpoints, training code, long-context support (65,536 token window), and provenance tooling. Starting with the Dolma 3 dataset (≈9 trillion tokens) and its disciplined mix of web text, scientific PDFs, code, and long-form documents, the pre-training, mid-training, and long-context phases shape the base models, which are then post-trained via supervised fine-tuning, direct preference optimisation, and RL with verifiable rewards to yield the Think and Instruct variants. The 32 B Think model is described as the strongest fully open reasoning model to date, competitively close to closed-weight peers in math, code, and complex reasoning.
Learn more