Go from Data Warehouse to Data and AI platform with BigQuery
Build, train, and run ML models with simple SQL. Automate data prep, analysis, and predictions with built-in AI assistance from Gemini.
BigQuery is more than a data warehouse—it's an autonomous data-to-AI platform. Use familiar SQL to train ML models, run time-series forecasts, and generate AI-powered insights with native Gemini integration. Built-in agents handle data engineering and data science workflows automatically. Get $300 in free credit, query 1 TB, and store 10 GB free monthly.
Try BigQuery Free
Build AI Apps with Gemini 3 on Vertex AI
Access Google’s most capable multimodal models. Train, test, and deploy AI with 200+ foundation models on one platform.
Vertex AI gives developers access to Gemini 3—Google’s most advanced reasoning and coding model—plus 200+ foundation models including Claude, Llama, and Gemma. Build generative AI apps with Vertex AI Studio, customize with fine-tuning, and deploy to production with enterprise-grade MLOps. New customers get $300 in free credits.
Win32 hosted tools to cross compile to DJGPP v1 (go32) for MS-DOS. Now includes GCC 1.27 aka the first version available to support the i386 processor.
Catalina is a C compiler plus a set of C libraries and device drivers for use with the Parallax Propeller microcontroller. Catalina is a cross-compiler based on the retargetable C compiler "lcc". Catalina runs on Windows or Linux.
RELIC (Retargetable Embedded Language Independent Compiler) is a compiler tool chain with the goal of complete language and processor independence for embedded microcontrollers and microprocessors.
Modern command-line cross-compiler of 8bit processor Z80(and R800) assembly language. Support many directives(even ZX-Spectrum specific as snapshot .SNA generation), has built-in Lua scripting engine, 3-pass design. Windows/DOS/FreeBSD.
Run everything from popular models with on-demand NVIDIA L4 GPUs to web apps without infrastructure management.
Run frontend and backend services, batch jobs, host LLMs, and queue processing workloads without the need to manage infrastructure. Cloud Run gives you on-demand GPU access for hosting LLMs and running real-time AI—with 5-second cold starts and automatic scale-to-zero so you only pay for actual usage. New customers get $300 in free credit to start.