Go from idea to deployed AI app without managing infrastructure. Vertex AI offers one platform for the entire AI development lifecycle.
Ship AI apps and features faster with Vertex AI—your end-to-end AI platform. Access Gemini 3 and 200+ foundation models, fine-tune for your needs, and deploy with enterprise-grade MLOps. Build chatbots, agents, or custom models. New customers get $300 in free credit.
Try Vertex AI Free
Go from Data Warehouse to Data and AI platform with BigQuery
Build, train, and run ML models with simple SQL. Automate data prep, analysis, and predictions with built-in AI assistance from Gemini.
BigQuery is more than a data warehouse—it's an autonomous data-to-AI platform. Use familiar SQL to train ML models, run time-series forecasts, and generate AI-powered insights with native Gemini integration. Built-in agents handle data engineering and data science workflows automatically. Get $300 in free credit, query 1 TB, and store 10 GB free monthly.
Reverse engineered communication protocol for the Mala ProEx and Easy Locator GPR implemented as a CLI & GUI running on unix systems. The communication protocol was ported to a graphical user interface applicable to embedded systems through the use of the Qt toolkit (tested on Armadeus Systems APF27 board). The GUI includes GPS datalogging capability assuming a receiver providing NMEA sentences at 4800 bauds is connected to the host computer.
Allows viewing and editing of weather data from
weather models as two dimensional grids. Python scripting support hooks in the ingest, editing and product creation subsystems allow for customization. All network communication made through HTTP services.
Enhanced version of the system used by national weather services.
The latest ESMF development is happening in GitHub:
https://github.com/esmf-org/esmf
https://earthsystemmodeling.org
The Earth System Modeling Framework provides high-performance software infrastructure and superstructure for the construction and coupling of climate, weather, and data assimilation applications.
Run everything from popular models with on-demand NVIDIA L4 GPUs to web apps without infrastructure management.
Run frontend and backend services, batch jobs, host LLMs, and queue processing workloads without the need to manage infrastructure. Cloud Run gives you on-demand GPU access for hosting LLMs and running real-time AI—with 5-second cold starts and automatic scale-to-zero so you only pay for actual usage. New customers get $300 in free credit to start.