From dev environments to AI training, choose preset or custom VMs with 1–96 vCPUs and industry-leading 99.95% uptime SLA.
Compute Engine delivers high-performance virtual machines for web apps, databases, containers, and AI workloads. Choose from general-purpose, compute-optimized, or GPU/TPU-accelerated machine types—or build custom VMs to match your exact specs. With live migration and automatic failover, your workloads stay online. New customers get $300 in free credits.
Try Compute Engine
Easily Host LLMs and Web Apps on Cloud Run
Run everything from popular models with on-demand NVIDIA L4 GPUs to web apps without infrastructure management.
Run frontend and backend services, batch jobs, host LLMs, and queue processing workloads without the need to manage infrastructure. Cloud Run gives you on-demand GPU access for hosting LLMs and running real-time AI—with 5-second cold starts and automatic scale-to-zero so you only pay for actual usage. New customers get $300 in free credit to start.
Titan allows service-oriented processing of context recognition applications on Wireless Sensor Networks. It includes an execution environment developed for TinyOS and a Java ME (Personal Profile) Network Manager intended to be run on a mobile phone.
MobiSoC is a middleware that enables Mobile Social Computing Application development and provides a common platform for capturing, managing, and sharing the social state of physical communities.
Retrieve files from a remote computer using a PDA/Pocket PC running Windows Mobile.
Can access remote computers regardless the geographical position. Can work under WiFi, GPRS or any other wireless networks.
Access Google’s most capable multimodal models. Train, test, and deploy AI with 200+ foundation models on one platform.
Vertex AI gives developers access to Gemini 3—Google’s most advanced reasoning and coding model—plus 200+ foundation models including Claude, Llama, and Gemma. Build generative AI apps with Vertex AI Studio, customize with fine-tuning, and deploy to production with enterprise-grade MLOps. New customers get $300 in free credits.
CAMA (Context-Aware Mobile Agents) is a middleware for supporting mobile agents coordination. It is based on Linda-style coordination and provides numerous extentions for interoperability and fault-tolerance. It supports portable platforms such as PDAs.
The Sino Project is an Open Source initiative that utilized
code developed by the Chinese government. Sino Linux Middleware v3.0 and Mobile Business Application Suites are built and tested in conjunction with the OSDL in a high volume, enterprise read