From dev environments to AI training, choose preset or custom VMs with 1–96 vCPUs and industry-leading 99.95% uptime SLA.
Compute Engine delivers high-performance virtual machines for web apps, databases, containers, and AI workloads. Choose from general-purpose, compute-optimized, or GPU/TPU-accelerated machine types—or build custom VMs to match your exact specs. With live migration and automatic failover, your workloads stay online. New customers get $300 in free credits.
Try Compute Engine
Ship AI Apps Faster with Vertex AI
Go from idea to deployed AI app without managing infrastructure. Vertex AI offers one platform for the entire AI development lifecycle.
Ship AI apps and features faster with Vertex AI—your end-to-end AI platform. Access Gemini 3 and 200+ foundation models, fine-tune for your needs, and deploy with enterprise-grade MLOps. Build chatbots, agents, or custom models. New customers get $300 in free credit.
AI Researcher is an experimental open-source project that demonstrates how multiple AI agents can collaborate to conduct complex research tasks from start to finish with minimal human intervention. It orchestrates agents that can generate research questions, perform literature reviews, execute experiments, analyze results, and synthesize findings into structured outputs like reports or code.
The simplest, fastest repository for training/finetuning models
NanoGPT is a minimalistic yet powerful reimplementation of GPT-style transformers created by Andrej Karpathy for educational and research use. It distills the GPT architecture into a few hundred lines of Python code, making it far easier to understand than large, production-scale implementations. The repo is organized with a training pipeline (dataset preprocessing, model definition, optimizer, training loop) and inference script so you can train a small GPT on text datasets like Shakespeare...
Experiments and code from Google Brain’s Tokyo research workshop
The Brain Tokyo Workshop repository hosts a collection of research materials and experimental code developed by the Google Brain team based in Tokyo. It showcases a variety of cutting-edge projects in artificial intelligence, particularly in the areas of neuroevolution, reinforcement learning, and model interpretability. Each project explores innovative approaches to learning, prediction, and creativity in neural networks, often through unconventional or biologically inspired methods. The...
Catalyst is a PyTorch framework for accelerated Deep Learning research and development. It allows you to write compact but full-featured Deep Learning pipelines with just a few lines of code. With Catalyst you get a full set of features including a training loop with metrics, model checkpointing and more, all without the boilerplate. Catalyst is focused on reproducibility, rapid experimentation, and codebase reuse so you can break the cycle of writing another regular train loop and make...
Host and run your applications without the need to manage infrastructure. Scales up from and down to zero automatically.
Cloud Run is the fastest way to deploy containerized apps. Push your code in Go, Python, Node.js, Java, or any language and Cloud Run builds and deploys it automatically. Get fast autoscaling, pay only when your code runs, and skip the infrastructure headaches. Two million requests free per month. And new customers get $300 in free credit.
...Installation Videos!
Part 1: http://youtu.be/rnv2VLcG-eI
Part 2: http://youtu.be/eFudbMWHNlQ
Special thanks to Wells Oliver for the code for downloading Retrosheet files. And the Chadwick project for its Retrosheet tools. https://sourceforge.net/projects/chadwick/?source=recommended
Basic and intermediate examples of DICOM library with Jupyter
Basic and intermediate examples to read, modify and write DICOM files with Python code using Jupyter - To install Jupyter - https://jupyter.org/install
======
All examples are based on Pydicom. An open source library - https://pydicom.github.io/
Make AsciiDoc part of your literate programming tool set. With eWEB you can weave and tangle literate programs written as AsciiDoc documents, using embedded WEB code snippets.