Go from idea to deployed AI app without managing infrastructure. Vertex AI offers one platform for the entire AI development lifecycle.
Ship AI apps and features faster with Vertex AI—your end-to-end AI platform. Access Gemini 3 and 200+ foundation models, fine-tune for your needs, and deploy with enterprise-grade MLOps. Build chatbots, agents, or custom models. New customers get $300 in free credit.
Try Vertex AI Free
$300 in Free Credit for Your Google Cloud Projects
Build, test, and explore on Google Cloud with $300 in free credit. No hidden charges. No surprise bills.
Launch your next project with $300 in free Google Cloud credit—no hidden charges. Test, build, and deploy without risk. Use your credit across the Google Cloud platform to find what works best for your needs. After your credits are used, continue building with free monthly usage products. Only pay when you're ready to scale. Sign up in minutes and start exploring.
An educational software system of a tiny self-compiling C compiler
An educational software system of a tiny self-compiling C compiler, a tiny self-executing RISC-V emulator, and a tiny self-hosting RISC-V hypervisor. Selfie is a project of the Computational Systems Group at the Department of Computer Sciences of the University of Salzburg in Austria. The Selfie Project provides an educational platform for teaching undergraduate and graduate students the design and implementation of programming languages and runtime systems.
Tiny (less then 1000 lines of code) education compiler that can compile itself. Implements small subset of C language without pointers, structures etc. Parameters and local variables also not implemented. To interface with OS Borland-style inline assembler are used (db directive only). Based on initial version of Tiny Context compiler (http://avhohlov.narod.ru/p1805en.htm).
...It may be used by computer languages as a quick path to have a compatible aeabi library.
It is not supposed to be as fast as assembler.
This project started as a laboratory to provide standalone aeabi for TCC (Tiny CC - http://savannah.nongnu.org/projects/tinycc) but may be used in any software requiring its own aeabi library.
GCC provides a much better implementation but requires gcc shared lib.
An open source COBOL compiler for the PC x86-32 platform.
Generates GNU assembler and requires GCC to produce a binary.
Supports Linux, BSD and MinGW 32 bit platforms.
Migrate from Snowflake, Databricks, or Redshift with free migration tools. Exabyte scale without the Exabyte price.
BigQuery delivers up to 54% lower TCO than cloud alternatives. Migrate from legacy or competing warehouses using free BigQuery Migration Service with automated SQL translation. Get serverless scale with no infrastructure to manage, compressed storage, and flexible pricing—pay per query or commit for deeper discounts. New customers get $300 in free credit.
...Mainframe, mid-range, Mac, PC, Hercules, Raspberry.
Signal interest, we'll set you up with some (vetted) file, distribution and communication space.
Wildcat, GnuCOBOL, Micro Focus, Elastic, RainCode, IBM, Tiny COBOL, all the others, all in.
COBOL-UP
latexcalc is a "LaTeX Calculator" that calculates values inside your LaTeX files before typesetting them. More accurately, it preprocesses files written in a superset of the LaTeX typesetting language and evaluates specified expressions in the text.
TScript is an opensource scripting engine, including two demos and some tiny games written in it by friends. It's easy to make your system configurable at a script level by adopting TScript. Please refer to http://tscripta.sf.net
Tiny C compiler, programmed in C with compiler/linker/assembler included. VERY fast and mainly usable as a library. WIN32 port of the http://tinycc.org version. Compiles code to memory currently, perhaps disk soon. Go to the homepage for more info.
Run everything from popular models with on-demand NVIDIA L4 GPUs to web apps without infrastructure management.
Run frontend and backend services, batch jobs, host LLMs, and queue processing workloads without the need to manage infrastructure. Cloud Run gives you on-demand GPU access for hosting LLMs and running real-time AI—with 5-second cold starts and automatic scale-to-zero so you only pay for actual usage. New customers get $300 in free credit to start.