Distributed LLM and StableDiffusion inference
A minimal, secure Python interpreter written in Rust for use by AI
Local AI coding agent CLI with multi-agent orchestration tools
Delivery infrastructure for agentic apps
The GPU-powered AI application database
Arch is an intelligent prompt gateway. Engineered with (fast) LLMs
A drop-in Apache Spark replacement written in Rust
CodeGeeX4-ALL-9B, a versatile model for all AI software development
A private, local meeting notes assistant
An improved implementation of the Ralph Wiggum technique
Convert codebases into structured prompts optimized for LLM analysis
Framework to prove inference of ML models blazingly fast
High-performance inference server for text embeddings models API layer
Package and deploy machine learning models using Docker containers
Fast ML inference & training for ONNX models in Rust
Serialize repositories into LLM-ready context w/ smart prioritization
Graph-vector database for building unified AI backends fast
Visual AI IDE for building agents with prompt chains and graphs
Secure local-first microVM sandbox for running untrusted code fast
High-performance API combining reasoning and creative AI models
Rust framework for building modular and scalable LLM-powered apps
Cloud-native open source data warehouse for analytics and AI queries
Open source IDE for orchestrating AI coding agents in large codebases
A tool to snap pixels to a perfect grid
Rust async runtime based on io-uring