Solve puzzles. Learn CUDA
Performance-optimized AI inference on your GPUs
Running large language models on a single GPU
The fundamental package for scientific computing with Python
AirLLM 70B inference with single 4GB GPU
Parallax is a distributed model serving framework
How to optimize some algorithm in cuda
Development repository for the Triton language and compiler
Run Local LLMs on Any Device. Open-source
SkyPilot: Run AI and batch jobs on any infra
NVIDIA Isaac Sim is an open-source application on NVIDIA Omniverse
AI agents running research on single-GPU nanochat training
Reinforcement Learning for Humanoid Robot with Zero-Shot Sim2Real
A high-quality rapid TTS voice cloning model
Ongoing research training transformer models at scale
A sound cloning tool with a web interface, using your voice
State-of-the-art Parameter-Efficient Fine-Tuning
Voice Recognition to Text Tool
An opinionated CLI to transcribe Audio files w/ Whisper on-device
Making large AI models cheaper, faster and more accessible
Pythonic tool for running machine-learning/high performance workflows
ChatGLM-6B: An Open Bilingual Dialogue Language Model
Simplifies the local serving of AI models from any source
High-Resolution Image Synthesis with Latent Diffusion Models
Official inference framework for 1-bit LLMs