A Unified Library for Parameter-Efficient Learning
Superduper: Integrate AI models and machine learning workflows
Images to inference with no labeling
Scripts for fine-tuning Meta Llama3 with composable FSDP & PEFT method
GPU environment management and cluster orchestration
INT4/INT5/INT8 and FP16 inference on CPU for RWKV language model
Open-source tool designed to enhance the efficiency of workloads
State-of-the-art Parameter-Efficient Fine-Tuning
PyTorch extensions for fast R&D prototyping and Kaggle farming
Probabilistic reasoning and statistical analysis in TensorFlow
Open platform for training, serving, and evaluating language models
Uncover insights, surface problems, monitor, and fine tune your LLM
Tensor search for humans
A toolkit to optimize ML models for deployment for Keras & TensorFlow
A lightweight vision library for performing large object detection
Powering Amazon custom machine learning chips
A GPU-accelerated library containing highly optimized building blocks
Serve machine learning models within a Docker container
LLMFlows - Simple, Explicit and Transparent LLM Apps
Run any Llama 2 locally with gradio UI on GPU or CPU from anywhere
Framework for Accelerating LLM Generation with Multiple Decoding Heads
Run 100B+ language models at home, BitTorrent-style
Implementation of "Tree of Thoughts
Toolbox of models, callbacks, and datasets for AI/ML researchers
Lightweight anchor-free object detection model