Large Language Model Text Generation Inference
Low-latency REST API for serving text-embeddings
Tensor search for humans
Efficient few-shot learning with Sentence Transformers
MII makes low-latency and high-throughput inference possible
State-of-the-art diffusion models for image and audio generation
Framework that is dedicated to making neural data processing
Phi-3.5 for Mac: Locally-run Vision and Language Models
Easy-to-use Speech Toolkit including Self-Supervised Learning model
LLM training code for MosaicML foundation models
Framework for Accelerating LLM Generation with Multiple Decoding Heads
Implementation of "Tree of Thoughts
Training & Implementation of chatbots leveraging GPT-like architecture
CPU/GPU inference server for Hugging Face transformer models