Efficient few-shot learning with Sentence Transformers
Large Language Model Text Generation Inference
On-device AI across mobile, embedded and edge for PyTorch
The official Python client for the Huggingface Hub
Uplift modeling and causal inference with machine learning algorithms
DoWhy is a Python library for causal inference
Integrate, train and manage any AI models and APIs with your database
Official inference library for Mistral models
GPU environment management and cluster orchestration
Sparsity-aware deep learning inference runtime for CPUs
Open-Source AI Camera. Empower any camera/CCTV
Lightweight inference library for ONNX files, written in C++
LLM training code for MosaicML foundation models
Fast inference engine for Transformer models
Easy-to-use deep learning framework with 3 key features
C#/.NET binding of llama.cpp, including LLaMa/GPT model inference
PArallel Distributed Deep LEarning: Machine Learning Framework
Training and deploying machine learning models on Amazon SageMaker
State-of-the-art Parameter-Efficient Fine-Tuning
Run serverless GPU workloads with fast cold starts on bare-metal
Pure C++ implementation of several models for real-time chatting
lightweight, standalone C++ inference engine for Google's Gemma models
Multilingual Automatic Speech Recognition with word-level timestamps
Database system for building simpler and faster AI-powered application
Lightweight Python library for adding real-time multi-object tracking