GPU environment management and cluster orchestration
Phi-3.5 for Mac: Locally-run Vision and Language Models
A Unified Library for Parameter-Efficient Learning
Scripts for fine-tuning Meta Llama3 with composable FSDP & PEFT method
Run local LLMs like llama, deepseek, kokoro etc. inside your browser
DoWhy is a Python library for causal inference
PyTorch library of curated Transformer models and their components
State-of-the-art Parameter-Efficient Fine-Tuning
Simplifies the local serving of AI models from any source
Superduper: Integrate AI models and machine learning workflows
OpenMLDB is an open-source machine learning database
Multilingual Automatic Speech Recognition with word-level timestamps
Turn your existing data infrastructure into a feature store
Python Package for ML-Based Heterogeneous Treatment Effects Estimation
A high-performance ML model serving framework, offers dynamic batching
A toolkit to optimize ML models for deployment for Keras & TensorFlow
A scalable inference server for models optimized with OpenVINO
Build Production-ready Agentic Workflow with Natural Language
Low-latency REST API for serving text-embeddings
A library for accelerating Transformer models on NVIDIA GPUs
The unofficial python package that returns response of Google Bard
Trainable, memory-efficient, and GPU-friendly PyTorch reproduction
Trainable models and NN optimization tools
Probabilistic reasoning and statistical analysis in TensorFlow
A GPU-accelerated library containing highly optimized building blocks