Phi-3.5 for Mac: Locally-run Vision and Language Models
Neural Network Compression Framework for enhanced OpenVINO
GPU environment management and cluster orchestration
MII makes low-latency and high-throughput inference possible
Trainable, memory-efficient, and GPU-friendly PyTorch reproduction
Probabilistic reasoning and statistical analysis in TensorFlow
A Unified Library for Parameter-Efficient Learning
Large Language Model Text Generation Inference
Images to inference with no labeling
Adversarial Robustness Toolbox (ART) - Python Library for ML security
DoWhy is a Python library for causal inference
Scripts for fine-tuning Meta Llama3 with composable FSDP & PEFT method
PyTorch library of curated Transformer models and their components
Run any Llama 2 locally with gradio UI on GPU or CPU from anywhere
State-of-the-art Parameter-Efficient Fine-Tuning
Framework that is dedicated to making neural data processing
Open platform for training, serving, and evaluating language models
Library for serving Transformers models on Amazon SageMaker
Superduper: Integrate AI models and machine learning workflows
Database system for building simpler and faster AI-powered application
Low-latency REST API for serving text-embeddings
Multilingual Automatic Speech Recognition with word-level timestamps
Integrate, train and manage any AI models and APIs with your database
A high-performance ML model serving framework, offers dynamic batching
The unofficial python package that returns response of Google Bard