Library for serving Transformers models on Amazon SageMaker
The official Python client for the Huggingface Hub
An easy-to-use LLMs quantization package with user-friendly apis
Openai style api for open large language models
Efficient few-shot learning with Sentence Transformers
A Unified Library for Parameter-Efficient Learning
The unofficial python package that returns response of Google Bard
MII makes low-latency and high-throughput inference possible
Database system for building simpler and faster AI-powered application
Implementation of model parallel autoregressive transformers on GPUs
CPU/GPU inference server for Hugging Face transformer models