DoWhy is a Python library for causal inference
Operating LLMs in production
The unofficial python package that returns response of Google Bard
Multi-Modal Neural Networks for Semantic Search, based on Mid-Fusion
Trainable models and NN optimization tools
MII makes low-latency and high-throughput inference possible
Official inference library for Mistral models
Superduper: Integrate AI models and machine learning workflows
AIMET is a library that provides advanced quantization and compression
INT4/INT5/INT8 and FP16 inference on CPU for RWKV language model
A set of Docker images for training and serving models in TensorFlow
Fast inference engine for Transformer models
Deep learning optimization library: makes distributed training easy
Run any Llama 2 locally with gradio UI on GPU or CPU from anywhere
Implementation of "Tree of Thoughts
Sequence-to-sequence framework, focused on Neural Machine Translation
Guide to deploying deep-learning inference networks
Toolkit for allowing inference and serving with MXNet in SageMaker