Pure C++ implementation of several models for real-time chatting
ONNX Runtime: cross-platform, high performance ML inferencing
Lightweight Python library for adding real-time multi-object tracking
Self-hosted, community-driven, local OpenAI compatible API
OpenMLDB is an open-source machine learning database
A toolkit to optimize ML models for deployment for Keras & TensorFlow
Protect and discover secrets using Gitleaks
C++ implementation of ChatGLM-6B & ChatGLM2-6B & ChatGLM3 & GLM4(V)
Openai style api for open large language models
Large Language Model Text Generation Inference
GPU environment management and cluster orchestration
Multilingual Automatic Speech Recognition with word-level timestamps
A scalable inference server for models optimized with OpenVINO
Integrate, train and manage any AI models and APIs with your database
PyTorch extensions for fast R&D prototyping and Kaggle farming
The Triton Inference Server provides an optimized cloud
Serving system for machine learning models
Deep Learning API and Server in C++14 support for Caffe, PyTorch
Powering Amazon custom machine learning chips
High quality, fast, modular reference implementation of SSD in PyTorch
A real time inference engine for temporal logical specifications
Database system for building simpler and faster AI-powered application
OpenFieldAI is an AI based Open Field Test Rodent Tracker
A computer vision framework to create and deploy apps in minutes
Toolbox of models, callbacks, and datasets for AI/ML researchers