The Triton Inference Server provides an optimized cloud
Deep Learning API and Server in C++14 support for Caffe, PyTorch
Python binding to the Apache Tika™ REST services
Low-latency REST API for serving text-embeddings
A client implementation for ChatGPT and Bing AI
Framework for intelligent service-based networks. Mobile compatible.
Leading free and open-source face recognition system
Deploy a ML inference service on a budget in 10 lines of code
Fast Coreference Resolution in spaCy with Neural Networks