Lightning-Fast RL for LLM Reasoning and Agents. Made Simple & Flexible
A Tree Search Library with Flexible API for LLM Inference-Time Scaling
Semantic cache for LLMs. Fully integrated with LangChain
Go package for computer vision using OpenCV 4 and beyond
Helping you get the most out of AWS, wherever you use MCP
Inference framework for 1-bit LLMs
Leveraging BERT and c-TF-IDF to create easily interpretable topics
FlashInfer: Kernel Library for LLM Serving
Optimizing inference proxy for LLMs
Go ahead and axolotl questions
HexStrike AI MCP Agents is an advanced MCP server
Fault-tolerant, highly scalable GPU orchestration
A Family of Open Foundation Models for Code Intelligence
Utilities intended for use with Llama models
Low-latency REST API for serving text-embeddings
BISHENG is an open LLM devops platform for next generation apps
Ray Aviary - evaluate multiple LLMs easily
Central interface to connect your LLM's with external data
Pruna is a model optimization framework built for developers
A Model Context Protocol (MCP) server that enables AI assistants
Build your chatbot within minutes on your favorite device
Easiest and laziest way for building multi-agent LLMs applications
A framework that facilitates all stages of LLM development
Learn AI and LLMs from scratch using free resources
GLM-4 series: Open Multilingual Multimodal Chat LMs