Test and evaluate LLMs and model configurations
Evaluate and compare LLM outputs, catch regressions, improve prompts
Open-source end-to-end LLM Development Platform
Low-code app builder for RAG and multi-agent AI applications
Tools like web browser, computer access and code runner for LLMs
Open-weight, large-scale hybrid-attention reasoning model
Open-source, developer-first LLMOps platform
AI agent that streamlines the entire process of data analysis
Distributed LLM and StableDiffusion inference
Code for Language models can explain neurons in language models paper
Ray Aviary - evaluate multiple LLMs easily
Beyond the Imitation Game collaborative benchmark for measuring
AI R&D Efficiency Improvement Research: Do-It-Yourself Training LoRA
llama.go is like llama.cpp in pure Golang
Community for applying LLMs to robotics and a robot simulator
Implements a reference architecture for creating information systems
8.5K high quality grade school math problems