Simple, Pythonic building blocks to evaluate LLM applications
Open-source AI hackers to find and fix your app’s vulnerabilities
A powerful tool for automated LLM fuzzing
AI Agent Evaluator & Red Team Platform
SDG is a specialized framework
270+ Claude Code plugins with 739 agent skills
Advanced LLM-powered brute-force tool combining AI intelligence
A Comprehensive Benchmark to Evaluate LLMs as Agents (ICLR'24)
Tools like web browser, computer access and code runner for LLMs
A security scanner for custom LLM applications
One-stop solution for creating your digital avatar from chat history
The open source post-building layer for agents
AI-powered penetration testing assistant using local LLM on linux
Evaluate your LLM's response with Prometheus and GPT4
A high-performance ML model serving framework, offers dynamic batching
Leaderboard Comparing LLM Performance at Producing Hallucinations
Semi-Structured Agentic Framework. Workflows build themselves
A.S.E (AICGSecEval) is a repository-level AI-generated code security
E2B Desktop Sandbox for LLMs. E2B Sandbox
Chinese Llama-3 LLMs) developed from Meta Llama 3
Retrieval Augmented Generation (RAG) framework
The unofficial python package that returns response of Google Bard
An implementation of model parallel GPT-2 and GPT-3-style models