Showing 66 open source projects for "performance testing"

View related business solutions
  • Our Free Plans just got better! | Auth0 Icon
    Our Free Plans just got better! | Auth0

    With up to 25k MAUs and unlimited Okta connections, our Free Plan lets you focus on what you do best—building great apps.

    You asked, we delivered! Auth0 is excited to expand our Free and Paid plans to include more options so you can focus on building, deploying, and scaling applications without having to worry about your security. Auth0 now, thank yourself later.
    Try free now
  • Fully Managed MySQL, PostgreSQL, and SQL Server Icon
    Fully Managed MySQL, PostgreSQL, and SQL Server

    Automatic backups, patching, replication, and failover. Focus on your app, not your database.

    Cloud SQL handles your database ops end to end, so you can focus on your app.
    Try Free
  • 1
    Giskard

    Giskard

    Collaborative & Open-Source Quality Assurance for all AI models

    The testing framework dedicated to ML models, from tabular to LLMs. Giskard is an open-source testing framework dedicated to ML models, from tabular models to LLMs. Testing Machine Learning applications can be tedious. Since ML models depend on data, testing scenarios depend on the domain specificities and are often infinite. At Giskard, we believe that Machine Learning needs its own testing framework.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 2
    mistral.rs

    mistral.rs

    Fast, flexible LLM inference

    mistral.rs is a fast and flexible LLM inference engine implemented in Rust, designed to run and serve modern language models with an emphasis on performance and practical deployment. It provides multiple entry points for developers, including a CLI for running models locally and an HTTP server that exposes an OpenAI-compatible API surface for easy integration with existing clients. The project includes hardware-aware tooling that can benchmark a system and choose sensible quantization and device-mapping strategies, helping users get strong performance without manual tuning. ...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 3
    Arcade AI

    Arcade AI

    Arcade Tool Development Kit (TDK), Worker, Evals, and CLI

    ...Core platform functionality and schemas. This repository contains the core Arcade libraries, organized as separate packages for maximum flexibility and modularity. Evaluation framework for testing tool performance. Test your MCP server's tools, resources, prompts, elicitation, and OAuth 2. MCPJam is compliant with the latest MCP specs. Connect to any MCP server. MCPJam inspector supports STDIO, SSE, and Streamable HTTP transports.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 4
    Nestia

    Nestia

    NestJS Helper + AI Chatbot Development

    Nestia is a high-performance toolkit and ecosystem built on top of NestJS that enhances backend development by introducing strongly typed APIs, automated SDK generation, and advanced tooling for scalable server applications. It is designed to eliminate much of the boilerplate typically associated with API development by leveraging pure TypeScript types to automatically generate validation logic, API documentation, and client SDKs. One of its defining advantages is its focus on performance...
    Downloads: 7 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build, govern, and optimize agents and models with Gemini Enterprise Agent Platform.
    Start Free
  • 5
    Ollama Grid Search

    Ollama Grid Search

    A multi-platform desktop application to evaluate and compare LLM

    ...The system integrates directly with local or remote Ollama servers, enabling seamless access to models already deployed in a user’s environment. It also includes experiment logging and A/B testing capabilities, which allow users to compare outputs side by side and track performance metrics such as latency or token usage.
    Downloads: 11 This Week
    Last Update:
    See Project
  • 6
    Tianji

    Tianji

    Evaluation suite designed to assess the performance of LLMs

    Tianji is a comprehensive evaluation suite designed to assess the performance of large language models (LLMs) across multiple dimensions. It focuses on measuring general capabilities such as reasoning, knowledge, commonsense, and language understanding. Tianji provides a curated set of benchmarks and a unified framework for systematically comparing LLMs, making it useful for research and model selection.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Claw Code

    Claw Code

    AI agent harness for AI coding agents

    Claw Code is an open-source AI agent harness project focused on building better tools for orchestrating and managing autonomous coding agents. It originated as a clean-room reimplementation inspired by the architecture of Claude Code, aiming to replicate core concepts without using proprietary code. The project provides a Python-based foundation for experimenting with agent workflows, tool integration, and task execution pipelines. It emphasizes harness engineering—how agents are structured,...
    Downloads: 36 This Week
    Last Update:
    See Project
  • 8
    FullTClash

    FullTClash

    General proxy performance testing tool based on Clash using Telegram

    Back end part useClash project(It can also be called nowmihomo)The relevant code is used as the outing agent. The front end part uses Telegram API as the interactive interface, which needs to be used in conjunction with Telegram, that is, a Telegram robot (bot), FullTClash bot is a Telegram robot (hereinafter referred to as bot) carrying its test tasks.
    Downloads: 34 This Week
    Last Update:
    See Project
  • 9
    AgentBench

    AgentBench

    A Comprehensive Benchmark to Evaluate LLMs as Agents (ICLR'24)

    ...These environments require agents to interpret instructions, take actions, and adapt their strategies based on feedback from the environment. AgentBench also includes an evaluation framework that measures success rates, rewards, and task completion performance across different agent implementations. By testing models across diverse scenarios, the benchmark highlights strengths and weaknesses in reasoning, long-term planning, and tool usage.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Build Securely on AWS with Proven Frameworks Icon
    Build Securely on AWS with Proven Frameworks

    Lay a foundation for success with Tested Reference Architectures developed by Fortinet’s experts. Learn more in this white paper.

    Moving to the cloud brings new challenges. How can you manage a larger attack surface while ensuring great network performance? Turn to Fortinet’s Tested Reference Architectures, blueprints for designing and securing cloud environments built by cybersecurity experts. Learn more and explore use cases in this white paper.
    Download Now
  • 10
    Easy DataSet

    Easy DataSet

    A powerful tool for creating datasets for LLM fine-tuning

    ...The system includes automated question-generation capabilities, hierarchical label trees, and answer generation pipelines that use LLM APIs to produce coherent paired data with customizable templates. Beyond dataset creation, Easy-dataset also provides a built-in evaluation system with model testing and blind-test features, helping teams validate model performance using curated test sets.
    Downloads: 9 This Week
    Last Update:
    See Project
  • 11
    Beehave

    Beehave

    Behavior tree AI for Godot Engine

    Beehave is a powerful AI behavior tree framework designed as an addon for the Godot game engine, enabling developers to create sophisticated and dynamic non-player character behaviors in games. It uses a node-based system that integrates directly into the Godot scene tree, allowing developers to visually design and organize complex AI logic in a structured and intuitive way. Behavior trees provide a modular approach to decision-making, making it easier to manage large and adaptive AI systems...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 12
    Simple Evals

    Simple Evals

    Lightweight framework for evaluating large language model performance

    simple-evals is a lightweight evaluation framework developed by OpenAI for quickly testing models against small, focused benchmarks. It is designed to help researchers and developers run targeted evaluations without the complexity of large-scale pipelines. By emphasizing simplicity, the framework makes it easy to define new tasks, run evaluations, and interpret results in a reproducible way. It is particularly useful for sanity checks, exploratory research, and comparing performance across different models or configurations. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 13
    TextWorld

    TextWorld

    ​TextWorld is a sandbox learning environment for the training

    TextWorld is a learning environment designed to train reinforcement learning agents to play text-based games, where actions and observations are entirely in natural language. Developed by Microsoft Research, TextWorld focuses on language understanding, planning, and interaction in complex, narrative-driven environments. It generates games procedurally, enabling scalable testing of agents’ natural language processing and decision-making abilities.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 14
    AutoAgent AI

    AutoAgent AI

    Autonomous harness engineering

    ...The system is particularly useful for building domain-specific agents that need continuous performance improvement.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 15
    langrocks

    langrocks

    Tools like web browser, computer access and code runner for LLMs

    Langrocks is a programming language experimentation toolkit that enables developers to create, test, and optimize custom programming languages.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 16
    Prompt flow

    Prompt flow

    Build high-quality LLM apps

    Prompt flow is a suite of development tools designed to streamline the end-to-end development cycle of LLM-based AI applications, from ideation, prototyping, testing, and evaluation to production deployment and monitoring. It makes prompt engineering much easier and enables you to build LLM apps with production quality.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Agent Behavior Monitoring

    Agent Behavior Monitoring

    The open source post-building layer for agents

    Agent Behavior Monitoring is an open-source framework designed to monitor, evaluate, and improve the behavior of AI agents operating in real or simulated environments. The system focuses on agent behavior monitoring by collecting interaction data and analyzing how agents perform across different scenarios and tasks. Developers can use the framework to observe agent actions in both online production environments and offline evaluation settings, making it useful for debugging and performance...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 18
    TensorFlow Serving

    TensorFlow Serving

    Serving system for machine learning models

    TensorFlow Serving is a flexible, high-performance serving system for machine learning models, designed for production environments. It deals with the inference aspect of machine learning, taking models after training and managing their lifetimes, providing clients with versioned access via a high-performance, reference-counted lookup table. TensorFlow Serving provides out-of-the-box integration with TensorFlow models, but can be easily extended to serve other types of models and data. The...
    Downloads: 8 This Week
    Last Update:
    See Project
  • 19
    LangWatch

    LangWatch

    The platform for LLM evaluations and AI agent testing

    LangWatch is an open-source observability and monitoring platform designed to help developers evaluate and improve applications built with large language models. The platform provides tools for tracking model interactions, analyzing prompt behavior, and identifying issues such as hallucinations, latency problems, or unexpected responses. By collecting telemetry data from AI applications, LangWatch allows developers to understand how their systems perform in real-world usage scenarios. The...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 20
    TapeAgents

    TapeAgents

    A framework that facilitates all stages of LLM development

    TapeAgents is a framework that facilitates all stages of the Large Language Model (LLM) agent development lifecycle, providing tools for building, testing, and deploying AI agents.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 21
    Paddler

    Paddler

    Open-source LLM load balancer and serving platform for hosting LLMs

    Paddler is an open-source LLM infrastructure platform designed to deploy, manage, and scale large language models on private infrastructure. The system acts as a specialized load balancer and serving layer for language models, enabling organizations to run inference workloads without relying on external API providers. It supports running models locally through engines such as llama.cpp while distributing requests across multiple compute nodes to improve performance and reliability. The...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    mlx

    mlx

    MLX: An array framework for Apple silicon

    MlX offers a local web interface to browse, download, and run ML models via Hugging Face or local sources. It supports searching by tags or tasks, visualization of model metadata, quick inference demos, automatic setup of runtime environments, and works with PyTorch, TensorFlow, and ONNX. Ideal for researchers exploring and testing models via browser.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 23
    NOFX

    NOFX

    Open source AI trading OS for autonomous multi-model trading systems

    NOFX is an open source AI-powered trading operating system designed to automate financial trading workflows using autonomous AI agents. It acts as an infrastructure layer that transforms market data into AI-driven trade decisions and execution. Instead of requiring users to manually configure machine learning models, data sources, and API integrations, the system allows AI components to perceive market conditions, select models, and perform trading actions automatically. It supports running...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 24
    Mosec

    Mosec

    A high-performance ML model serving framework, offers dynamic batching

    Mosec is a high-performance and flexible model-serving framework for building ML model-enabled backend and microservices. It bridges the gap between any machine learning models you just trained and the efficient online service API.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 25
    GLM-4.1V

    GLM-4.1V

    GLM-4.6V/4.5V/4.1V-Thinking, towards versatile multimodal reasoning

    ...Given its open-source availability under the same project repository, it provides an accessible entry point for testing multimodal reasoning and building proof-of-concept applications.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • Next
MongoDB Logo MongoDB