Showing 63 open source projects for "performance testing"

View related business solutions
  • Compliant and Reliable File Transfers Backed by Top Security Certifications Icon
    Compliant and Reliable File Transfers Backed by Top Security Certifications

    Cerberus FTP Server delivers SOC 2 Type II certified security and FIPS 140-2 validated encryption.

    Stop relying on non-certified, legacy file transfer tools that creak under the weight of modern security demands. Get full audit trails, advanced access controls and more supported by an award-winning team of experts. Start your free 25-day trial today.
    Start Free Trial
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • 1
    mistral.rs

    mistral.rs

    Fast, flexible LLM inference

    mistral.rs is a fast and flexible LLM inference engine implemented in Rust, designed to run and serve modern language models with an emphasis on performance and practical deployment. It provides multiple entry points for developers, including a CLI for running models locally and an HTTP server that exposes an OpenAI-compatible API surface for easy integration with existing clients. The project includes hardware-aware tooling that can benchmark a system and choose sensible quantization and device-mapping strategies, helping users get strong performance without manual tuning. ...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 2
    Arcade AI

    Arcade AI

    Arcade Tool Development Kit (TDK), Worker, Evals, and CLI

    ...Core platform functionality and schemas. This repository contains the core Arcade libraries, organized as separate packages for maximum flexibility and modularity. Evaluation framework for testing tool performance. Test your MCP server's tools, resources, prompts, elicitation, and OAuth 2. MCPJam is compliant with the latest MCP specs. Connect to any MCP server. MCPJam inspector supports STDIO, SSE, and Streamable HTTP transports.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 3
    Nestia

    Nestia

    NestJS Helper + AI Chatbot Development

    Nestia is a high-performance toolkit and ecosystem built on top of NestJS that enhances backend development by introducing strongly typed APIs, automated SDK generation, and advanced tooling for scalable server applications. It is designed to eliminate much of the boilerplate typically associated with API development by leveraging pure TypeScript types to automatically generate validation logic, API documentation, and client SDKs. One of its defining advantages is its focus on performance...
    Downloads: 7 This Week
    Last Update:
    See Project
  • 4
    Ollama Grid Search

    Ollama Grid Search

    A multi-platform desktop application to evaluate and compare LLM

    ...The system integrates directly with local or remote Ollama servers, enabling seamless access to models already deployed in a user’s environment. It also includes experiment logging and A/B testing capabilities, which allow users to compare outputs side by side and track performance metrics such as latency or token usage.
    Downloads: 11 This Week
    Last Update:
    See Project
  • Auth0 B2B Essentials: SSO, MFA, and RBAC Built In Icon
    Auth0 B2B Essentials: SSO, MFA, and RBAC Built In

    Unlimited organizations, 3 enterprise SSO connections, role-based access control, and pro MFA included. Dev and prod tenants out of the box.

    Auth0's B2B Essentials plan gives you everything you need to ship secure multi-tenant apps. Unlimited orgs, enterprise SSO, RBAC, audit log streaming, and higher auth and API limits included. Add on M2M tokens, enterprise MFA, or additional SSO connections as you scale.
    Sign Up Free
  • 5
    Tianji

    Tianji

    Evaluation suite designed to assess the performance of LLMs

    Tianji is a comprehensive evaluation suite designed to assess the performance of large language models (LLMs) across multiple dimensions. It focuses on measuring general capabilities such as reasoning, knowledge, commonsense, and language understanding. Tianji provides a curated set of benchmarks and a unified framework for systematically comparing LLMs, making it useful for research and model selection.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    Claw Code

    Claw Code

    AI agent harness for AI coding agents

    Claw Code is an open-source AI agent harness project focused on building better tools for orchestrating and managing autonomous coding agents. It originated as a clean-room reimplementation inspired by the architecture of Claude Code, aiming to replicate core concepts without using proprietary code. The project provides a Python-based foundation for experimenting with agent workflows, tool integration, and task execution pipelines. It emphasizes harness engineering—how agents are structured,...
    Downloads: 36 This Week
    Last Update:
    See Project
  • 7
    FullTClash

    FullTClash

    General proxy performance testing tool based on Clash using Telegram

    Back end part useClash project(It can also be called nowmihomo)The relevant code is used as the outing agent. The front end part uses Telegram API as the interactive interface, which needs to be used in conjunction with Telegram, that is, a Telegram robot (bot), FullTClash bot is a Telegram robot (hereinafter referred to as bot) carrying its test tasks.
    Downloads: 34 This Week
    Last Update:
    See Project
  • 8
    AgentBench

    AgentBench

    A Comprehensive Benchmark to Evaluate LLMs as Agents (ICLR'24)

    ...These environments require agents to interpret instructions, take actions, and adapt their strategies based on feedback from the environment. AgentBench also includes an evaluation framework that measures success rates, rewards, and task completion performance across different agent implementations. By testing models across diverse scenarios, the benchmark highlights strengths and weaknesses in reasoning, long-term planning, and tool usage.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    Easy DataSet

    Easy DataSet

    A powerful tool for creating datasets for LLM fine-tuning

    ...The system includes automated question-generation capabilities, hierarchical label trees, and answer generation pipelines that use LLM APIs to produce coherent paired data with customizable templates. Beyond dataset creation, Easy-dataset also provides a built-in evaluation system with model testing and blind-test features, helping teams validate model performance using curated test sets.
    Downloads: 9 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build, govern, and optimize agents and models with Gemini Enterprise Agent Platform.
    Start Free
  • 10
    Beehave

    Beehave

    Behavior tree AI for Godot Engine

    Beehave is a powerful AI behavior tree framework designed as an addon for the Godot game engine, enabling developers to create sophisticated and dynamic non-player character behaviors in games. It uses a node-based system that integrates directly into the Godot scene tree, allowing developers to visually design and organize complex AI logic in a structured and intuitive way. Behavior trees provide a modular approach to decision-making, making it easier to manage large and adaptive AI systems...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 11
    Simple Evals

    Simple Evals

    Lightweight framework for evaluating large language model performance

    simple-evals is a lightweight evaluation framework developed by OpenAI for quickly testing models against small, focused benchmarks. It is designed to help researchers and developers run targeted evaluations without the complexity of large-scale pipelines. By emphasizing simplicity, the framework makes it easy to define new tasks, run evaluations, and interpret results in a reproducible way. It is particularly useful for sanity checks, exploratory research, and comparing performance across different models or configurations. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 12
    TextWorld

    TextWorld

    ​TextWorld is a sandbox learning environment for the training

    TextWorld is a learning environment designed to train reinforcement learning agents to play text-based games, where actions and observations are entirely in natural language. Developed by Microsoft Research, TextWorld focuses on language understanding, planning, and interaction in complex, narrative-driven environments. It generates games procedurally, enabling scalable testing of agents’ natural language processing and decision-making abilities.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 13
    AutoAgent AI

    AutoAgent AI

    Autonomous harness engineering

    ...The system is particularly useful for building domain-specific agents that need continuous performance improvement.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 14
    langrocks

    langrocks

    Tools like web browser, computer access and code runner for LLMs

    Langrocks is a programming language experimentation toolkit that enables developers to create, test, and optimize custom programming languages.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 15
    Prompt flow

    Prompt flow

    Build high-quality LLM apps

    Prompt flow is a suite of development tools designed to streamline the end-to-end development cycle of LLM-based AI applications, from ideation, prototyping, testing, and evaluation to production deployment and monitoring. It makes prompt engineering much easier and enables you to build LLM apps with production quality.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Agent Behavior Monitoring

    Agent Behavior Monitoring

    The open source post-building layer for agents

    Agent Behavior Monitoring is an open-source framework designed to monitor, evaluate, and improve the behavior of AI agents operating in real or simulated environments. The system focuses on agent behavior monitoring by collecting interaction data and analyzing how agents perform across different scenarios and tasks. Developers can use the framework to observe agent actions in both online production environments and offline evaluation settings, making it useful for debugging and performance...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 17
    LangWatch

    LangWatch

    The platform for LLM evaluations and AI agent testing

    LangWatch is an open-source observability and monitoring platform designed to help developers evaluate and improve applications built with large language models. The platform provides tools for tracking model interactions, analyzing prompt behavior, and identifying issues such as hallucinations, latency problems, or unexpected responses. By collecting telemetry data from AI applications, LangWatch allows developers to understand how their systems perform in real-world usage scenarios. The...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 18
    TensorFlow Serving

    TensorFlow Serving

    Serving system for machine learning models

    TensorFlow Serving is a flexible, high-performance serving system for machine learning models, designed for production environments. It deals with the inference aspect of machine learning, taking models after training and managing their lifetimes, providing clients with versioned access via a high-performance, reference-counted lookup table. TensorFlow Serving provides out-of-the-box integration with TensorFlow models, but can be easily extended to serve other types of models and data. The...
    Downloads: 8 This Week
    Last Update:
    See Project
  • 19
    TapeAgents

    TapeAgents

    A framework that facilitates all stages of LLM development

    TapeAgents is a framework that facilitates all stages of the Large Language Model (LLM) agent development lifecycle, providing tools for building, testing, and deploying AI agents.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 20
    Paddler

    Paddler

    Open-source LLM load balancer and serving platform for hosting LLMs

    Paddler is an open-source LLM infrastructure platform designed to deploy, manage, and scale large language models on private infrastructure. The system acts as a specialized load balancer and serving layer for language models, enabling organizations to run inference workloads without relying on external API providers. It supports running models locally through engines such as llama.cpp while distributing requests across multiple compute nodes to improve performance and reliability. The...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    mlx

    mlx

    MLX: An array framework for Apple silicon

    MlX offers a local web interface to browse, download, and run ML models via Hugging Face or local sources. It supports searching by tags or tasks, visualization of model metadata, quick inference demos, automatic setup of runtime environments, and works with PyTorch, TensorFlow, and ONNX. Ideal for researchers exploring and testing models via browser.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 22
    NOFX

    NOFX

    Open source AI trading OS for autonomous multi-model trading systems

    NOFX is an open source AI-powered trading operating system designed to automate financial trading workflows using autonomous AI agents. It acts as an infrastructure layer that transforms market data into AI-driven trade decisions and execution. Instead of requiring users to manually configure machine learning models, data sources, and API integrations, the system allows AI components to perceive market conditions, select models, and perform trading actions automatically. It supports running...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 23
    Mosec

    Mosec

    A high-performance ML model serving framework, offers dynamic batching

    Mosec is a high-performance and flexible model-serving framework for building ML model-enabled backend and microservices. It bridges the gap between any machine learning models you just trained and the efficient online service API.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 24
    GLM-4.1V

    GLM-4.1V

    GLM-4.6V/4.5V/4.1V-Thinking, towards versatile multimodal reasoning

    ...Given its open-source availability under the same project repository, it provides an accessible entry point for testing multimodal reasoning and building proof-of-concept applications.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Made With ML

    Made With ML

    Learn how to develop, deploy and iterate on production-grade ML

    ...The project focuses on bridging the gap between experimental machine learning notebooks and real-world software systems that can be deployed, monitored, and maintained at scale. It provides structured lessons and practical code examples that demonstrate how to design machine learning workflows, manage datasets, train models, evaluate performance, and deploy inference services. The repository organizes these concepts into modular Python scripts that follow software engineering best practices such as testing, configuration management, logging, and version control. Through a combination of tutorials, notebooks, and production-ready scripts, the project demonstrates how machine learning applications should be developed as maintainable systems rather than isolated experiments.
    Downloads: 1 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • Next
MongoDB Logo MongoDB