Pezzo
Pezzo is the open-source LLMOps platform built for developers and teams. In just two lines of code, you can seamlessly troubleshoot and monitor your AI operations, collaborate and manage your prompts in one place, and instantly deploy changes to any environment.
Learn more
HoneyHive
AI engineering doesn't have to be a black box. Get full visibility with tools for tracing, evaluation, prompt management, and more. HoneyHive is an AI observability and evaluation platform designed to assist teams in building reliable generative AI applications. It offers tools for evaluating, testing, and monitoring AI models, enabling engineers, product managers, and domain experts to collaborate effectively. Measure quality over large test suites to identify improvements and regressions with each iteration. Track usage, feedback, and quality at scale, facilitating the identification of issues and driving continuous improvements. HoneyHive supports integration with various model providers and frameworks, offering flexibility and scalability to meet diverse organizational needs. It is suitable for teams aiming to ensure the quality and performance of their AI agents, providing a unified platform for evaluation, monitoring, and prompt management.
Learn more
Parea
The prompt engineering platform to experiment with different prompt versions, evaluate and compare prompts across a suite of tests, optimize prompts with one-click, share, and more. Optimize your AI development workflow. Key features to help you get and identify the best prompts for your production use cases. Side-by-side comparison of prompts across test cases with evaluation. CSV import test cases, and define custom evaluation metrics. Improve LLM results with automatic prompt and template optimization. View and manage all prompt versions and create OpenAI functions. Access all of your prompts programmatically, including observability and analytics. Determine the costs, latency, and efficacy of each prompt. Start enhancing your prompt engineering workflow with Parea today. Parea makes it easy for developers to improve the performance of their LLM apps through rigorous testing and version control.
Learn more
PromptHub
Test, collaborate, version, and deploy prompts, from a single place, with PromptHub. Put an end to continuous copy and pasting and utilize variables to simplify prompt creation. Say goodbye to spreadsheets, and easily compare outputs side-by-side when tweaking prompts. Bring your datasets and test prompts at scale with batch testing. Make sure your prompts are consistent by testing with different models, variables, and parameters. Stream two conversations and test different models, system messages, or chat templates. Commit prompts, create branches, and collaborate seamlessly. We detect prompt changes, so you can focus on outputs. Review changes as a team, approve new versions, and keep everyone on the same page. Easily monitor requests, costs, and latencies. PromptHub makes it easy to test, version, and collaborate on prompts with your team. Our GitHub-style versioning and collaboration makes it easy to iterate your prompts with your team, and store them in one place.
Learn more