Humiris AI
Humiris AI is a next-generation AI infrastructure platform that enables developers to build advanced applications by integrating multiple Large Language Models (LLMs). It offers a multi-LLM routing and reasoning layer, allowing users to optimize generative AI workflows with a flexible, scalable infrastructure. Humiris AI supports various use cases, including chatbot development, fine-tuning multiple LLMs simultaneously, retrieval-augmented generation, building super reasoning agents, advanced data analysis, and code generation. The platform's unique data format adapts to all foundation models, facilitating seamless integration and optimization. To get started, users can register for an account, create a project, add LLM provider API keys, and define parameters to generate a mixed model tailored to their specific needs. It allows deployment on users' own infrastructure, ensuring full data sovereignty and compliance with internal and external regulations.
Learn more
Entry Point AI
Entry Point AI is the modern AI optimization platform for proprietary and open source language models. Manage prompts, fine-tunes, and evals all in one place. When you reach the limits of prompt engineering, it’s time to fine-tune a model, and we make it easy. Fine-tuning is showing a model how to behave, not telling. It works together with prompt engineering and retrieval-augmented generation (RAG) to leverage the full potential of AI models. Fine-tuning can help you to get better quality from your prompts. Think of it like an upgrade to few-shot learning that bakes the examples into the model itself. For simpler tasks, you can train a lighter model to perform at or above the level of a higher-quality model, greatly reducing latency and cost. Train your model not to respond in certain ways to users, for safety, to protect your brand, and to get the formatting right. Cover edge cases and steer model behavior by adding examples to your dataset.
Learn more
StableVicuna
StableVicuna is the first large-scale open source chatbot trained via reinforced learning from human feedback (RHLF). StableVicuna is a further instruction fine tuned and RLHF trained version of Vicuna v0 13b, which is an instruction fine tuned LLaMA 13b model.
In order to achieve StableVicuna’s strong performance, we utilize Vicuna as the base model and follow the typical three-stage RLHF pipeline outlined by Steinnon et al. and Ouyang et al. Concretely, we further train the base Vicuna model with supervised finetuning (SFT) using a mixture of three datasets:
OpenAssistant Conversations Dataset (OASST1), a human-generated, human-annotated assistant-style conversation corpus comprising 161,443 messages distributed across 66,497 conversation trees, in 35 different languages;
GPT4All Prompt Generations, a dataset of 437,605 prompts and responses generated by GPT-3.5 Turbo;
And Alpaca, a dataset of 52,000 instructions and demonstrations generated by OpenAI's text-davinci-003.
Learn more
BGE
BGE (BAAI General Embedding) is a comprehensive retrieval toolkit designed for search and Retrieval-Augmented Generation (RAG) applications. It offers inference, evaluation, and fine-tuning capabilities for embedding models and rerankers, facilitating the development of advanced information retrieval systems. The toolkit includes components such as embedders and rerankers, which can be integrated into RAG pipelines to enhance search relevance and accuracy. BGE supports various retrieval methods, including dense retrieval, multi-vector retrieval, and sparse retrieval, providing flexibility to handle different data types and retrieval scenarios. The models are available through platforms like Hugging Face, and the toolkit provides tutorials and APIs to assist users in implementing and customizing their retrieval systems. By leveraging BGE, developers can build robust and efficient search solutions tailored to their specific needs.
Learn more