Best Artificial Intelligence Software for Hugging Face - Page 2

Compare the Top Artificial Intelligence Software that integrates with Hugging Face as of December 2025 - Page 2

This a list of Artificial Intelligence software that integrates with Hugging Face. Use the filters on the left to add additional filters for products that have integrations with Hugging Face. View the products that work with Hugging Face in the table below.

  • 1
    AIxBlock

    AIxBlock

    AIxBlock

    AIxBlock: The first unified and decentralized platform for end-to-end AI development and workflow automation - built natively on MCP. AIxBlock is a MCP-based, decentralized end-to-end AI development and workflow automation platform purpose-built for AI engineer teams. It empowers users to build, train, deploy AI models and build AI automation workflows using those models through a unified environment that integrates decentralized compute, models, datasets, and labeling resources - all at a fraction of the traditional cost. AIxBlock is the modular AI ecosystem - purpose-built for custom model creation, workflow automation, and open interoperability across MCP client tools like Cursor, Claude, WindSurf, etc.
    Starting Price: $19 per month
  • 2
    Lamini

    Lamini

    Lamini

    Lamini makes it possible for enterprises to turn proprietary data into the next generation of LLM capabilities, by offering a platform for in-house software teams to uplevel to OpenAI-level AI teams and to build within the security of their existing infrastructure. Guaranteed structured output with optimized JSON decoding. Photographic memory through retrieval-augmented fine-tuning. Improve accuracy, and dramatically reduce hallucinations. Highly parallelized inference for large batch inference. Parameter-efficient finetuning that scales to millions of production adapters. Lamini is the only company that enables enterprise companies to safely and quickly develop and control their own LLMs anywhere. It brings several of the latest technologies and research to bear that was able to make ChatGPT from GPT-3, as well as Github Copilot from Codex. These include, among others, fine-tuning, RLHF, retrieval-augmented training, data augmentation, and GPU optimization.
    Starting Price: $99 per month
  • 3
    CodeQwen

    CodeQwen

    Alibaba

    CodeQwen is the code version of Qwen, the large language model series developed by the Qwen team, Alibaba Cloud. It is a transformer-based decoder-only language model pre-trained on a large amount of data of codes. Strong code generation capabilities and competitive performance across a series of benchmarks. Supporting long context understanding and generation with the context length of 64K tokens. CodeQwen supports 92 coding languages and provides excellent performance in text-to-SQL, bug fixes, etc. You can just write several lines of code with transformers to chat with CodeQwen. Essentially, we build the tokenizer and the model from pre-trained methods, and we use the generate method to perform chatting with the help of the chat template provided by the tokenizer. We apply the ChatML template for chat models following our previous practice. The model completes the code snippets according to the given prompts, without any additional formatting.
    Starting Price: Free
  • 4
    Agenta

    Agenta

    Agenta

    Collaborate on prompts, evaluate, and monitor LLM apps with confidence. Agenta is a comprehensive platform that enables teams to quickly build robust LLM apps. Create a playground connected to your code where the whole team can experiment and collaborate. Systematically compare different prompts, models, and embeddings before going to production. Share a link to gather human feedback from the rest of the team. Agenta works out of the box with all frameworks (Langchain, Lama Index, etc.) and model providers (OpenAI, Cohere, Huggingface, self-hosted models, etc.). Gain visibility into your LLM app's costs, latency, and chain of calls. You have the option to create simple LLM apps directly from the UI. However, if you would like to write customized applications, you need to write code with Python. Agenta is model agnostic and works with all model providers and frameworks. The only limitation at present is that our SDK is available only in Python.
    Starting Price: Free
  • 5
    OpenLIT

    OpenLIT

    OpenLIT

    OpenLIT is an OpenTelemetry-native application observability tool. It's designed to make the integration process of observability into AI projects with just a single line of code. Whether you're working with popular LLM libraries such as OpenAI and HuggingFace. OpenLIT's native support makes adding it to your projects feel effortless and intuitive. Analyze LLM and GPU performance, and costs to achieve maximum efficiency and scalability. Streams data to let you visualize your data and make quick decisions and modifications. Ensures that data is processed quickly without affecting the performance of your application. OpenLIT UI helps you explore LLM costs, token consumption, performance indicators, and user interactions in a straightforward interface. Connect to popular observability systems with ease, including Datadog and Grafana Cloud, to export data automatically. OpenLIT ensures your applications are monitored seamlessly.
    Starting Price: Free
  • 6
    Qwen2

    Qwen2

    Alibaba

    Qwen2 is the large language model series developed by Qwen team, Alibaba Cloud. Qwen2 is a series of large language models developed by the Qwen team at Alibaba Cloud. It includes both base language models and instruction-tuned models, ranging from 0.5 billion to 72 billion parameters, and features both dense models and a Mixture-of-Experts model. The Qwen2 series is designed to surpass most previous open-weight models, including its predecessor Qwen1.5, and to compete with proprietary models across a broad spectrum of benchmarks in language understanding, generation, multilingual capabilities, coding, mathematics, and reasoning.
    Starting Price: Free
  • 7
    Msty

    Msty

    Msty

    Chat with any AI model in a single click. No prior model setup experience is needed. Msty is designed to function seamlessly offline, ensuring reliability and privacy. For added flexibility, it also supports popular online model vendors, giving you the best of both worlds. Revolutionize your research with split chats. Compare and contrast multiple AI models' responses in real time, streamlining your workflow and uncovering new insights. Msty puts you in the driver's seat. Take your conversations wherever you want, and stop whenever you're satisfied. Replace an existing answer or create and iterate through several conversation branches. Delete branches that don't sound quite right. With delve mode, every response becomes a gateway to new knowledge, waiting to be discovered. Click on a keyword, and embark on a journey of discovery. Leverage Msty's split chat feature to move your desired conversation branches into a new split chat or a new chat session.
    Starting Price: $50 per year
  • 8
    Qwen2-VL

    Qwen2-VL

    Alibaba

    Qwen2-VL is the latest version of the vision language models based on Qwen2 in the Qwen model familities. Compared with Qwen-VL, Qwen2-VL has the capabilities of: SoTA understanding of images of various resolution & ratio: Qwen2-VL achieves state-of-the-art performance on visual understanding benchmarks, including MathVista, DocVQA, RealWorldQA, MTVQA, etc. Understanding videos of 20 min+: Qwen2-VL can understand videos over 20 minutes for high-quality video-based question answering, dialog, content creation, etc. Agent that can operate your mobiles, robots, etc.: with the abilities of complex reasoning and decision making, Qwen2-VL can be integrated with devices like mobile phones, robots, etc., for automatic operation based on visual environment and text instructions. Multilingual Support: to serve global users, besides English and Chinese, Qwen2-VL now supports the understanding of texts in different languages inside images
    Starting Price: Free
  • 9
    OpenHands

    OpenHands

    All Hands AI

    We believe agentic technology is too important to be controlled by a few corporations. So we're building all our agents in the open on GitHub, under the MIT license. Our agents can do anything a human developer can, they write code, run commands, and use the web. We're partnering with AI safety experts like Invariant Labs to balance innovation with security. Thousands of developers are working together to build the AI-powered future they want to see. Our agents are compatible with any large language model provider.
    Starting Price: Free
  • 10
    LLMWare.ai

    LLMWare.ai

    LLMWare.ai

    Our open source research efforts are focused both on the new "ware" ("middleware" and "software" that will wrap and integrate LLMs), as well as building high-quality, automation-focused enterprise models available in Hugging Face. LLMWare also provides a coherent, high-quality, integrated, and organized framework for development in an open system that provides the foundation for building LLM-applications for AI Agent workflows, Retrieval Augmented Generation (RAG), and other use cases, which include many of the core objects for developers to get started instantly. Our LLM framework is built from the ground up to handle the complex needs of data-sensitive enterprise use cases. Use our pre-built specialized LLMs for your industry or we can customize and fine-tune an LLM for specific use cases and domains. From a robust, integrated AI framework to specialized models and implementation, we provide an end-to-end solution.
    Starting Price: Free
  • 11
    ID Privacy AI

    ID Privacy AI

    ID Privacy AI

    At ID Privacy, we are shaping the future of AI with a focus on privacy-first solutions. Our mission is simple, to deliver cutting-edge AI technologies that empower businesses to innovate without compromising the security and trust of their users. ID Privacy AI delivers secure, adaptable AI models built with privacy at the core. We empower businesses across industries to harness advanced AI, whether optimizing workflows, enhancing customer AI chat experiences, or driving insights, while safeguarding data. Built under a cloak of stealth, the team at ID Privacy began meeting and formulating the plan for our AI as a service solution. Launched with multi-modal, multi-lingual capabilities and the deepest knowledge base on ad tech currently available anywhere. ID Privacy AI is focused on privacy-first AI development for businesses and enterprises. Empowering businesses with a flexible AI framework that protects data while solving complex challenges across any vertical.
    Starting Price: $15 per month
  • 12
    Maxim

    Maxim

    Maxim

    Maxim is an agent simulation, evaluation, and observability platform that empowers modern AI teams to deploy agents with quality, reliability, and speed. Maxim's end-to-end evaluation and data management stack covers every stage of the AI lifecycle, from prompt engineering to pre & post release testing and observability, data-set creation & management, and fine-tuning. Use Maxim to simulate and test your multi-turn workflows on a wide variety of scenarios and across different user personas before taking your application to production. Features: Agent Simulation Agent Evaluation Prompt Playground Logging/Tracing Workflows Custom Evaluators- AI, Programmatic and Statistical Dataset Curation Human-in-the-loop Use Case: Simulate and test AI agents Evals for agentic workflows: pre and post-release Tracing and debugging multi-agent workflows Real-time alerts on performance and quality Creating robust datasets for evals and fine-tuning Human-in-the-loop workflows
    Starting Price: $29/seat/month
  • 13
    Lunary

    Lunary

    Lunary

    Lunary is an AI developer platform designed to help AI teams manage, improve, and protect Large Language Model (LLM) chatbots. It offers features such as conversation and feedback tracking, analytics on costs and performance, debugging tools, and a prompt directory for versioning and team collaboration. Lunary supports integration with various LLMs and frameworks, including OpenAI and LangChain, and provides SDKs for Python and JavaScript. Guardrails to deflect malicious prompts and sensitive data leaks. Deploy in your VPC with Kubernetes or Docker. Allow your team to judge responses from your LLMs. Understand what languages your users are speaking. Experiment with prompts and LLM models. Search and filter anything in milliseconds. Receive notifications when agents are not performing as expected. Lunary's core platform is 100% open-source. Self-host or in the cloud, get started in minutes.
    Starting Price: $20 per month
  • 14
    DeepEval

    DeepEval

    Confident AI

    DeepEval is a simple-to-use, open source LLM evaluation framework, for evaluating and testing large-language model systems. It is similar to Pytest but specialized for unit testing LLM outputs. DeepEval incorporates the latest research to evaluate LLM outputs based on metrics such as G-Eval, hallucination, answer relevancy, RAGAS, etc., which uses LLMs and various other NLP models that run locally on your machine for evaluation. Whether your application is implemented via RAG or fine-tuning, LangChain, or LlamaIndex, DeepEval has you covered. With it, you can easily determine the optimal hyperparameters to improve your RAG pipeline, prevent prompt drifting, or even transition from OpenAI to hosting your own Llama2 with confidence. The framework supports synthetic dataset generation with advanced evolution techniques and integrates seamlessly with popular frameworks, allowing for efficient benchmarking and optimization of LLM systems.
    Starting Price: Free
  • 15
    Marco-o1

    Marco-o1

    AIDC-AI

    Marco-o1 is a robust, next-generation AI model tailored for high-performance natural language processing and real-time problem-solving. It is engineered to deliver precise and contextually rich responses, combining deep language comprehension with a streamlined architecture for speed and efficiency. Marco-o1 excels in a variety of applications, including conversational AI, content creation, technical support, and decision-making tasks, adapting seamlessly to diverse user needs. With a focus on intuitive interactions, reliability, and ethical AI principles, Marco-o1 stands out as a cutting-edge solution for individuals and organizations seeking intelligent, adaptive, and scalable AI-driven tools. MCTS allows the exploration of multiple reasoning paths using confidence scores derived from softmax-applied log probabilities of the top-k alternative tokens, guiding the model to optimal solutions.
    Starting Price: Free
  • 16
    Teuken 7B

    Teuken 7B

    OpenGPT-X

    Teuken-7B is a multilingual, open source language model developed under the OpenGPT-X initiative, specifically designed to cater to Europe's diverse linguistic landscape. It has been trained on a dataset comprising over 50% non-English texts, encompassing all 24 official languages of the European Union, ensuring robust performance across these languages. A key innovation in Teuken-7B is its custom multilingual tokenizer, optimized for European languages, which enhances training efficiency and reduces inference costs compared to standard monolingual tokenizers. The model is available in two versions, Teuken-7B-Base, the foundational pre-trained model, and Teuken-7B-Instruct, which has undergone instruction tuning for improved performance in following user prompts. Both versions are accessible on Hugging Face, promoting transparency and collaboration within the AI community. The development of Teuken-7B underscores a commitment to creating AI models that reflect Europe's diversity.
    Starting Price: Free
  • 17
    Qwen2.5-Coder
    Qwen2.5-Coder-32B-Instruct has become the current SOTA open source code model, matching the coding capabilities of GPT-4o. While demonstrating strong and comprehensive coding abilities, it also possesses good general and mathematical skills. As of now, Qwen2.5-Coder has covered six mainstream model sizes to meet the needs of different developers. We explore the practicality of Qwen2.5-Coder in two scenarios, including code assistants and artifacts, with some examples showcasing the potential applications of Qwen2.5-Coder in real-world scenarios. Qwen2.5-Coder-32B-Instruct, as the flagship model of this open source release, has achieved the best performance among open source models on multiple popular code generation benchmarks and has competitive performance with GPT-4o. Code repair is an important programming skill. Qwen2.5-Coder-32B-Instruct can help users fix errors in their code, making programming more efficient.
    Starting Price: Free
  • 18
    NVIDIA TensorRT
    NVIDIA TensorRT is an ecosystem of APIs for high-performance deep learning inference, encompassing an inference runtime and model optimizations that deliver low latency and high throughput for production applications. Built on the CUDA parallel programming model, TensorRT optimizes neural network models trained on all major frameworks, calibrating them for lower precision with high accuracy, and deploying them across hyperscale data centers, workstations, laptops, and edge devices. It employs techniques such as quantization, layer and tensor fusion, and kernel tuning on all types of NVIDIA GPUs, from edge devices to PCs to data centers. The ecosystem includes TensorRT-LLM, an open source library that accelerates and optimizes inference performance of recent large language models on the NVIDIA AI platform, enabling developers to experiment with new LLMs for high performance and quick customization through a simplified Python API.
    Starting Price: Free
  • 19
    SmythOS

    SmythOS

    SmythOS

    Say goodbye to manual coding and build agents faster than ever. Describe what you need, and SmythOS builds it from your chat or image, using the best AI models and APIs for your task. Use any AI model or API. Integrate with OpenAI, Hugging Face, Amazon Bedrock, and hundreds of vendors without a line of code. A pre-built agent template library gives you agents that already work out of the box for dozens of use cases. Just hit the button and connect with your own API keys. Because your marketing team should not have access to agents that work with your code. We got you covered. Create a space for each client, team, and project with full user and permission management. Deploy on-prem or to AWS. Integrate with Bedrock, Vertex, Adobe, Salesforce, etc. Explainable AI with full control over data flows, audit logs, encryption, and auth. Chat with your agents, give them bulk work, inspect their work logs, assign them work schedules, and more.
    Starting Price: $30 per month
  • 20
    Bakery

    Bakery

    Bakery

    Easily fine-tune & monetize your AI models with one click. For AI startups, ML engineers, and researchers. Bakery is a platform that enables AI startups, machine learning engineers, and researchers to fine-tune and monetize AI models with ease. Users can create or upload datasets, adjust model settings, and publish their models on the marketplace. The platform supports various model types and provides access to community-driven datasets for project development. Bakery's fine-tuning process is streamlined, allowing users to build, test, and deploy models efficiently. The platform integrates with tools like Hugging Face and supports decentralized storage solutions, ensuring flexibility and scalability for diverse AI projects. The bakery empowers contributors to collaboratively build AI models without exposing model parameters or data to one another. It ensures proper attribution and fair revenue distribution to all contributors.
    Starting Price: Free
  • 21
    Weave

    Weave

    Chasm

    Weave is a no-code AI workflow builder that enables users to automate tasks by implementing multiple Large Language Models (LLMs) and connecting prompts without the need for coding. With an intuitive interface, users can select templates, personalize them, and transform workflows into automated solutions. Weave supports various AI models, including those from OpenAI, Meta, Hugging Face, and Mistral AI, allowing for seamless integration and fine-tuning to achieve industry-specific results. Key features include intuitive dataflow management, app-ready APIs for easy integration, AI hosting, cost-effective AI models, effortless personalization, and user-friendly modules. Weave is ideal for applications such as generating character dialogue and backstories, developing intelligent chatbots, and automating written content.
    Starting Price: $10
  • 22
    FauxPilot

    FauxPilot

    FauxPilot

    FauxPilot is an open source, self-hosted alternative to GitHub Copilot. It utilizes the SalesForce CodeGen models on NVIDIA's Triton Inference Server with the FasterTransformer backend for local code generation. It requires Docker, an NVIDIA GPU with sufficient VRAM, and the ability to split the model across multiple GPUs if needed. The setup involves downloading models from Hugging Face and converting them for FasterTransformer compatibility.
    Starting Price: Free
  • 23
    Qwen2.5-Max
    Qwen2.5-Max is a large-scale Mixture-of-Experts (MoE) model developed by the Qwen team, pretrained on over 20 trillion tokens and further refined through Supervised Fine-Tuning (SFT) and Reinforcement Learning from Human Feedback (RLHF). In evaluations, it outperforms models like DeepSeek V3 in benchmarks such as Arena-Hard, LiveBench, LiveCodeBench, and GPQA-Diamond, while also demonstrating competitive results in other assessments, including MMLU-Pro. Qwen2.5-Max is accessible via API through Alibaba Cloud and can be explored interactively on Qwen Chat.
    Starting Price: Free
  • 24
    Qwen2.5-VL

    Qwen2.5-VL

    Alibaba

    Qwen2.5-VL is the latest vision-language model from the Qwen series, representing a significant advancement over its predecessor, Qwen2-VL. This model excels in visual understanding, capable of recognizing a wide array of objects, including text, charts, icons, graphics, and layouts within images. It functions as a visual agent, capable of reasoning and dynamically directing tools, enabling applications such as computer and phone usage. Qwen2.5-VL can comprehend videos exceeding one hour in length and can pinpoint relevant segments within them. Additionally, it accurately localizes objects in images by generating bounding boxes or points and provides stable JSON outputs for coordinates and attributes. The model also supports structured outputs for data like scanned invoices, forms, and tables, benefiting sectors such as finance and commerce. Available in base and instruct versions across 3B, 7B, and 72B sizes, Qwen2.5-VL is accessible through platforms like Hugging Face and ModelScope.
    Starting Price: Free
  • 25
    Zyphra Zonos
    Zyphra is excited to announce the release of Zonos-v0.1 beta, featuring two expressive and real-time text-to-speech models with high-fidelity voice cloning. We are releasing our 1.6B transformer and 1.6B hybrid under an Apache 2.0 license. It is difficult to quantitatively measure quality in the audio domain; we find that Zonos’ generation quality matches or exceeds that of leading proprietary TTS model providers. Further, we believe that openly releasing models of this caliber will significantly advance TTS research. Zonos model weights are available on Huggingface, and sample inference code for the models is available on our GitHub. You can also access Zonos through our model playground and API with simple and competitive flat-rate pricing. We have found that quantitative evaluations struggle to measure the quality of outputs in the audio domain, so for demonstration, we present a number of samples of Zonos vs both proprietary models.
    Starting Price: $0.02 per minute
  • 26
    txtai

    txtai

    NeuML

    txtai is an all-in-one open source embeddings database designed for semantic search, large language model orchestration, and language model workflows. It unifies vector indexes (both sparse and dense), graph networks, and relational databases, providing a robust foundation for vector search and serving as a powerful knowledge source for LLM applications. With txtai, users can build autonomous agents, implement retrieval augmented generation processes, and develop multi-modal workflows. Key features include vector search with SQL support, object storage integration, topic modeling, graph analysis, and multimodal indexing capabilities. It supports the creation of embeddings for various data types, including text, documents, audio, images, and video. Additionally, txtai offers pipelines powered by language models that handle tasks such as LLM prompting, question-answering, labeling, transcription, translation, and summarization.
    Starting Price: Free
  • 27
    Patched

    Patched

    Patched

    Patched is a managed service that leverages the open-source framework Patchwork to automate development tasks such as code reviews, bug fixing, security patching, and documentation. By utilizing large language models, Patched enables developers to build and deploy AI-assisted workflow, referred to as "patch flows", that autonomously handle post-code activities, thereby enhancing code quality and accelerating development cycles. The platform offers a user-friendly graphical interface and a visual workflow builder, allowing for the customization of patch flows without the need to manage infrastructure or LLM endpoints. For those who prefer self-hosting, Patchwork provides a self-hosted command-line interface agent that integrates seamlessly with existing development pipelines. Patched emphasizes privacy and control, enabling deployment within an organization's infrastructure using its own LLM API keys.
    Starting Price: $99 per month
  • 28
    SmolLM2

    SmolLM2

    Hugging Face

    SmolLM2 is a collection of state-of-the-art, compact language models developed for on-device applications. The models in this collection range from 1.7B parameters to smaller 360M and 135M versions, designed to perform efficiently even on less powerful hardware. These models excel in text generation tasks and are optimized for real-time, low-latency applications, providing high-quality results across various use cases, including content creation, coding assistance, and natural language processing. SmolLM2's flexibility makes it a suitable choice for developers looking to integrate powerful AI into mobile devices, edge computing, and other resource-constrained environments.
    Starting Price: Free
  • 29
    LiteLLM

    LiteLLM

    LiteLLM

    ​LiteLLM is a versatile platform designed to streamline interactions with over 100 Large Language Models (LLMs) through a unified interface. It offers both a Proxy Server (LLM Gateway) and a Python SDK, enabling developers to integrate various LLMs seamlessly into their applications. The Proxy Server facilitates centralized management, allowing for load balancing, cost tracking across projects, and consistent input/output formatting compatible with OpenAI standards. This setup supports multiple providers. It ensures robust observability by generating unique call IDs for each request, aiding in precise tracking and logging across systems. Developers can leverage pre-defined callbacks to log data using various tools. For enterprise users, LiteLLM offers advanced features like Single Sign-On (SSO), user management, and professional support through dedicated channels like Discord and Slack.
    Starting Price: Free
  • 30
    EigentBot

    EigentBot

    EigentBot

    ​EigentBot is an intelligent agent solution that integrates Retrieval-Augmented Generation (RAG) capabilities and function-call features. This design enables EigentBot to efficiently process user inputs, access relevant information, and execute functions to provide accurate and context-aware responses. By leveraging these advanced technologies, EigentBot aims to deliver enhanced user experiences across various platforms. ​The easiest way to build a secure & efficient AI knowledge base in just 5 seconds, perfect for streamlining customer service, technical QA, and more. Easily switch between different AI providers without disruption, keeping your AI assistant up-to-date with the best models available. Eigentbot continuously updates itself with the latest data from sources like Notion, GitHub, and Google Scholar. Enhance AI retrieval accuracy with structured, visualized knowledge graphs for better contextual understanding.
    Starting Price: $8 per month