Best Artificial Intelligence Software for Hugging Face - Page 3

Compare the Top Artificial Intelligence Software that integrates with Hugging Face as of December 2025 - Page 3

This a list of Artificial Intelligence software that integrates with Hugging Face. Use the filters on the left to add additional filters for products that have integrations with Hugging Face. View the products that work with Hugging Face in the table below.

  • 1
    Axolotl

    Axolotl

    Axolotl

    ​Axolotl is an open source tool designed to streamline the fine-tuning of various AI models, offering support for multiple configurations and architectures. It enables users to train models, supporting methods like full fine-tuning, LoRA, QLoRA, ReLoRA, and GPTQ. Users can customize configurations using simple YAML files or command-line interface overrides, and load different dataset formats, including custom or pre-tokenized datasets. Axolotl integrates with technologies like xFormers, Flash Attention, Liger kernel, RoPE scaling, and multipacking, and works with single or multiple GPUs via Fully Sharded Data Parallel (FSDP) or DeepSpeed. It can be run locally or on the cloud using Docker and supports logging results and checkpoints to several platforms. It is designed to make fine-tuning AI models friendly, fast, and fun, without sacrificing functionality or scale.
    Starting Price: Free
  • 2
    Skott

    Skott

    Lyzr AI

    ​Skott is an AI marketing agent that autonomously researches, writes, and posts content, allowing your team to focus more on strategy and creative endeavors. It offers a customizable UI and workflow, providing actionable insights to guide your strategy, stay ahead of trends with real-time data, conduct in-depth competitive analysis, and gain audience insights to tailor your content effectively. Skott excels in stellar content creation by crafting high-impact blog posts, engaging social media content, SEO-optimized writing, and maintaining a consistent brand voice across all platforms. It ensures seamless publishing by allowing you to publish across multiple channels effortlessly, maintain consistent formatting and optimization, automate scheduling, and integrate with major blogging and social media platforms. Skott is cost-effective, offering affordable, high-quality marketing solutions that maximize your ROI without overspending or hiring additional resources.
    Starting Price: $99 per month
  • 3
    Mistral Small 3.1
    ​Mistral Small 3.1 is a state-of-the-art, multimodal, and multilingual AI model released under the Apache 2.0 license. Building upon Mistral Small 3, this enhanced version offers improved text performance, and advanced multimodal understanding, and supports an expanded context window of up to 128,000 tokens. It outperforms comparable models like Gemma 3 and GPT-4o Mini, delivering inference speeds of 150 tokens per second. Designed for versatility, Mistral Small 3.1 excels in tasks such as instruction following, conversational assistance, image understanding, and function calling, making it suitable for both enterprise and consumer-grade AI applications. Its lightweight architecture allows it to run efficiently on a single RTX 4090 or a Mac with 32GB RAM, facilitating on-device deployments. It is available for download on Hugging Face, accessible via Mistral AI's developer playground, and integrated into platforms like Google Cloud Vertex AI, with availability on NVIDIA NIM and
    Starting Price: Free
  • 4
    ML Console

    ML Console

    ML Console

    ​ML Console is a web-based application that enables users to build powerful machine learning models without writing a single line of code. Designed for accessibility, it allows individuals from various backgrounds, including marketing professionals, e-commerce store owners, and larger enterprises, to create AI models in less than a minute. It operates entirely within the user's browser, ensuring that data remains local and secure. By leveraging modern web technologies like WebAssembly and WebGL, ML Console achieves training speeds comparable to traditional Python-based methods. Its user-friendly interface simplifies the machine learning process, making it approachable for users with no advanced AI expertise. Additionally, ML Console is free to use, eliminating barriers to entry for those interested in exploring machine learning solutions. ​
    Starting Price: Free
  • 5
    Pruna AI

    Pruna AI

    Pruna AI

    Pruna uses generative AI to enable companies to produce professional-grade visual content quickly and affordably. By eliminating the traditional need for studios and manual editing, it empowers brands to create consistent, customized images for advertising, product displays, and digital campaigns with minimal effort.
    Starting Price: $0.40 per runtime hour
  • 6
    Hugging Face Transformers
    ​Transformers is a library of pretrained natural language processing, computer vision, audio, and multimodal models for inference and training. Use Transformers to train models on your data, build inference applications, and generate text with large language models. Explore the Hugging Face Hub today to find a model and use Transformers to help you get started right away.​ Simple and optimized inference class for many machine learning tasks like text generation, image segmentation, automatic speech recognition, document question answering, and more. A comprehensive trainer that supports features such as mixed precision, torch.compile, and FlashAttention for training and distributed training for PyTorch models.​ Fast text generation with large language models and vision language models. Every model is implemented from only three main classes (configuration, model, and preprocessor) and can be quickly used for inference or training.
    Starting Price: $9 per month
  • 7
    Qwen3

    Qwen3

    Alibaba

    Qwen3, the latest iteration of the Qwen family of large language models, introduces groundbreaking features that enhance performance across coding, math, and general capabilities. With models like the Qwen3-235B-A22B and Qwen3-30B-A3B, Qwen3 achieves impressive results compared to top-tier models, thanks to its hybrid thinking modes that allow users to control the balance between deep reasoning and quick responses. The platform supports 119 languages and dialects, making it an ideal choice for global applications. Its pre-training process, which uses 36 trillion tokens, enables robust performance, and advanced reinforcement learning (RL) techniques continue to refine its capabilities. Available on platforms like Hugging Face and ModelScope, Qwen3 offers a powerful tool for developers and researchers working in diverse fields.
    Starting Price: Free
  • 8
    Flower

    Flower

    Flower

    Flower is an open source federated learning framework designed to simplify the development and deployment of machine learning models across decentralized data sources. It enables training on data located on devices or servers without transferring the data itself, thereby enhancing privacy and reducing bandwidth usage. Flower supports a wide range of machine learning frameworks, including PyTorch, TensorFlow, Hugging Face Transformers, scikit-learn, and XGBoost, and is compatible with various platforms and cloud services like AWS, GCP, and Azure. It offers flexibility through customizable strategies and supports both horizontal and vertical federated learning scenarios. Flower's architecture allows for scalable experiments, with the capability to handle workloads involving tens of millions of clients. It also provides built-in support for privacy-preserving techniques like differential privacy and secure aggregation.
    Starting Price: Free
  • 9
    Open Computer Agent
    The Open Computer Agent is a browser-based AI assistant developed by Hugging Face that automates web interactions such as browsing, form-filling, and data retrieval. It leverages vision-language models like Qwen-VL to simulate mouse and keyboard actions, enabling tasks like booking tickets, checking store hours, and finding directions. Operating within a web browser, the agent can locate and interact with webpage elements using their image coordinates. As part of Hugging Face's smolagents project, it emphasizes flexibility and transparency, offering an open-source platform for developers to inspect, modify, and build upon for niche applications. While still in its early stages and facing challenges, the agent represents a new approach to AI as an active digital assistant, capable of performing online tasks without direct user input.
    Starting Price: Free
  • 10
    Devstral

    Devstral

    Mistral AI

    Devstral is an open source, agentic large language model (LLM) developed by Mistral AI in collaboration with All Hands AI, specifically designed for software engineering tasks. It excels at navigating complex codebases, editing multiple files, and resolving real-world issues, outperforming all open source models on the SWE-Bench Verified benchmark with a score of 46.8%. Devstral is fine-tuned from Mistral-Small-3.1 and features a long context window of up to 128,000 tokens. It is optimized for local deployment on high-end hardware, such as a Mac with 32GB RAM or an Nvidia RTX 4090 GPU, and is compatible with inference frameworks like vLLM, Transformers, and Ollama. Released under the Apache 2.0 license, Devstral is available for free and can be accessed via Hugging Face, Ollama, Kaggle, Unsloth, and LM Studio.
    Starting Price: $0.1 per million input tokens
  • 11
    BGE

    BGE

    BGE

    BGE (BAAI General Embedding) is a comprehensive retrieval toolkit designed for search and Retrieval-Augmented Generation (RAG) applications. It offers inference, evaluation, and fine-tuning capabilities for embedding models and rerankers, facilitating the development of advanced information retrieval systems. The toolkit includes components such as embedders and rerankers, which can be integrated into RAG pipelines to enhance search relevance and accuracy. BGE supports various retrieval methods, including dense retrieval, multi-vector retrieval, and sparse retrieval, providing flexibility to handle different data types and retrieval scenarios. The models are available through platforms like Hugging Face, and the toolkit provides tutorials and APIs to assist users in implementing and customizing their retrieval systems. By leveraging BGE, developers can build robust and efficient search solutions tailored to their specific needs.
    Starting Price: Free
  • 12
    Pinecone Rerank v0
    Pinecone Rerank V0 is a cross-encoder model optimized for precision in reranking tasks, enhancing enterprise search and retrieval-augmented generation (RAG) systems. It processes queries and documents together to capture fine-grained relevance, assigning a relevance score from 0 to 1 for each query-document pair. The model's maximum context length is set to 512 tokens to preserve ranking quality. Evaluations on the BEIR benchmark demonstrated that Pinecone Rerank V0 achieved the highest average NDCG@10, outperforming other models on 6 out of 12 datasets. For instance, it showed up to a 60% boost on the Fever dataset compared to Google Semantic Ranker and over 40% on the Climate-Fever dataset relative to cohere-v3-multilingual or voyageai-rerank-2. The model is accessible through Pinecone Inference and is available to all users in public preview.
    Starting Price: $25 per month
  • 13
    RankGPT

    RankGPT

    Weiwei Sun

    RankGPT is a Python toolkit designed to explore the use of generative Large Language Models (LLMs) like ChatGPT and GPT-4 for relevance ranking in Information Retrieval (IR). It introduces methods such as instructional permutation generation and a sliding window strategy to enable LLMs to effectively rerank documents. It supports various LLMs, including GPT-3.5, GPT-4, Claude, Cohere, and Llama2 via LiteLLM. RankGPT provides modules for retrieval, reranking, evaluation, and response analysis, facilitating end-to-end workflows. It includes a module for detailed analysis of input prompts and LLM responses, addressing reliability concerns with LLM APIs and non-deterministic behavior in Mixture-of-Experts (MoE) models. The toolkit supports various backends, including SGLang and TensorRT-LLM, and is compatible with a wide range of LLMs. RankGPT's Model Zoo includes models like LiT5 and MonoT5, hosted on Hugging Face.
    Starting Price: Free
  • 14
    HumanSignal

    HumanSignal

    HumanSignal

    HumanSignal's Label Studio Enterprise is a comprehensive platform designed for creating high-quality labeled data and evaluating model outputs with human supervision. It supports labeling and evaluating multi-modal data, image, video, audio, text, and time series, all in one place. It offers customizable labeling interfaces with pre-built templates and powerful plugins, allowing users to tailor the UI and workflows to specific use cases. Label Studio Enterprise integrates seamlessly with popular cloud storage providers and ML/AI models, facilitating pre-annotation, AI-assisted labeling, and prediction generation for model evaluation. The Prompts feature enables users to leverage LLMs to swiftly generate accurate predictions, enabling instant labeling of thousands of tasks. It supports various labeling use cases, including text classification, named entity recognition, sentiment analysis, summarization, and image captioning.
    Starting Price: $99 per month
  • 15
    FriendliAI

    FriendliAI

    FriendliAI

    FriendliAI is a generative AI infrastructure platform that offers fast, efficient, and reliable inference solutions for production environments. It provides a suite of tools and services designed to optimize the deployment and serving of large language models (LLMs) and other generative AI workloads at scale. Key offerings include Friendli Endpoints, which allow users to build and serve custom generative AI models, saving GPU costs and accelerating AI inference. It supports seamless integration with popular open source models from the Hugging Face Hub, enabling lightning-fast, high-performance inference. FriendliAI's cutting-edge technologies, such as Iteration Batching, Friendli DNN Library, Friendli TCache, and Native Quantization, contribute to significant cost savings (50–90%), reduced GPU requirements (6× fewer GPUs), higher throughput (10.7×), and lower latency (6.2×).
    Starting Price: $5.9 per hour
  • 16
    ZenCtrl

    ZenCtrl

    Fotographer AI

    ZenCtrl is an open source AI image generation toolkit developed by Fotographer AI, designed to produce high-quality, multi-view, and diverse-scene outputs from a single image without any training. It enables precise regeneration of objects and subjects from any angle and background, offering real-time element regeneration that provides both stability and flexibility in creative workflows. ZenCtrl allows users to regenerate subjects from any angle, swap backgrounds or clothing with just a click, and start generating results immediately without the need for additional training. By leveraging advanced image processing techniques, it ensures high accuracy without the need for extensive training data. The model's architecture is composed of lightweight sub-models, each fine-tuned on task-specific data to excel at a single job, resulting in a lean system that delivers sharper, more controllable results.
    Starting Price: Free
  • 17
    Bitext

    Bitext

    Bitext

    Bitext provides multilingual, hybrid synthetic training datasets specifically designed for intent detection and LLM fine‑tuning. These datasets blend large-scale synthetic text generation with expert curation and linguistic annotation, covering lexical, syntactic, semantic, register, and stylistic variation, to enhance conversational models’ understanding, accuracy, and domain adaptation. For example, their open source customer‑support dataset features ~27,000 question–answer pairs (≈3.57 million tokens), 27 intents across 10 categories, 30 entity types, and 12 language‑generation tags, all anonymized to comply with privacy, bias, and anti‑hallucination standards. Bitext also offers vertical-specific datasets (e.g., travel, banking) and supports over 20 industries in multiple languages with more than 95% accuracy. Their hybrid approach ensures scalable, multilingual training data, privacy-compliant, bias-mitigated, and ready for seamless LLM improvement and deployment.
    Starting Price: Free
  • 18
    Portia

    Portia

    Portia

    Portia AI is an open source developer framework (with optional cloud services) that lets teams rapidly build, deploy, and monitor stateful, authenticated AI agents with full visibility and control. Developers start by prompting the SDK to generate explicit, structured multi-step “plans” that weave together LLM reasoning and tool calls, then run those plans step-by-step, enriching plan state at each stage and pausing for clarifications (human or machine) whenever authentication or missing data is required. With its unified auth framework and plug-and-play tool catalog, Portia handles credentials and permissions for remote API and MCP tool invocations automatically. The complementary cloud offering adds persistent storage of plan run states, historical logs, telemetry dashboards, and managed scaling so production deployments stay reliable, auditable, and compliant in regulated environments.
    Starting Price: $30 per month
  • 19
    NuExtract

    NuExtract

    NuExtract

    NuExtract is a large language model specialized in extracting structured information from documents of any format, including raw text, scanned images, PDFs, PowerPoints, spreadsheets, and more, supporting over a dozen languages and mixed‑language inputs. It delivers JSON‑formatted output that faithfully follows user‑defined templates, with built‑in verification and null‑value handling to minimize hallucinations. Users define extraction tasks by creating a template, either by describing the desired fields or importing existing schemas—and can improve accuracy by adding document, output examples in the example set. The NuExtract Platform provides an intuitive workspace for designing templates, testing extractions in a playground, managing teaching examples, and fine‑tuning settings such as model temperature and document rasterization DPI. Once validated, projects can be deployed via a RESTful API endpoint that processes documents in real time.
    Starting Price: $5 per 1M tokens
  • 20
    AI SDK

    AI SDK

    AI SDK

    The AI SDK is a free, open source TypeScript toolkit from the creators of Next.js that gives developers unified, high-level primitives to build AI-powered features quickly across any model provider by changing a single line of code. It abstracts common complexities like streaming responses, multi-turn tool execution, error handling and recovery, and model switching while remaining framework-agnostic so builders can go from idea to working application in minutes. With a unified provider API, developers can generate typed objects, compose generative UIs, and deliver instant, streamed AI responses without reinventing plumbing, and the SDK includes documentation, cookbooks, a playground, and community-driven extensibility to accelerate development. It handles the hard parts under the hood while exposing enough control to get under the hood when needed, making integration with multiple LLMs seamless.
    Starting Price: Free
  • 21
    Qwen-Image

    Qwen-Image

    Alibaba

    Qwen-Image is a multimodal diffusion transformer (MMDiT) foundation model offering state-of-the-art image generation, text rendering, editing, and understanding. It excels at complex text integration, seamlessly embedding alphabetic and logographic scripts into visuals with typographic fidelity, and supports diverse artistic styles from photorealism to impressionism, anime, and minimalist design. Beyond creation, it enables advanced image editing operations such as style transfer, object insertion or removal, detail enhancement, in-image text editing, and human pose manipulation through intuitive prompts. Its built-in vision understanding tasks, including object detection, semantic segmentation, depth and edge estimation, novel view synthesis, and super-resolution, extend its capabilities into intelligent visual comprehension. Qwen-Image is accessible via popular libraries like Hugging Face Diffusers and integrates prompt-enhancement tools for multilingual support.
    Starting Price: Free
  • 22
    Anything

    Anything

    Anything

    Anything is an AI-powered, no-code app and website builder that transforms natural language prompts into fully functional projects without manual coding. The platform employs a mix of frontier and custom AI models to reason, plan actions, generate code, explore app behavior, and implement rapid updates, so you never look at source code. Anything includes built-in design tools for creating polished UX in seconds, instant development and production Postgres databases with 1 GB+ free per app, Stripe integration for subscriptions and one-time payments, and secure authentication via email, Google, Facebook, X, and more, all with zero setup or API keys. You can integrate the latest AI models (GPT-4o, GPT-3.5, Claude Sonnet 4, Gemini 2.5) through single prompts; generate image assets on demand; leverage web search for API documentation, references, and examples; and access hundreds of third-party APIs like Zapier and Google Maps.
    Starting Price: $16 per month
  • 23
    NVIDIA Cosmos
    NVIDIA Cosmos is a developer-first platform of state-of-the-art generative World Foundation Models (WFMs), advanced video tokenizers, guardrails, and an accelerated data processing and curation pipeline designed to supercharge physical AI development. It enables developers working on autonomous vehicles, robotics, and video analytics AI agents to generate photorealistic, physics-aware synthetic video data, trained on an immense dataset including 20 million hours of real-world and simulated video, to rapidly simulate future scenarios, train world models, and fine‑tune custom behaviors. It includes three core WFM types; Cosmos Predict, capable of generating up to 30 seconds of continuous video from multimodal inputs; Cosmos Transfer, which adapts simulations across environments and lighting for versatile domain augmentation; and Cosmos Reason, a vision-language model that applies structured reasoning to interpret spatial-temporal data for planning and decision-making.
    Starting Price: Free
  • 24
    DeepSeek V3.1
    DeepSeek V3.1 is a groundbreaking open-weight large language model featuring a massive 685-billion parameters and an extended 128,000‑token context window, enabling it to process documents equivalent to 400-page books in a single prompt. It delivers integrated capabilities for chat, reasoning, and code generation within a unified hybrid architecture, seamlessly blending these functions into one coherent model. V3.1 supports a variety of tensor formats to give developers flexibility in optimizing performance across different hardware. Early benchmark results show robust performance, including a 71.6% score on the Aider coding benchmark, putting it on par with or ahead of systems like Claude Opus 4 and doing so at a far lower cost. Made available under an open source license on Hugging Face with minimal fanfare, DeepSeek V3.1 is poised to reshape access to high-performance AI, challenging traditional proprietary models.
    Starting Price: Free
  • 25
    Google AI Edge Gallery
    Google AI Edge Gallery is an experimental, open source Android app that demonstrates on-device machine learning and generative AI use cases, letting users download and run models locally (so they work offline once installed). It offers several features including AI Chat (multi-turn conversation), Ask Image (upload or use images to ask questions, identify objects, get descriptions), Audio Scribe (transcribe or translate recorded/uploaded audio), Prompt Lab (for single-turn tasks such as summarization, rewriting, code generation), and performance insights (metrics like latency, decode speed, etc.). Users can switch between different compatible models (including Gemma 3n and models from Hugging Face), bring their own LiteRT models, and explore model cards and source code for transparency. The app aims to protect privacy by doing all processing on the device, no internet connection needed for core operations after models are loaded, reducing latency, and enhancing data security.
    Starting Price: Free
  • 26
    DeepSeek-V3.1-Terminus
    DeepSeek has released DeepSeek-V3.1-Terminus, which enhances the V3.1 architecture by incorporating user feedback to improve output stability, consistency, and agent performance. It notably reduces instances of mixed Chinese/English character output and unintended garbled characters, resulting in cleaner, more consistent language generation. The update upgrades both the code agent and search agent subsystems to yield stronger, more reliable performance across benchmarks. DeepSeek-V3.1-Terminus is also available as an open source model, and its weights are published on Hugging Face. The model structure remains the same as DeepSeek-V3, ensuring compatibility with existing deployment methods, with updated inference demos provided for community use. While trained at a scale of 685B parameters, the model includes FP8, BF16, and F32 tensor formats, offering flexibility across environments.
    Starting Price: Free
  • 27
    DeepSeek-V3.2-Exp
    Introducing DeepSeek-V3.2-Exp, our latest experimental model built on V3.1-Terminus, debuting DeepSeek Sparse Attention (DSA) for faster and more efficient inference and training on long contexts. DSA enables fine-grained sparse attention with minimal loss in output quality, boosting performance for long-context tasks while reducing compute costs. Benchmarks indicate that V3.2-Exp performs on par with V3.1-Terminus despite these efficiency gains. The model is now live across app, web, and API. Alongside this, the DeepSeek API prices have been cut by over 50% immediately to make access more affordable. For a transitional period, users can still access V3.1-Terminus via a temporary API endpoint until October 15, 2025. DeepSeek welcomes feedback on DSA via its feedback portal. In conjunction with the release, DeepSeek-V3.2-Exp has been open-sourced: the model weights and supporting technology (including key GPU kernels in TileLang and CUDA) are available on Hugging Face.
    Starting Price: Free
  • 28
    BotDojo

    BotDojo

    BotDojo

    BotDojo is an enterprise-grade AI enablement platform that empowers organizations to design, deploy, monitor, and scale intelligent agents across chat, voice, email, and web channels using a low-code visual workflow builder, while integrating deeply with enterprise data sources and systems. It provides over 100 ready-made templates to accelerate common use-cases (such as support automation, knowledge search, sales insights, and internal ops), supports branching logic, memory, tool orchestration (code, RPA, web browse), and connects to CRMs, ticketing systems, and databases. BotDojo also delivers human-feedback loops and continuous agent learning by enabling employees to coach agents via feedback queues, codifying corrections into memory and prompts, and evaluating performance through robust observability (audit trails, metrics such as deflection, first-contact resolution, and cost per interaction).
    Starting Price: $89 per month
  • 29
    Ekinox

    Ekinox

    Ekinox

    Ekinox is a visual AI automation platform that enables users to build, deploy, and manage AI-driven workflows without writing code; through its intuitive drag-and-drop canvas, you can design intelligent agents that connect to more than 100 pre-built integrations and trigger actions across a wide array of productivity, data, and communication tools. The platform supports real-time processing and collaboration, providing team workspaces, version control, and instant deployment. It features enterprise-grade security with SOC 2 compliance, bank-grade encryption, custom API connector support, and advanced access controls. Users can monitor workflows via analytics dashboards, track cost and performance across models and integrations, and scale with predictive auto-scaling and log retention. With setup times measured in minutes rather than months, Ekinox streamlines everything from simple task automation.
    Starting Price: $30 per month
  • 30
    Kimi K2 Thinking

    Kimi K2 Thinking

    Moonshot AI

    Kimi K2 Thinking is an advanced open source reasoning model developed by Moonshot AI, designed specifically for long-horizon, multi-step workflows where the system interleaves chain-of-thought processes with tool invocation across hundreds of sequential tasks. The model uses a mixture-of-experts architecture with a total of 1 trillion parameters, yet only about 32 billion parameters are activated per inference pass, optimizing efficiency while maintaining vast capacity. It supports a context window of up to 256,000 tokens, enabling the handling of extremely long inputs and reasoning chains without losing coherence. Native INT4 quantization is built in, which reduces inference latency and memory usage without performance degradation. Kimi K2 Thinking is explicitly built for agentic workflows; it can autonomously call external tools, manage sequential logic steps (up to and typically between 200-300 tool calls in a single chain), and maintain consistent reasoning.
    Starting Price: Free