Compare the Top AI Models that integrate with Hugging Face as of November 2025

This a list of AI Models that integrate with Hugging Face. Use the filters on the left to add additional filters for products that have integrations with Hugging Face. View the products that work with Hugging Face in the table below.

What are AI Models for Hugging Face?

AI models are systems designed to simulate human intelligence by learning from data and solving complex tasks. They include specialized types like Large Language Models (LLMs) for text generation, image models for visual recognition and editing, and video models for processing and analyzing dynamic content. These models power applications such as chatbots, facial recognition, video summarization, and personalized recommendations. Their capabilities rely on advanced algorithms, extensive training datasets, and robust computational resources. AI models are transforming industries by automating processes, enhancing decision-making, and enabling creative innovations. Compare and read user reviews of the best AI Models for Hugging Face currently available using the table below. This list is updated regularly.

  • 1
    Qwen

    Qwen

    Alibaba

    Qwen LLM refers to a family of large language models (LLMs) developed by Alibaba Cloud's Damo Academy. These models are trained on a massive dataset of text and code, allowing them to understand and generate human-like text, translate languages, write different kinds of creative content, and answer your questions in an informative way. Here are some key features of Qwen LLMs: Variety of sizes: The Qwen series ranges from 1.8 billion to 72 billion parameters, offering options for different needs and performance levels. Open source: Some versions of Qwen are open-source, which means their code is publicly available for anyone to use and modify. Multilingual support: Qwen can understand and translate multiple languages, including English, Chinese, and French. Diverse capabilities: Besides generation and translation, Qwen models can be used for tasks like question answering, text summarization, and code generation.
    Starting Price: Free
  • 2
    Qwen-7B

    Qwen-7B

    Alibaba

    Qwen-7B is the 7B-parameter version of the large language model series, Qwen (abbr. Tongyi Qianwen), proposed by Alibaba Cloud. Qwen-7B is a Transformer-based large language model, which is pretrained on a large volume of data, including web texts, books, codes, etc. Additionally, based on the pretrained Qwen-7B, we release Qwen-7B-Chat, a large-model-based AI assistant, which is trained with alignment techniques. The features of the Qwen-7B series include: Trained with high-quality pretraining data. We have pretrained Qwen-7B on a self-constructed large-scale high-quality dataset of over 2.2 trillion tokens. The dataset includes plain texts and codes, and it covers a wide range of domains, including general domain data and professional domain data. Strong performance. In comparison with the models of the similar model size, we outperform the competitors on a series of benchmark datasets, which evaluates natural language understanding, mathematics, coding, etc. And more.
    Starting Price: Free
  • 3
    Codestral Mamba
    As a tribute to Cleopatra, whose glorious destiny ended in tragic snake circumstances, we are proud to release Codestral Mamba, a Mamba2 language model specialized in code generation, available under an Apache 2.0 license. Codestral Mamba is another step in our effort to study and provide new architectures. It is available for free use, modification, and distribution, and we hope it will open new perspectives in architecture research. Mamba models offer the advantage of linear time inference and the theoretical ability to model sequences of infinite length. It allows users to engage with the model extensively with quick responses, irrespective of the input length. This efficiency is especially relevant for code productivity use cases, this is why we trained this model with advanced code and reasoning capabilities, enabling it to perform on par with SOTA transformer-based models.
    Starting Price: Free
  • 4
    Mathstral

    Mathstral

    Mistral AI

    As a tribute to Archimedes, whose 2311th anniversary we’re celebrating this year, we are proud to release our first Mathstral model, a specific 7B model designed for math reasoning and scientific discovery. The model has a 32k context window published under the Apache 2.0 license. We’re contributing Mathstral to the science community to bolster efforts in advanced mathematical problems requiring complex, multi-step logical reasoning. The Mathstral release is part of our broader effort to support academic projects, it was produced in the context of our collaboration with Project Numina. Akin to Isaac Newton in his time, Mathstral stands on the shoulders of Mistral 7B and specializes in STEM subjects. It achieves state-of-the-art reasoning capacities in its size category across various industry-standard benchmarks. In particular, it achieves 56.6% on MATH and 63.47% on MMLU, with the following MMLU performance difference by subject between Mathstral 7B and Mistral 7B.
    Starting Price: Free
  • 5
    Qwen2.5

    Qwen2.5

    Alibaba

    Qwen2.5 is an advanced multimodal AI model designed to provide highly accurate and context-aware responses across a wide range of applications. It builds on the capabilities of its predecessors, integrating cutting-edge natural language understanding with enhanced reasoning, creativity, and multimodal processing. Qwen2.5 can seamlessly analyze and generate text, interpret images, and interact with complex data to deliver precise solutions in real time. Optimized for adaptability, it excels in personalized assistance, data analysis, creative content generation, and academic research, making it a versatile tool for professionals and everyday users alike. Its user-centric design emphasizes transparency, efficiency, and alignment with ethical AI practices.
    Starting Price: Free
  • 6
    LTXV

    LTXV

    Lightricks

    LTXV offers a suite of AI-powered creative tools designed to empower content creators across various platforms. LTX provides AI-driven video generation capabilities, allowing users to craft detailed video sequences with full control over every stage of production. It leverages Lightricks' proprietary AI models to deliver high-quality, efficient, and user-friendly editing experiences. LTX Video uses a breakthrough called multiscale rendering, starting with fast, low-res passes to capture motion and lighting, then refining with high-res detail. Unlike traditional upscalers, LTXV-13B analyzes motion over time, front-loading the heavy computation to deliver up to 30× faster, high-quality renders.
    Starting Price: Free
  • 7
    CodeQwen

    CodeQwen

    Alibaba

    CodeQwen is the code version of Qwen, the large language model series developed by the Qwen team, Alibaba Cloud. It is a transformer-based decoder-only language model pre-trained on a large amount of data of codes. Strong code generation capabilities and competitive performance across a series of benchmarks. Supporting long context understanding and generation with the context length of 64K tokens. CodeQwen supports 92 coding languages and provides excellent performance in text-to-SQL, bug fixes, etc. You can just write several lines of code with transformers to chat with CodeQwen. Essentially, we build the tokenizer and the model from pre-trained methods, and we use the generate method to perform chatting with the help of the chat template provided by the tokenizer. We apply the ChatML template for chat models following our previous practice. The model completes the code snippets according to the given prompts, without any additional formatting.
    Starting Price: Free
  • 8
    Qwen2

    Qwen2

    Alibaba

    Qwen2 is the large language model series developed by Qwen team, Alibaba Cloud. Qwen2 is a series of large language models developed by the Qwen team at Alibaba Cloud. It includes both base language models and instruction-tuned models, ranging from 0.5 billion to 72 billion parameters, and features both dense models and a Mixture-of-Experts model. The Qwen2 series is designed to surpass most previous open-weight models, including its predecessor Qwen1.5, and to compete with proprietary models across a broad spectrum of benchmarks in language understanding, generation, multilingual capabilities, coding, mathematics, and reasoning.
    Starting Price: Free
  • 9
    Qwen2-VL

    Qwen2-VL

    Alibaba

    Qwen2-VL is the latest version of the vision language models based on Qwen2 in the Qwen model familities. Compared with Qwen-VL, Qwen2-VL has the capabilities of: SoTA understanding of images of various resolution & ratio: Qwen2-VL achieves state-of-the-art performance on visual understanding benchmarks, including MathVista, DocVQA, RealWorldQA, MTVQA, etc. Understanding videos of 20 min+: Qwen2-VL can understand videos over 20 minutes for high-quality video-based question answering, dialog, content creation, etc. Agent that can operate your mobiles, robots, etc.: with the abilities of complex reasoning and decision making, Qwen2-VL can be integrated with devices like mobile phones, robots, etc., for automatic operation based on visual environment and text instructions. Multilingual Support: to serve global users, besides English and Chinese, Qwen2-VL now supports the understanding of texts in different languages inside images
    Starting Price: Free
  • 10
    Marco-o1

    Marco-o1

    AIDC-AI

    Marco-o1 is a robust, next-generation AI model tailored for high-performance natural language processing and real-time problem-solving. It is engineered to deliver precise and contextually rich responses, combining deep language comprehension with a streamlined architecture for speed and efficiency. Marco-o1 excels in a variety of applications, including conversational AI, content creation, technical support, and decision-making tasks, adapting seamlessly to diverse user needs. With a focus on intuitive interactions, reliability, and ethical AI principles, Marco-o1 stands out as a cutting-edge solution for individuals and organizations seeking intelligent, adaptive, and scalable AI-driven tools. MCTS allows the exploration of multiple reasoning paths using confidence scores derived from softmax-applied log probabilities of the top-k alternative tokens, guiding the model to optimal solutions.
    Starting Price: Free
  • 11
    Teuken 7B

    Teuken 7B

    OpenGPT-X

    Teuken-7B is a multilingual, open source language model developed under the OpenGPT-X initiative, specifically designed to cater to Europe's diverse linguistic landscape. It has been trained on a dataset comprising over 50% non-English texts, encompassing all 24 official languages of the European Union, ensuring robust performance across these languages. A key innovation in Teuken-7B is its custom multilingual tokenizer, optimized for European languages, which enhances training efficiency and reduces inference costs compared to standard monolingual tokenizers. The model is available in two versions, Teuken-7B-Base, the foundational pre-trained model, and Teuken-7B-Instruct, which has undergone instruction tuning for improved performance in following user prompts. Both versions are accessible on Hugging Face, promoting transparency and collaboration within the AI community. The development of Teuken-7B underscores a commitment to creating AI models that reflect Europe's diversity.
    Starting Price: Free
  • 12
    Qwen2.5-Max
    Qwen2.5-Max is a large-scale Mixture-of-Experts (MoE) model developed by the Qwen team, pretrained on over 20 trillion tokens and further refined through Supervised Fine-Tuning (SFT) and Reinforcement Learning from Human Feedback (RLHF). In evaluations, it outperforms models like DeepSeek V3 in benchmarks such as Arena-Hard, LiveBench, LiveCodeBench, and GPQA-Diamond, while also demonstrating competitive results in other assessments, including MMLU-Pro. Qwen2.5-Max is accessible via API through Alibaba Cloud and can be explored interactively on Qwen Chat.
    Starting Price: Free
  • 13
    Qwen2.5-VL

    Qwen2.5-VL

    Alibaba

    Qwen2.5-VL is the latest vision-language model from the Qwen series, representing a significant advancement over its predecessor, Qwen2-VL. This model excels in visual understanding, capable of recognizing a wide array of objects, including text, charts, icons, graphics, and layouts within images. It functions as a visual agent, capable of reasoning and dynamically directing tools, enabling applications such as computer and phone usage. Qwen2.5-VL can comprehend videos exceeding one hour in length and can pinpoint relevant segments within them. Additionally, it accurately localizes objects in images by generating bounding boxes or points and provides stable JSON outputs for coordinates and attributes. The model also supports structured outputs for data like scanned invoices, forms, and tables, benefiting sectors such as finance and commerce. Available in base and instruct versions across 3B, 7B, and 72B sizes, Qwen2.5-VL is accessible through platforms like Hugging Face and ModelScope.
    Starting Price: Free
  • 14
    Zyphra Zonos
    Zyphra is excited to announce the release of Zonos-v0.1 beta, featuring two expressive and real-time text-to-speech models with high-fidelity voice cloning. We are releasing our 1.6B transformer and 1.6B hybrid under an Apache 2.0 license. It is difficult to quantitatively measure quality in the audio domain; we find that Zonos’ generation quality matches or exceeds that of leading proprietary TTS model providers. Further, we believe that openly releasing models of this caliber will significantly advance TTS research. Zonos model weights are available on Huggingface, and sample inference code for the models is available on our GitHub. You can also access Zonos through our model playground and API with simple and competitive flat-rate pricing. We have found that quantitative evaluations struggle to measure the quality of outputs in the audio domain, so for demonstration, we present a number of samples of Zonos vs both proprietary models.
    Starting Price: $0.02 per minute
  • 15
    SmolLM2

    SmolLM2

    Hugging Face

    SmolLM2 is a collection of state-of-the-art, compact language models developed for on-device applications. The models in this collection range from 1.7B parameters to smaller 360M and 135M versions, designed to perform efficiently even on less powerful hardware. These models excel in text generation tasks and are optimized for real-time, low-latency applications, providing high-quality results across various use cases, including content creation, coding assistance, and natural language processing. SmolLM2's flexibility makes it a suitable choice for developers looking to integrate powerful AI into mobile devices, edge computing, and other resource-constrained environments.
    Starting Price: Free
  • 16
    Mistral Small 3.1
    ​Mistral Small 3.1 is a state-of-the-art, multimodal, and multilingual AI model released under the Apache 2.0 license. Building upon Mistral Small 3, this enhanced version offers improved text performance, and advanced multimodal understanding, and supports an expanded context window of up to 128,000 tokens. It outperforms comparable models like Gemma 3 and GPT-4o Mini, delivering inference speeds of 150 tokens per second. Designed for versatility, Mistral Small 3.1 excels in tasks such as instruction following, conversational assistance, image understanding, and function calling, making it suitable for both enterprise and consumer-grade AI applications. Its lightweight architecture allows it to run efficiently on a single RTX 4090 or a Mac with 32GB RAM, facilitating on-device deployments. It is available for download on Hugging Face, accessible via Mistral AI's developer playground, and integrated into platforms like Google Cloud Vertex AI, with availability on NVIDIA NIM and
    Starting Price: Free
  • 17
    Qwen3

    Qwen3

    Alibaba

    Qwen3, the latest iteration of the Qwen family of large language models, introduces groundbreaking features that enhance performance across coding, math, and general capabilities. With models like the Qwen3-235B-A22B and Qwen3-30B-A3B, Qwen3 achieves impressive results compared to top-tier models, thanks to its hybrid thinking modes that allow users to control the balance between deep reasoning and quick responses. The platform supports 119 languages and dialects, making it an ideal choice for global applications. Its pre-training process, which uses 36 trillion tokens, enables robust performance, and advanced reinforcement learning (RL) techniques continue to refine its capabilities. Available on platforms like Hugging Face and ModelScope, Qwen3 offers a powerful tool for developers and researchers working in diverse fields.
    Starting Price: Free
  • 18
    Devstral

    Devstral

    Mistral AI

    Devstral is an open source, agentic large language model (LLM) developed by Mistral AI in collaboration with All Hands AI, specifically designed for software engineering tasks. It excels at navigating complex codebases, editing multiple files, and resolving real-world issues, outperforming all open source models on the SWE-Bench Verified benchmark with a score of 46.8%. Devstral is fine-tuned from Mistral-Small-3.1 and features a long context window of up to 128,000 tokens. It is optimized for local deployment on high-end hardware, such as a Mac with 32GB RAM or an Nvidia RTX 4090 GPU, and is compatible with inference frameworks like vLLM, Transformers, and Ollama. Released under the Apache 2.0 license, Devstral is available for free and can be accessed via Hugging Face, Ollama, Kaggle, Unsloth, and LM Studio.
    Starting Price: $0.1 per million input tokens
  • 19
    ZenCtrl

    ZenCtrl

    Fotographer AI

    ZenCtrl is an open source AI image generation toolkit developed by Fotographer AI, designed to produce high-quality, multi-view, and diverse-scene outputs from a single image without any training. It enables precise regeneration of objects and subjects from any angle and background, offering real-time element regeneration that provides both stability and flexibility in creative workflows. ZenCtrl allows users to regenerate subjects from any angle, swap backgrounds or clothing with just a click, and start generating results immediately without the need for additional training. By leveraging advanced image processing techniques, it ensures high accuracy without the need for extensive training data. The model's architecture is composed of lightweight sub-models, each fine-tuned on task-specific data to excel at a single job, resulting in a lean system that delivers sharper, more controllable results.
    Starting Price: Free
  • 20
    Qwen-Image

    Qwen-Image

    Alibaba

    Qwen-Image is a multimodal diffusion transformer (MMDiT) foundation model offering state-of-the-art image generation, text rendering, editing, and understanding. It excels at complex text integration, seamlessly embedding alphabetic and logographic scripts into visuals with typographic fidelity, and supports diverse artistic styles from photorealism to impressionism, anime, and minimalist design. Beyond creation, it enables advanced image editing operations such as style transfer, object insertion or removal, detail enhancement, in-image text editing, and human pose manipulation through intuitive prompts. Its built-in vision understanding tasks, including object detection, semantic segmentation, depth and edge estimation, novel view synthesis, and super-resolution, extend its capabilities into intelligent visual comprehension. Qwen-Image is accessible via popular libraries like Hugging Face Diffusers and integrates prompt-enhancement tools for multilingual support.
    Starting Price: Free
  • 21
    NVIDIA Cosmos
    NVIDIA Cosmos is a developer-first platform of state-of-the-art generative World Foundation Models (WFMs), advanced video tokenizers, guardrails, and an accelerated data processing and curation pipeline designed to supercharge physical AI development. It enables developers working on autonomous vehicles, robotics, and video analytics AI agents to generate photorealistic, physics-aware synthetic video data, trained on an immense dataset including 20 million hours of real-world and simulated video, to rapidly simulate future scenarios, train world models, and fine‑tune custom behaviors. It includes three core WFM types; Cosmos Predict, capable of generating up to 30 seconds of continuous video from multimodal inputs; Cosmos Transfer, which adapts simulations across environments and lighting for versatile domain augmentation; and Cosmos Reason, a vision-language model that applies structured reasoning to interpret spatial-temporal data for planning and decision-making.
    Starting Price: Free
  • 22
    DeepSeek V3.1
    DeepSeek V3.1 is a groundbreaking open-weight large language model featuring a massive 685-billion parameters and an extended 128,000‑token context window, enabling it to process documents equivalent to 400-page books in a single prompt. It delivers integrated capabilities for chat, reasoning, and code generation within a unified hybrid architecture, seamlessly blending these functions into one coherent model. V3.1 supports a variety of tensor formats to give developers flexibility in optimizing performance across different hardware. Early benchmark results show robust performance, including a 71.6% score on the Aider coding benchmark, putting it on par with or ahead of systems like Claude Opus 4 and doing so at a far lower cost. Made available under an open source license on Hugging Face with minimal fanfare, DeepSeek V3.1 is poised to reshape access to high-performance AI, challenging traditional proprietary models.
    Starting Price: Free
  • 23
    DeepSeek-V3.1-Terminus
    DeepSeek has released DeepSeek-V3.1-Terminus, which enhances the V3.1 architecture by incorporating user feedback to improve output stability, consistency, and agent performance. It notably reduces instances of mixed Chinese/English character output and unintended garbled characters, resulting in cleaner, more consistent language generation. The update upgrades both the code agent and search agent subsystems to yield stronger, more reliable performance across benchmarks. DeepSeek-V3.1-Terminus is also available as an open source model, and its weights are published on Hugging Face. The model structure remains the same as DeepSeek-V3, ensuring compatibility with existing deployment methods, with updated inference demos provided for community use. While trained at a scale of 685B parameters, the model includes FP8, BF16, and F32 tensor formats, offering flexibility across environments.
    Starting Price: Free
  • 24
    HunyuanVideo
    HunyuanVideo is an advanced AI-powered video generation model developed by Tencent, designed to seamlessly blend virtual and real elements, offering limitless creative possibilities. It delivers cinematic-quality videos with natural movements and precise expressions, capable of transitioning effortlessly between realistic and virtual styles. This technology overcomes the constraints of short dynamic images by presenting complete, fluid actions and rich semantic content, making it ideal for applications in advertising, film production, and other commercial industries.
  • 25
    Qwen2.5-1M

    Qwen2.5-1M

    Alibaba

    Qwen2.5-1M is an open-source language model developed by the Qwen team, designed to handle context lengths of up to one million tokens. This release includes two model variants, Qwen2.5-7B-Instruct-1M and Qwen2.5-14B-Instruct-1M, marking the first time Qwen models have been upgraded to support such extensive context lengths. To facilitate efficient deployment, the team has also open-sourced an inference framework based on vLLM, integrated with sparse attention methods, enabling processing of 1M-token inputs with a 3x to 7x speed improvement. Comprehensive technical details, including design insights and ablation experiments, are available in the accompanying technical report.
    Starting Price: Free
  • 26
    Yi-Large
    Yi-Large is a proprietary large language model developed by 01.AI, offering a 32k context length with both input and output costs at $2 per million tokens. It stands out with its advanced capabilities in natural language processing, common-sense reasoning, and multilingual support, performing on par with leading models like GPT-4 and Claude3 in various benchmarks. Yi-Large is designed for tasks requiring complex inference, prediction, and language understanding, making it suitable for applications like knowledge search, data classification, and creating human-like chatbots. Its architecture is based on a decoder-only transformer with enhancements such as pre-normalization and Group Query Attention, and it has been trained on a vast, high-quality multilingual dataset. This model's versatility and cost-efficiency make it a strong contender in the AI market, particularly for enterprises aiming to deploy AI solutions globally.
    Starting Price: $0.19 per 1M input token
  • 27
    Gemma 3

    Gemma 3

    Google

    Gemma 3, introduced by Google, is a new AI model built on the Gemini 2.0 architecture, designed to offer enhanced performance and versatility. This model is capable of running efficiently on a single GPU or TPU, making it accessible for a wide range of developers and researchers. Gemma 3 focuses on improving natural language understanding, generation, and other AI-driven tasks. By offering scalable, powerful AI capabilities, Gemma 3 aims to advance the development of AI systems across various industries and use cases.
    Starting Price: Free
  • 28
    Gemma 3n

    Gemma 3n

    Google DeepMind

    Gemma 3n is our state-of-the-art open multimodal model, engineered for on-device performance and efficiency. Made for responsive, low-footprint local inference, Gemma 3n empowers a new wave of intelligent, on-the-go applications. It analyzes and responds to combined images and text, with video and audio coming soon. Build intelligent, interactive features that put user privacy first and work reliably offline. Mobile-first architecture, with a significantly reduced memory footprint. Co-designed by Google's mobile hardware teams and industry leaders. 4B active memory footprint with the ability to create submodels for quality-latency tradeoffs. Gemma 3n is our first open model built on this groundbreaking, shared architecture, allowing developers to begin experimenting with this technology today in an early preview.
  • 29
    Orpheus TTS

    Orpheus TTS

    Canopy Labs

    Canopy Labs has introduced Orpheus, a family of state-of-the-art speech large language models (LLMs) designed for human-level speech generation. These models are built on the Llama-3 architecture and are trained on over 100,000 hours of English speech data, enabling them to produce natural intonation, emotion, and rhythm that surpasses current state-of-the-art closed source models. Orpheus supports zero-shot voice cloning, allowing users to replicate voices without prior fine-tuning, and offers guided emotion and intonation control through simple tags. The models achieve low latency, with approximately 200ms streaming latency for real-time applications, reducible to around 100ms with input streaming. Canopy Labs has released both pre-trained and fine-tuned 3B-parameter models under the permissive Apache 2.0 license, with plans to release smaller models of 1B, 400M, and 150M parameters for use on resource-constrained devices.
  • 30
    MiniMax-M1

    MiniMax-M1

    MiniMax

    MiniMax‑M1 is a large‑scale hybrid‑attention reasoning model released by MiniMax AI under the Apache 2.0 license. It supports an unprecedented 1 million‑token context window and up to 80,000-token outputs, enabling extended reasoning across long documents. Trained using large‑scale reinforcement learning with a novel CISPO algorithm, MiniMax‑M1 completed full training on 512 H800 GPUs in about three weeks. It achieves state‑of‑the‑art performance on benchmarks in mathematics, coding, software engineering, tool usage, and long‑context understanding, matching or outperforming leading models. Two model variants are available (40K and 80K thinking budgets), with weights and deployment scripts provided via GitHub and Hugging Face.
  • Previous
  • You're on page 1
  • 2
  • Next