Hugging Face TransformersHugging Face
|
MiMo-V2-FlashXiaomi Technology
|
|||||
Related Products
|
||||||
About
Transformers is a library of pretrained natural language processing, computer vision, audio, and multimodal models for inference and training. Use Transformers to train models on your data, build inference applications, and generate text with large language models. Explore the Hugging Face Hub today to find a model and use Transformers to help you get started right away. Simple and optimized inference class for many machine learning tasks like text generation, image segmentation, automatic speech recognition, document question answering, and more. A comprehensive trainer that supports features such as mixed precision, torch.compile, and FlashAttention for training and distributed training for PyTorch models. Fast text generation with large language models and vision language models. Every model is implemented from only three main classes (configuration, model, and preprocessor) and can be quickly used for inference or training.
|
About
MiMo-V2-Flash is an open weight large language model developed by Xiaomi based on a Mixture-of-Experts (MoE) architecture that blends high performance with inference efficiency. It has 309 billion total parameters but activates only 15 billion active parameters per inference, letting it balance reasoning quality and computational efficiency while supporting extremely long context handling, for tasks like long-document understanding, code generation, and multi-step agent workflows. It incorporates a hybrid attention mechanism that interleaves sliding-window and global attention layers to reduce memory usage and maintain long-range comprehension, and it uses a Multi-Token Prediction (MTP) design that accelerates inference by processing batches of tokens in parallel. MiMo-V2-Flash delivers very fast generation speeds (up to ~150 tokens/second) and is optimized for agentic applications requiring sustained reasoning and multi-turn interactions.
|
|||||
Platforms Supported
Windows
Mac
Linux
Cloud
On-Premises
iPhone
iPad
Android
Chromebook
|
Platforms Supported
Windows
Mac
Linux
Cloud
On-Premises
iPhone
iPad
Android
Chromebook
|
|||||
Audience
Machine learning practitioners looking for a tool to train and deploy state-of-the-art models across NLP, vision, and audio tasks
|
Audience
Developers and researchers requiring a solution to build high-performance AI applications involving long-context reasoning, coding, and agentic workflows
|
|||||
Support
Phone Support
24/7 Live Support
Online
|
Support
Phone Support
24/7 Live Support
Online
|
|||||
API
Offers API
|
API
Offers API
|
|||||
Screenshots and Videos |
Screenshots and Videos |
|||||
Pricing
$9 per month
Free Version
Free Trial
|
Pricing
Free
Free Version
Free Trial
|
|||||
Reviews/
|
Reviews/
|
|||||
Training
Documentation
Webinars
Live Online
In Person
|
Training
Documentation
Webinars
Live Online
In Person
|
|||||
Company InformationHugging Face
Founded: 2016
United States
huggingface.co/docs/transformers/en/index
|
Company InformationXiaomi Technology
Founded: 2010
China
mimo.xiaomi.com/blog/mimo-v2-flash
|
|||||
Alternatives |
Alternatives |
|||||
|
|
|
|||||
|
|
|
|||||
|
|
||||||
|
|
|
|||||
Categories |
Categories |
|||||
Integrations
Hugging Face
Claude Code
PyTorch
Xiaomi MiMo
Xiaomi MiMo Studio
|
Integrations
Hugging Face
Claude Code
PyTorch
Xiaomi MiMo
Xiaomi MiMo Studio
|
|||||
|
|
|