DeepScaleRAgentica Project
|
DeepSeek-V3.2DeepSeek
|
|||||
Related Products
|
||||||
About
DeepScaleR is a 1.5-billion-parameter language model fine-tuned from DeepSeek-R1-Distilled-Qwen-1.5B using distributed reinforcement learning and a novel iterative context-lengthening strategy that gradually increases its context window from 8K to 24K tokens during training. It was trained on ~40,000 carefully curated mathematical problems drawn from competition-level datasets like AIME (1984–2023), AMC (pre-2023), Omni-MATH, and STILL. DeepScaleR achieves 43.1% accuracy on AIME 2024, a roughly 14.3 percentage point boost over the base model, and surpasses the performance of the proprietary O1-Preview model despite its much smaller size. It also posts strong results on a suite of math benchmarks (e.g., MATH-500, AMC 2023, Minerva Math, OlympiadBench), demonstrating that small, efficient models tuned with RL can match or exceed larger baselines on reasoning tasks.
|
About
DeepSeek-V3.2 is a next-generation open large language model designed for efficient reasoning, complex problem solving, and advanced agentic behavior. It introduces DeepSeek Sparse Attention (DSA), a long-context attention mechanism that dramatically reduces computation while preserving performance. The model is trained with a scalable reinforcement learning framework, allowing it to achieve results competitive with GPT-5 and even surpass it in its Speciale variant. DeepSeek-V3.2 also includes a large-scale agent task synthesis pipeline that generates structured reasoning and tool-use demonstrations for post-training. The model features an updated chat template with new tool-calling logic and the optional developer role for agent workflows. With gold-medal performance in the IMO and IOI 2025 competitions, DeepSeek-V3.2 demonstrates elite reasoning capabilities for both research and applied AI scenarios.
|
|||||
Platforms Supported
Windows
Mac
Linux
Cloud
On-Premises
iPhone
iPad
Android
Chromebook
|
Platforms Supported
Windows
Mac
Linux
Cloud
On-Premises
iPhone
iPad
Android
Chromebook
|
|||||
Audience
Researchers, students, and developers interested in an AI model capable of mathematical reasoning and logic tasks without requiring heavy hardware
|
Audience
DeepSeek-V3.2 is ideal for researchers, developers, and enterprises seeking an open, high-performance LLM optimized for deep reasoning, agent workflows, and long-context tasks
|
|||||
Support
Phone Support
24/7 Live Support
Online
|
Support
Phone Support
24/7 Live Support
Online
|
|||||
API
Offers API
|
API
Offers API
|
|||||
Screenshots and Videos |
Screenshots and Videos |
|||||
Pricing
Free
Free Version
Free Trial
|
Pricing
Free
Free Version
Free Trial
|
|||||
Reviews/
|
Reviews/
|
|||||
Training
Documentation
Webinars
Live Online
In Person
|
Training
Documentation
Webinars
Live Online
In Person
|
|||||
Company InformationAgentica Project
Founded: 2025
United States
agentica-project.com
|
Company InformationDeepSeek
Founded: 2023
China
deepseek.com
|
|||||
Alternatives |
Alternatives |
|||||
|
|
|
|||||
|
|
|
|||||
|
|
|
|||||
|
|
|
|||||
Categories |
Categories |
|||||
Integrations
DeepSeek
Hugging Face
|
||||||
|
|
|