Lightweight multimodal translation model for 55 languages
Efficient MoE model for million-token reasoning and coding
ClinicalBERT model trained on MIMIC notes for clinical NLP tasks
Efficient 14B multimodal instruct model with edge deployment and FP8
Large-scale xAI model for local inference with SGLang, Grok-2.5
Tiny pre-trained IBM model for multivariate time series forecasting
Quantized 675B multimodal instruct model optimized for NVFP4
Ultra-efficient 3B multimodal instruct model built for edge deployment