T5-Small: Lightweight text-to-text transformer for NLP tasks
Llama 3.2–1B: Multilingual, instruction-tuned model for mobile AI
CTC-based forced aligner for audio-text in 158 languages
Multimodal 7B model for image, video, and text understanding tasks
CLIP model fine-tuned for zero-shot fashion product classification
Frontier-scale 675B multimodal base model for custom AI training
VaultGemma: 1B DP-trained Gemma variant for private NLP tasks
Metric monocular depth estimation (vision model)
Custom BLEURT model for evaluating text similarity using PyTorch
Text-to-image model optimized for artistic quality and safe generation
Instruction-tuned 7B language model for chat and complex tasks
Lightweight 24B agentic coding model with vision and long context
High-compute ultra-reasoning model surpassing model surpassing GPT-5
Qwen3-Next: 80B instruct LLM with ultra-long context up to 1M tokens
JetBrains’ 4B parameter code model for completions
Instruction-tuned 1.2B LLM for multilingual text generation by Meta
CLIP ViT-bigG/14: Zero-shot image-text model trained on LAION-2B
Efficient 13B MoE language model with long context and reasoning modes
Russian ASR model fine-tuned on Common Voice and CSS10 datasets
Speculative-decoding accelerator for the 675B Mistral Large 3
Quantized 675B multimodal instruct model optimized for NVFP4
Open, non-commercial SDXL model for quality image generation