Code for running inference and finetuning with SAM 3 model
Phi-3.5 for Mac: Locally-run Vision and Language Models
A Powerful Native Multimodal Model for Image Generation
INT4/INT5/INT8 and FP16 inference on CPU for RWKV language model
HY-Motion model for 3D character animation generation
Collection of Gemma 3 variants that are trained for performance
Provides convenient access to the Anthropic REST API from any Python 3
GLM-130B: An Open Bilingual Pre-Trained Model (ICLR 2023)
PyTorch implementation of VALL-E (Zero-Shot Text-To-Speech)
A method to increase the speed and lower the memory footprint
An implementation of model parallel GPT-2 and GPT-3-style models
Large language model developed and released by NVIDIA
High-performance MoE model with MLA, MTP, and multilingual reasoning
Small 3B-base multimodal model ideal for custom AI on edge hardware
Versatile 8B-base multimodal LLM, flexible foundation for custom AI
Powerful 14B-base multimodal model — flexible base for fine-tuning
Compact 3B-param multimodal model for efficient on-device reasoning
Efficient 8B multimodal model tuned for advanced reasoning tasks.
High-precision 14B multimodal model built for advanced reasoning tasks
Ultra-efficient 3B multimodal instruct model built for edge deployment
Compact 8B multimodal instruct model optimized for edge deployment
Efficient 14B multimodal instruct model with edge deployment and FP8
Llama 3.2–1B: Multilingual, instruction-tuned model for mobile AI
Frontier-scale 675B multimodal base model for custom AI training
Frontier-scale 675B multimodal instruct MoE model for enterprise AIMis