Code for running inference and finetuning with SAM 3 model
AlphaFold 3 inference pipeline
Phi-3.5 for Mac: Locally-run Vision and Language Models
A Powerful Native Multimodal Model for Image Generation
INT4/INT5/INT8 and FP16 inference on CPU for RWKV language model
HY-Motion model for 3D character animation generation
Collection of Gemma 3 variants that are trained for performance
Uncommon Objects in 3D dataset
GPT4V-level open-source multi-modal model based on Llama3-8B
LLM-based Reinforcement Learning audio edit model
Provides convenient access to the Anthropic REST API from any Python 3
Runtime extension of Proximus enabling Deployment on AMD Ryzen™ AI
GLM-130B: An Open Bilingual Pre-Trained Model (ICLR 2023)
PyTorch implementation of VALL-E (Zero-Shot Text-To-Speech)
A method to increase the speed and lower the memory footprint
An implementation of model parallel GPT-2 and GPT-3-style models
Large language model developed and released by NVIDIA
High-performance MoE model with MLA, MTP, and multilingual reasoning
Small 3B-base multimodal model ideal for custom AI on edge hardware
Versatile 8B-base multimodal LLM, flexible foundation for custom AI
Powerful 14B-base multimodal model — flexible base for fine-tuning
Compact 3B-param multimodal model for efficient on-device reasoning
Efficient 8B multimodal model tuned for advanced reasoning tasks.
High-precision 14B multimodal model built for advanced reasoning tasks
Ultra-efficient 3B multimodal instruct model built for edge deployment