Implementation of Vision Transformer, a simple way to achieve SOTA
Multilingual sentence & image embeddings with BERT
The open-source data curation platform for LLMs
A Powerful Native Multimodal Model for Image Generation
Composable Loggers for the Julia Logging StdLib
Fast State-of-the-Art Tokenizers optimized for Research and Production
MiniMax M2.1, a SOTA model for real-world dev & agents.
BitNet: Scaling 1-bit Transformers for Large Language Models
A library for accelerating Transformer models on NVIDIA GPUs
Scalable and user friendly neural forecasting algorithms.
An agentic Machine Learning Engineer
A coding-free framework built on PyTorch
A very simple framework for state-of-the-art NLP
End-to-end HTTP and REST API testing for Go
MII makes low-latency and high-throughput inference possible
LLM Finetuning with peft
Super Comprehensive Deep Learning Notes
Hunyuan Translation Model Version 1.5
Interview guide for machine learning, mathematics, and deep learning
Training Large Language Model to Reason in a Continuous Latent Space
A series of math-specific large language models of our Qwen2 series
A game theoretic approach to explain the output of ml models
Rust native ready-to-use NLP pipelines and transformer-based models
Implementation for MatMul-free LM
Accelerate local LLM inference and finetuning