BitNet: Scaling 1-bit Transformers for Large Language Models
An agentic Machine Learning Engineer
A library for accelerating Transformer models on NVIDIA GPUs
LLM Finetuning with peft
Flower: A Friendly Federated Learning Framework
Collaborative & Open-Source Quality Assurance for all AI models
MiniMax M2.1, a SOTA model for real-world dev & agents.
A coding-free framework built on PyTorch
A very simple framework for state-of-the-art NLP
Super Comprehensive Deep Learning Notes
Hunyuan Translation Model Version 1.5
Composable Loggers for the Julia Logging StdLib
Fast State-of-the-Art Tokenizers optimized for Research and Production
An easy to use Fractal wrapper built for Laravel and Lumen
Deep learning optimization library making distributed training easy
Interview guide for machine learning, mathematics, and deep learning
Unifying 3D Mesh Generation with Language Models
End-to-end speech processing toolkit
A game theoretic approach to explain the output of ml models
Rust native ready-to-use NLP pipelines and transformer-based models
Implementation for MatMul-free LM
Accelerate local LLM inference and finetuning
Implementation of Vision Transformer, a simple way to achieve SOTA
Hackable and optimized Transformers building blocks
Scalable and user friendly neural forecasting algorithms.