MoBA: Mixture of Block Attention for Long-Context LLMs
User toolkit for analyzing and interfacing with Large Language Models
Repo of Qwen2-Audio chat & pretrained large audio language model
95% token savings. 155x faster queries. 16 languages
Capable of understanding text, audio, vision, video
Chat & pretrained large audio language model proposed by Alibaba Cloud
Chat & pretrained large vision language model
Code for Language models can explain neurons in language models paper
Open-source, high-performance Mixture-of-Experts large language model
Chat language model that can use tools and interpret the results
The first Chinese LLaMA2 model in the open source community
An interpretable and efficient predictor using pre-trained models