Open-LLM-VTuber is an open source platform designed to create AI-powered VTuber characters that can interact with users through voice and animated avatars. It enables hands-free conversations with large language models by combining speech recognition, language processing, and text-to-speech synthesis into a single system. Users can speak directly to the AI character, and the system can respond with a generated voice while animating a Live2D avatar to simulate a talking virtual personality. Open-LLM-VTuber is modular, allowing developers to swap or configure different language models, speech recognition engines, and voice synthesis systems depending on their needs. It can run locally and supports both offline and online AI services, giving users flexibility in how models and resources are used. Open-LLM-VTuber was originally inspired by the goal of recreating an AI VTuber experience using open source tools that work across multiple operating systems.
Features
- Hands-free voice interaction with large language models
- Live2D avatar integration that animates during conversations
- Modular architecture for swapping LLM, ASR, and TTS engines
- Support for both offline local models and online AI services
- Long-term memory options for persistent conversational context
- Cross-platform deployment with GPU or CPU execution options