A multimodal AI story teller, built with Stable Diffusion, GPT, and neural text-to-speech (TTS). Given a prompt as an opening line of a story, GPT writes the rest of the plot; Stable Diffusion draws an image for each sentence; a TTS model narrates each line, resulting in a fully animated video of a short story, replete with audio and visuals. To develop locally, install dev dependencies and install pre-commit hooks. This will automatically trigger linting and code quality checks before each commit. The final video will be saved as /out/out.mp4, alongside other intermediate images, audio files, and subtitles. For more advanced use cases, you can also directly interface with Story Teller in Python code.
Features
- Story Teller is available on PyPI
- The quickest way to run a demo is through the CLI. Simply type
- The final video will be saved as /out/out.mp4, alongside other intermediate images, audio files, and subtitles
- To adjust the defaults with custom parametes, toggle the CLI flags as needed
- For more advanced use cases, you can also directly interface with Story Teller in Python code
- Configure the model with custom settings
License
MIT LicenseFollow StoryTeller
Other Useful Business Software
Go From AI Idea to AI App Fast
Access Gemini 3 and 200+ models. Build chatbots, agents, or custom models with built-in monitoring and scaling.
Rate This Project
Login To Rate This Project
User Reviews
Be the first to post a review of StoryTeller!