char-rnn
Multi-layer Recurrent Neural Networks (LSTM, GRU, RNN)
char-rnn is a classic codebase for training multi-layer recurrent neural networks on raw text to build character-level language models that learn to predict the next character in a sequence. It supports common recurrent architectures including vanilla RNNs as well as LSTM and GRU variants, letting users compare behavior and output quality across model types. It is straightforward: you provide a single text file, train the model to minimize next-character prediction loss, then sample from the trained network to generate new text one character at a time in the style of the dataset. ...