Showing 2 open source projects for "lstm"

View related business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • 1
    torch-rnn

    torch-rnn

    Efficient, reusable RNNs and LSTMs for torch

    The torch-rnn project is a lightweight and efficient implementation of recurrent neural networks built on the Torch framework, focusing on flexibility and reusability for sequence modeling tasks. It provides implementations of standard RNNs and long short-term memory networks, enabling users to train models for tasks such as text generation, language modeling, and sequence prediction. The repository emphasizes simplicity and performance, offering a streamlined pipeline for preprocessing...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    char-rnn

    char-rnn

    Multi-layer Recurrent Neural Networks (LSTM, GRU, RNN)

    char-rnn is a classic codebase for training multi-layer recurrent neural networks on raw text to build character-level language models that learn to predict the next character in a sequence. It supports common recurrent architectures including vanilla RNNs as well as LSTM and GRU variants, letting users compare behavior and output quality across model types. It is straightforward: you provide a single text file, train the model to minimize next-character prediction loss, then sample from the trained network to generate new text one character at a time in the style of the dataset. The project is designed for experimentation, offering tunable settings for depth, hidden size, dropout, sequence length, and sampling temperature to control creativity and coherence. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB