Open Source Linux Deep Learning Frameworks - Page 10

Deep Learning Frameworks for Linux

View 20 business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 1
    TensorLayer

    TensorLayer

    Deep learning and reinforcement learning library for scientists

    TensorLayer is a novel TensorFlow-based deep learning and reinforcement learning library designed for researchers and engineers. It provides an extensive collection of customizable neural layers to build advanced AI models quickly, based on this, the community open-sourced mass tutorials and applications. TensorLayer is awarded the 2017 Best Open Source Software by the ACM Multimedia Society. This project can also be found at OpenI and Gitee. 3.0.0 has been pre-released, the current version supports TensorFlow, MindSpore and PaddlePaddle (partial) as the backends, allowing users to run the code on different hardware like Nvidia-GPU and Huawei-Ascend. In the future, it will support TensorFlow, MindSpore, PaddlePaddle, PyTorch and other backends. TensorLayer has a high-level layer/model abstraction which is effortless to learn. You can learn how deep learning can benefit your AI tasks in minutes through the massive examples.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    Tez

    Tez

    Tez is a super-simple and lightweight Trainer for PyTorch

    Tez is a super-simple and lightweight Trainer for PyTorch. It also comes with many utils that you can use to tackle over 90% of deep learning projects in PyTorch. tez (तेज़ / تیز) means sharp, fast & active. This is a simple, to-the-point, library to make your PyTorch training easy. This library is in early-stage currently! So, there might be breaking changes. Currently, tez supports cpu, single gpu and multi-gpu & tpu training. More coming soon! Using tez is super-easy. We don't want you to be far away from pytorch. So, you do everything on your own and just use tez to make a few things simpler.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    The Deep Review

    The Deep Review

    A collaboratively written review paper on deep learning, genomics, etc

    This repository is home to the Deep Review, a review article on deep learning in precision medicine. The Deep Review is collaboratively written on GitHub using a tool called Manubot (see below). The project operates on an open contribution model, welcoming contributions from anyone. To see what's incoming, check the open pull requests. For project discussion and planning see the Issues. As of writing, we are aiming to publish an update of the deep review. We will continue to make project preprints available on bioRxiv or another preprint service and aim to continue publishing the finished reviews in a peer-reviewed venue as well. Like the initial release, we are planning for an open and collaborative effort. New contributors are welcome and will be listed as version 2.0 authors. Manubot is a system for writing scholarly manuscripts via GitHub. Manubot automates citations and references, versions manuscripts using git, and enables collaborative writing via GitHub.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    The fastai book

    The fastai book

    The fastai book, published as Jupyter Notebooks

    These notebooks cover an introduction to deep learning, fastai, and PyTorch. fastai is a layered API for deep learning; for more information, see the fastai paper. These notebooks are used for a MOOC and form the basis of this book, which is currently available for purchase. It does not have the same GPL restrictions that are on this repository. The code in the notebooks and python .py files is covered by the GPL v3 license; see the LICENSE file for details. The remainder (including all markdown cells in the notebooks and other prose) is not licensed for any redistribution or change of format or medium, other than making copies of the notebooks or forking this repo for your own private use. No commercial or broadcast use is allowed. We are making these materials freely available to help you learn deep learning, so please respect our copyright and these restrictions.
    Downloads: 0 This Week
    Last Update:
    See Project
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • 5
    Trax

    Trax

    Deep learning with clear code and speed

    Trax is an end-to-end library for deep learning that focuses on clear code and speed. It is actively used and maintained in the Google Brain team. Run a pre-trained Transformer, create a translator in a few lines of code. Features and resources, API docs, where to talk to us, how to open an issue and more. Walkthrough, how Trax works, how to make new models and train on your own data. Trax includes basic models (like ResNet, LSTM, Transformer) and RL algorithms (like REINFORCE, A2C, PPO). It is also actively used for research and includes new models like the Reformer and new RL algorithms like AWR. Trax has bindings to a large number of deep learning datasets, including Tensor2Tensor and TensorFlow datasets. You can use Trax either as a library from your own python scripts and notebooks or as a binary from the shell, which can be more convenient for training large models. It runs without any changes on CPUs, GPUs and TPUs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    VibeTensor

    VibeTensor

    Our first fully AI generated deep learning system

    VibeTensor is a groundbreaking open-source research system software stack for deep learning that was uniquely generated almost entirely by AI coding agents under guided human supervision, demonstrating a new frontier in AI-assisted software engineering. It implements a PyTorch-style eager tensor library with a modern C++20 core that supports both CPU and CUDA backends, giving it the ability to manage tensors, automatic differentiation (autograd), and complex computation flows similar to mainstream frameworks. What makes VibeTensor remarkable is that every major component, from core libraries and dispatch systems to CUDA runtime support, caching allocators, and language bindings, was created and validated by coding agents using automated builds and tests rather than manual line-by-line human coding. The system includes both a Python frontend via a torch-like API and an experimental Node.js/TypeScript interface.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    X-DeepLearning

    X-DeepLearning

    An industrial deep learning framework for high-dimension sparse data

    X-DeepLearning (XDL for short) is a complete set of deep optimization solutions for high-dimensional sparse data scenarios (such as advertising/recommendation/search, etc.). XDL version 1.2 has been released recently. Performance optimization for large batch/low concurrency scenarios, 50-100% performance improvement in such scenarios. Storage and communication optimization, parameters are automatically allocated globally without manual intervention, and requests are merged to completely eliminate computing/storage/communication hotspots of ps. Complete streaming training features including feature admission, feature elimination, model incremental export, feature counting statistics, etc. Background: XDL1.0 focuses on throughput optimization and adopts the one request per thread processing model, which can significantly improve the limit throughput under ultra-high concurrency.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    XZVoice

    XZVoice

    Free and open source text-to-speech software

    Text-to-speech software developed by Electron + vue + ElementUI + js. The high-fidelity and flexible configuration of speech synthesis products opens up the closed loop of human-computer interaction and enables applications to sound realistically. A variety of timbres are available, and functions such as adjusting speech rate, intonation, and volume are provided. Technically, multi-level rhythmic pauses are taken into account to achieve the purpose of natural synthesizing rhythm, and comprehensively use acoustic parameters and linguistic parameters to establish multiple automatic prediction models based on deep learning. Using massive audio data to train the pronunciation model, the synthetic sound is real, full, cadenced, and expressive, and the MOS score has reached the professional level in the industry.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    Zero to Mastery Deep Learning TensorFlow

    Zero to Mastery Deep Learning TensorFlow

    All course materials for the Zero to Mastery Deep Learning with TF

    This project is a comprehensive, code-first deep learning curriculum built around TensorFlow and Keras, designed to guide learners from foundational concepts to practical model deployment through hands-on experimentation. It is structured as a series of progressively complex Jupyter notebooks that emphasize writing and understanding code before diving into theory, reinforcing learning through repetition and application. The material covers core machine learning workflows including regression, classification, computer vision, natural language processing, and time series forecasting, allowing users to build a well-rounded understanding of modern AI tasks. It also integrates milestone projects that simulate real-world scenarios, helping users translate abstract concepts into deployable solutions.
    Downloads: 0 This Week
    Last Update:
    See Project
  • AI-powered service management for IT and enterprise teams Icon
    AI-powered service management for IT and enterprise teams

    Enterprise-grade ITSM, for every business

    Give your IT, operations, and business teams the ability to deliver exceptional services—without the complexity. Maximize operational efficiency with refreshingly simple, AI-powered Freshservice.
    Try it Free
  • 10
    ai trading

    ai trading

    The aim is to build a prototype of agent able to trade by himself

    The aim is to build a prototype of agent able to trade by himself selecting his own criterias and strategies based on its AI capability. Deep learning AI autonomous agent to build FX trading strategies
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    deep-q-learning

    deep-q-learning

    Minimal Deep Q Learning (DQN & DDQN) implementations in Keras

    The deep-q-learning repository authored by keon provides a Python-based implementation of the Deep Q-Learning algorithm — a cornerstone method in reinforcement learning. It implements the core logic needed to train an agent using Q-learning with neural networks (i.e. approximating Q-values via deep nets), setting up environment interaction loops, experience replay, network updates, and policy behavior. For learners and researchers interested in reinforcement learning, this repo offers a concrete, runnable example bridging theory and practice: you can execute the code, play with hyperparameters, observe convergence behavior, and see how deep Q-learning learns policies over time in standard environments. Because it’s self-contained and Python-based, it's well-suited for experimentation, modifications, or extension — for instance adapting to custom Gym environments, tweaking network architecture, or combining with other RL techniques.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12

    evoNet

    Deep learning for population genetic inference

    This software implements the algorithms described in the following paper: Sheehan, S. and Song, Y.S. Deep learning for population genetic inference. PLoS Computational Biology, in press.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    fastai

    fastai

    Deep learning library

    fastai is a deep learning library which provides practitioners with high-level components that can quickly and easily provide state-of-the-art results in standard deep learning domains, and provides researchers with low-level components that can be mixed and matched to build new approaches. It aims to do both things without substantial compromises in ease of use, flexibility, or performance. This is possible thanks to a carefully layered architecture, which expresses common underlying patterns of many deep learning and data processing techniques in terms of decoupled abstractions. These abstractions can be expressed concisely and clearly by leveraging the dynamism of the underlying Python language and the flexibility of the PyTorch library. fastai is organized around two main design goals: to be approachable and rapidly productive, while also being deeply hackable and configurable. It is built on top of a hierarchy of lower-level APIs which provide composable building blocks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14

    irs-elcho

    Telemetryczny system monitoringu stanu technicznego urządzeń

    Celem projektu jest opracowanie telemetrycznego systemu monitoringu stanu technicznego urządzeń typu chłodnie wentylatorowe, który wykorzystując algorytmy typu deep learning będzie stanowił element strategii predykcyjnego utrzymania ruchu. System wpisuje się w ideę Przemysłu 4.0 poprzez umożliwianie monitoringu i predykcji stanu urządzeń. Projekt zakłada badania przemysłowe i prace rozwojowe nad systemem dokonującym kompleksowej akwizycji i analizy danych dot. stanu działania urządzenia. Rejestracji będą podlegać wielkości fizyczne obrazujące działanie urządzeń przy pomocy zbioru czujników. Będą to zarówno sygnały wolnozmienne takie jak temperatura czy wilgotność jak i szybkozmienne takie jak wibracje. Duża ilość danych wejściowych determinuje wykorzystanie sztucznej inteligencji do analizy i predykcji stanu urządzenia. Upraszcza to etap wdrożenia gdyż nie wymaga kosztownej kalibracji a jedynie etapu uczenia, czyli uzupełnienia danych wejściowych o klasyfikację stanu.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    kcws

    kcws

    Deep Learning Chinese Word Segment

    Deep learning chinese word segment. Install the bazel code construction tool and install tensorflow (currently this project requires tf 1.0.0alpha version or above) Switch to the code directory of this project and run ./configure. Compile background service. Pay attention to the public account of waiting for words and reply to kcws to get the corpus download address. Extract the corpus to a directory. Change to the code directory.After installing tensorflow, switch to the kcws code directory. Currently, the custom dictionary is supported in the decoding stage. Please refer to kcws/cc/test_seg.cc for specific usage. The dictionary is in text format.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    nGraph

    nGraph

    nGraph has moved to OpenVINO

    Frameworks using nGraph Compiler stack to execute workloads have shown up to 45X performance boost when compared to native framework implementations. We've also seen performance boosts running workloads that are not included on the list of Validated workloads, thanks to nGraph's powerful subgraph pattern matching. Additionally, we have integrated nGraph with PlaidML to provide deep learning performance acceleration on Intel, nVidia, & AMD GPUs. nGraph Compiler aims to accelerate developing AI workloads using any deep learning framework and deploying to a variety of hardware targets. We strongly believe in providing freedom, performance, and ease of use to AI developers. Our documentation has extensive information about how to use nGraph Compiler stack to create an nGraph computational graph, integrate custom frameworks, and to interact with supported backends.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    satellite-image-deep-learning

    satellite-image-deep-learning

    Resources for deep learning with satellite & aerial imagery

    This page lists resources for performing deep learning on satellite imagery. To a lesser extent classical Machine learning (e.g. random forests) are also discussed, as are classical image processing techniques. Note there is a huge volume of academic literature published on these topics, and this repository does not seek to index them all but rather list approachable resources with published code that will benefit both the research and developer communities. If you find this work useful please give it a star and consider sponsoring it. You can also follow me on Twitter and LinkedIn where I aim to post frequent updates on my new discoveries, and I have created a dedicated group on LinkedIn. I have also started a blog here and have published a post on the history of this repository called Dissecting the satellite-image-deep-learning repo.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    stanford-tensorflow-tutorials

    stanford-tensorflow-tutorials

    This repository contains code examples for the Stanford's course

    This repository contains code examples for the course CS 20: TensorFlow for Deep Learning Research. It will be updated as the class progresses. Detailed syllabus and lecture notes can be found in the site. For this course, I use python3.6 and TensorFlow 1.4.1.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    tf2_course

    tf2_course

    Notebooks for my "Deep Learning with TensorFlow 2 and Keras" course

    tf2_course provides the notebooks for the “Deep Learning with TensorFlow 2 and Keras” course authored by the same author, Aurélien Géron. It is structured as a teaching toolkit: you’ll find notebooks covering neural networks with Keras, lower-level TensorFlow APIs, data loading & preprocessing, convolutional and recurrent networks, and deployment/distribution of models. The material is intended for learners who already have foundational knowledge of ML and wish to deepen their understanding of deep learning frameworks and practices. The repo supports experimentation: you can run code, tweak hyperparameters, and follow guided exercises that strengthen practical mastery. Rather than being book-based, it is course-based, meaning the flow, examples and structure lean toward interactive teaching and incremental builds. It’s well-suited for those who want a focused, deep-learning path rather than a broad ML textbook.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    tvm

    tvm

    Open deep learning compiler stack for cpu, gpu, etc.

    Apache TVM is an open source machine learning compiler framework for CPUs, GPUs, and machine learning accelerators. It aims to enable machine learning engineers to optimize and run computations efficiently on any hardware backend. The vision of the Apache TVM Project is to host a diverse community of experts and practitioners in machine learning, compilers, and systems architecture to build an accessible, extensible, and automated open-source framework that optimizes current and emerging machine learning models for any hardware platform. Compilation of deep learning models in Keras, MXNet, PyTorch, Tensorflow, CoreML, DarkNet and more. Start using TVM with Python today, build out production stacks using C++, Rust, or Java the next day.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    vJEPA-2

    vJEPA-2

    PyTorch code and models for VJEPA2 self-supervised learning from video

    VJEPA2 is a next-generation self-supervised learning framework for video that extends the “predict in representation space” idea from i-JEPA to the temporal domain. Instead of reconstructing pixels, it predicts the missing high-level embeddings of masked space-time regions using a context encoder and a slowly updated target encoder. This objective encourages the model to learn semantics, motion, and long-range structure without the shortcuts that pixel-level losses can invite. The architecture is designed to scale: spatiotemporal ViT backbones, flexible masking schedules, and efficient sampling let it train on long clips while remaining stable. Trained representations transfer well to downstream tasks such as action recognition, temporal localization, and video retrieval, often with simple linear probes or light fine-tuning. The repository typically includes end-to-end recipes—data pipelines, augmentation policies, training scripts, and evaluation harnesses.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB