Open Source Linux Machine Learning Software - Page 22

Machine Learning Software for Linux

View 58 business solutions
  • Full-stack observability with actually useful AI | Grafana Cloud Icon
    Full-stack observability with actually useful AI | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • Fully Managed MySQL, PostgreSQL, and SQL Server Icon
    Fully Managed MySQL, PostgreSQL, and SQL Server

    Automatic backups, patching, replication, and failover. Focus on your app, not your database.

    Cloud SQL handles your database ops end to end, so you can focus on your app.
    Try Free
  • 1
    Deepchecks

    Deepchecks

    Test Suites for validating ML models & data

    Deepchecks is the leading tool for testing and for validating your machine learning models and data, and it enables doing so with minimal effort. Deepchecks accompany you through various validation and testing needs such as verifying your data’s integrity, inspecting its distributions, validating data splits, evaluating your model and comparing between different models. While you’re in the research phase, and want to validate your data, find potential methodological problems, and/or validate your model and evaluate it. To run a specific single check, all you need to do is import it and then to run it with the required (check-dependent) input parameters. More details about the existing checks and the parameters they can receive can be found in our API Reference. An ordered collection of checks, that can have conditions added to them. The Suite enables displaying a concluding report for all of the Checks that ran.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    Deepo

    Deepo

    Set up deep learning environment in a single command line

    Deepo is a series of Docker images that allows you to quickly set up your deep learning research environment, supports almost all commonly used deep learning frameworks, supports GPU acceleration (CUDA and cuDNN included), also works in CPU-only mode, and works on Linux (CPU version/GPU version), Windows (CPU version) and OS X (CPU version). Their Dockerfile generator that allows you to customize your own environment with Lego-like modules, and automatically resolves the dependencies for you. For users in China who may suffer from slow speeds when pulling the image from the public Docker registry, you can pull deepo images from the China registry mirror by specifying the full path, including the registry, in your docker pull command. This should work and enables Deepo to use the GPU from inside a docker container.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Deepvoice3_pytorch

    Deepvoice3_pytorch

    PyTorch implementation of convolutional neural networks

    An open source implementation of Deep Voice 3: Scaling Text-to-Speech with Convolutional Sequence Learning.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4

    Delayed Response Network

    Neural network based on signal delays.

    An artificial neural network, currently specialized to save a specific bit pattern, mainly by changing the signal propagation delays in links. More features, variables and algorithms will be added in time.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 5
    Delta ML

    Delta ML

    Deep learning based natural language and speech processing platform

    DELTA is a deep learning-based end-to-end natural language and speech processing platform. DELTA aims to provide easy and fast experiences for using, deploying, and developing natural language processing and speech models for both academia and industry use cases. DELTA is mainly implemented using TensorFlow and Python 3. DELTA has been used for developing several state-of-the-art algorithms for publications and delivering real production to serve millions of users. It helps you to train, develop, and deploy NLP and/or speech models. Use configuration files to easily tune parameters and network structures. What you see in training is what you get in serving: all data processing and features extraction are integrated into a model graph. Text classification, named entity recognition, question and answering, text summarization, etc. Uniform I/O interfaces and no changes for new models.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    Denoising Diffusion Probabilistic Model

    Denoising Diffusion Probabilistic Model

    Implementation of Denoising Diffusion Probabilistic Model in Pytorch

    Implementation of Denoising Diffusion Probabilistic Model in Pytorch. It is a new approach to generative modeling that may have the potential to rival GANs. It uses denoising score matching to estimate the gradient of the data distribution, followed by Langevin sampling to sample from the true distribution. If you simply want to pass in a folder name and the desired image dimensions, you can use the Trainer class to easily train a model.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Density-ratio based clustering

    Density-ratio based clustering

    Discovering clusters with varying densities

    This site provides the source code of two approaches for density-ratio based clustering, used for discovering clusters with varying densities. One approach is to modify a density-based clustering algorithm to do density-ratio based clustering by using its density estimator to compute density-ratio. The other approach involves rescaling the given dataset only. An existing density-based clustering algorithm, which is applied to the rescaled dataset, can find all clusters with varying densities that would otherwise impossible had the same algorithm been applied to the unscaled dataset. Reference: Zhu, Y., Ting, K. M., & Carman, M. J. (2016). Density-ratio based clustering for discovering clusters with varying densities. Pattern Recognition. http://www.sciencedirect.com/science/article/pii/S0031320316301571
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    Determined

    Determined

    Determined, deep learning training platform

    The fastest and easiest way to build deep learning models. Distributed training without changing your model code. Determined takes care of provisioning machines, networking, data loading, and fault tolerance. Build more accurate models faster with scalable hyperparameter search, seamlessly orchestrated by Determined. Use state-of-the-art algorithms and explore results with our hyperparameter search visualizations. Interpret your experiment results using the Determined UI and TensorBoard, and reproduce experiments with artifact tracking. Deploy your model using Determined's built-in model registry. Easily share on-premise or cloud GPUs with your team. Determined’s cluster scheduling offers first-class support for deep learning and seamless spot instance support. Check out examples of how you can use Determined to train popular deep learning models at scale.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    DialoGPT

    DialoGPT

    Large-scale pretraining for dialogue

    DialoGPT is an open-source conversational language model developed by Microsoft Research for generating natural dialogue responses using large-scale transformer architectures. The system is built on the GPT-2 architecture and is designed specifically for multi-turn conversation tasks, enabling machines to produce coherent responses during interactive dialogue. The model was trained on a massive dataset of approximately 147 million conversational exchanges extracted from Reddit discussion threads, allowing it to learn patterns of natural human conversation. DialoGPT provides multiple pretrained model sizes and includes code for training, fine-tuning, and evaluating dialogue generation models. The repository also contains scripts for preparing conversation datasets and reproducing experimental benchmarks related to conversational AI research.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 10
    Diff Zoo

    Diff Zoo

    Differentiation for Hackers

    Diff-zoo is a learning-focused handbook designed to demystify algorithmic differentiation (AD), the core technique powering modern machine learning frameworks. The project introduces AD from a foundational calculus perspective and gradually builds towards toy implementations that resemble systems like PyTorch and TensorFlow. It clarifies the differences and connections between forward mode, reverse mode, symbolic, numeric, tracing, and source transformation approaches to differentiation. Unlike production-grade AD systems that are often obscured by complex implementation details, these examples are deliberately simple and coherent to highlight the fundamental ideas. The repository is organized as a set of Julia notebooks, allowing learners to explore concepts interactively and compare different methods side by side. By stripping away unnecessary complexity, diff-zoo serves as both an educational resource and a practical guide for anyone.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    DiffEqFlux.jl

    DiffEqFlux.jl

    Pre-built implicit layer architectures with O(1) backprop, GPUs

    DiffEqFlux.jl is a Julia library that combines differential equations with neural networks, enabling the creation of neural differential equations (neural ODEs), universal differential equations, and physics-informed learning models. It serves as a bridge between the DifferentialEquations.jl and Flux.jl libraries, allowing for end-to-end differentiable simulations and model training in scientific machine learning. DiffEqFlux.jl is widely used for modeling dynamical systems with learnable components.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Diffrax

    Diffrax

    Numerical differential equation solvers in JAX

    Diffrax is a numerical differential equation solving library built for the JAX ecosystem, with a strong focus on composability, differentiability, and high-performance scientific computing. The project provides tools for solving ordinary differential equations, stochastic differential equations, controlled differential equations, and related systems in a way that fits naturally into modern machine learning and differentiable programming workflows. Because it is written to work closely with JAX, it supports just-in-time compilation, automatic differentiation, vectorization, and accelerator-backed execution on hardware such as GPUs and TPUs. This makes it especially appealing for researchers who need equation solvers that can be embedded inside trainable models or simulation-heavy learning systems.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Diffusion for World Modeling

    Diffusion for World Modeling

    Learning agent trained in a diffusion world model

    Diffusion for World Modeling is an experimental reinforcement learning system that trains intelligent agents inside a simulated environment generated by a diffusion-based world model. The project introduces the idea of using diffusion models, commonly used for image generation, to simulate the dynamics of an environment and predict future states based on previous observations and actions. Instead of interacting directly with a real environment, the reinforcement learning agent learns within a generative model that produces frames representing the environment. This approach allows training to occur in a simulated world that captures detailed visual dynamics while reducing the need for costly interactions with real environments. The system has been applied to tasks such as Atari game simulations and demonstrations involving complex environments like first-person shooter games.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Distance Scaling

    Distance Scaling

    A Distance Scaling Method to Improve Density-Based Clustering

    These functions implement a distance scaling method, proposed by Ye Zhu, Kai Ming Ting, and Maia Angelova, "A Distance Scaling Method to Improve Density-Based Clustering", in PAKDD2018 proceedings: https://doi.org/10.1007/978-3-319-93040-4_31.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Dive-into-DL-TensorFlow2.0

    Dive-into-DL-TensorFlow2.0

    Dive into Deep Learning

    This project changes the MXNet code implementation in the original book "Learning Deep Learning by Hand" to TensorFlow2 implementation. After consulting Mr. Li Mu by the tutor of archersama , the implementation of this project has been agreed by Mr. Li Mu. Original authors: Aston Zhang, Li Mu, Zachary C. Lipton, Alexander J. Smola and other community contributors. There are some differences between the Chinese and English versions of this book . This project mainly focuses on TensorFlow2 reconstruction for the Chinese version of this book. In addition, this project also refers to the project Dive-into-DL-PyTorch , which refactored PyTorch in the Chinese version of this book, and I would like to express my gratitude here. This repository mainly contains two folders, code and docs (plus some data stored in data). The code folder is the relevant jupyter notebook code for each chapter (based on TensorFlow2); the docs folder is the relevant content in the book.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Django friendly finite state machine

    Django friendly finite state machine

    Django friendly finite state machine support

    Django-fsm adds simple declarative state management for Django models. If you need parallel task execution, view, and background task code reuse over different flows - check my new project Django-view flow. Instead of adding a state field to a Django model and managing its values by hand, you use FSMField and mark model methods with the transition decorator. These methods could contain side effects of the state change. You may also take a look at the Django-fsm-admin project containing a mixin and template tags to integrate Django-fsm state transitions into the Django admin. FSM really helps to structure the code, especially when a new developer comes to the project. FSM is most effective when you use it for some sequential steps. Transition logging support could be achieved with help of django-fsm-log package.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    DoWhy

    DoWhy

    DoWhy is a Python library for causal inference

    DoWhy is a Python library for causal inference that supports explicit modeling and testing of causal assumptions. DoWhy is based on a unified language for causal inference, combining causal graphical models and potential outcomes frameworks. Much like machine learning libraries have done for prediction, DoWhy is a Python library that aims to spark causal thinking and analysis. DoWhy provides a wide variety of algorithms for effect estimation, causal structure learning, diagnosis of causal structures, root cause analysis, interventions and counterfactuals. DoWhy builds on two of the most powerful frameworks for causal inference: graphical causal models and potential outcomes. For effect estimation, it uses graph-based criteria and do-calculus for modeling assumptions and identifying a non-parametric causal effect. For estimation, it switches to methods based primarily on potential outcomes.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    DocCO

    DocCO

    Non-disjoint groupping of Documents based on word sequence approach

    This is a GUI for learning non disjoint groups of documents based on Weka machine learning framework. It offers the possibility to make non disjoint clustering of documents using both vectorial and sequential representation (word sequence approach based on WSK kernel). All data format supported by WEKA could be used in DocCO. Data could be loaded from files, from databases or from specified URL. All the preprocessing techniques implemented in WEKA could be used before performing the learning.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    Docker Machine

    Docker Machine

    Machine management for a container-centric world

    Docker Machine is a tool that lets you install Docker Engine on virtual hosts, and manage the hosts with docker-machine commands. You can use Machine to create Docker hosts on your local Mac or Windows box, on your company network, in your data center, or on cloud providers like Azure, AWS, or DigitalOcean. Using docker-machine commands, you can start, inspect, stop, and restart a managed host, upgrade the Docker client and daemon, and configure a Docker client to talk to your host. Point the Machine CLI at a running, managed host, and you can run docker commands directly on that host. For example, run docker-machine env default to point to a host called default, follow on-screen instructions to complete env setup, and run docker ps, docker run hello-world, and so forth. Machine was the only way to run Docker on Mac or Windows previous to Docker v1.12.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20

    Drug Extraction

    Drug name extraction

    Drug name recognition and normalisation/grounding to DrugBank ids and standard names. Package provides 2 taggers: 1. DrugTagger - CRF-based with DrugBank presence feature (see feature set for details). 2. DrugnameGazetteer - gazetteer/dictionary-based. Dictionary created from DrugBank.ca database. Both taggers include grounding/normalisation to DrugBank ids and standard names. Feature set: Word, Word-1, Word+1, Word-1_Word, Word_Word+1, DrugBankPresence, POS DrugBankPresence feature indicates the presence of the drug name in the DrugBank. Using CONLL-Evaluation: processed 32065 tokens with 3656 phrases; found: 3251 phrases; correct: 2786. accuracy: 95.25%; precision: 85.70%; recall: 76.20%; FB1: 80.67 Using GATE Corpus Benchmark: Strict: P: 0.65 R: 0.73 F1: 0.69 Lenient: P: 0.74 R: 0.84 F1: 0.78 The details of how to reproduce evaluation, see README. To use standalone version for tagging download DrugExtractionStandalone.tar.gz from Files.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    Dynamic Routing Between Capsules

    Dynamic Routing Between Capsules

    A PyTorch implementation of the NIPS 2017 paper

    Dynamic Routing Between Capsules is a PyTorch implementation of the Capsule Network architecture originally proposed to address limitations in traditional convolutional neural networks. Capsule networks aim to improve how neural models represent spatial hierarchies and relationships between objects within images. Instead of scalar neuron activations, capsules output vectors that encode both the presence of features and their spatial properties such as orientation or pose. The repository implements the dynamic routing algorithm between capsules, which allows lower-level features to route their outputs to higher-level structures that best represent the detected patterns. This approach enables the model to capture part-to-whole relationships in visual data more effectively than standard CNNs. The project serves primarily as a research implementation that demonstrates how capsule networks can be built and trained using modern deep learning frameworks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    ECOC PAK is a C++ Library for the Error Correcting Output Codes classification framework. It supports several coding and decoding strategies as well as several classifiers.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    This project is devoted to the creation of an open source Error-Correcting Output Codes (ECOC) library for the Machine Learning community. The ECOC framework is a powerful tool to deal with multi-class categorization problems.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    EEG Seizure Prediction

    EEG Seizure Prediction

    Seizure prediction from EEG data using machine learning

    The Kaggle-EEG project is a machine learning solution developed for seizure prediction from EEG data, achieving 3rd place in the Kaggle/University of Melbourne Seizure Prediction competition. The repository processes EEG data to predict seizures by training machine learning models, specifically using SVM (Support Vector Machine) and RUS Boosted Tree ensemble models. The framework processes EEG data into features, trains models, and outputs predictions, handling temporal data to ensure accuracy.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    ELI5

    ELI5

    A library for debugging/inspecting machine learning classifiers

    ELI5 is a Python library designed to help developers interpret, debug, and explain the predictions of machine learning models. The project focuses on improving model transparency by providing tools that visualize feature importance and prediction reasoning. It supports several popular machine learning frameworks including scikit-learn, XGBoost, LightGBM, CatBoost, and Keras. The library allows users to inspect model weights, analyze decision trees, and compute permutation feature importance for black-box models. It also provides specialized tools such as TextExplainer, which can highlight important words in text classification tasks to explain why a model produced a particular prediction. Additionally, the library integrates explanation algorithms such as LIME to interpret predictions from arbitrary machine learning models.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB