Open Source Linux Machine Learning Software - Page 37

Machine Learning Software for Linux

View 58 business solutions
  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 1
    PySyft

    PySyft

    Data science on data without acquiring a copy

    Most software libraries let you compute over the information you own and see inside of machines you control. However, this means that you cannot compute on information without first obtaining (at least partial) ownership of that information. It also means that you cannot compute using machines without first obtaining control over those machines. This is very limiting to human collaboration and systematically drives the centralization of data, because you cannot work with a bunch of data without first putting it all in one (central) place. The Syft ecosystem seeks to change this system, allowing you to write software which can compute over information you do not own on machines you do not have (total) control over. This not only includes servers in the cloud, but also personal desktops, laptops, mobile phones, websites, and edge devices. Wherever your data wants to live in your ownership, the Syft ecosystem exists to help keep it there while allowing it to be used privately.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    PyTensor

    PyTensor

    Python library for defining and optimizing mathematical expressions

    PyTensor is a fork of Aesara, a Python library for defining, optimizing, and efficiently evaluating mathematical expressions involving multi-dimensional arrays. PyTensor is based on Theano, which has been powering large-scale computationally intensive scientific investigations since 2007. A hackable, pure-Python codebase. Extensible graph framework is suitable for rapid development of custom operators and symbolic optimizations. Implements an extensible graph transpilation framework that currently provides compilation via C, JAX, and Numba. Based on one of the most widely-used Python tensor libraries: Theano.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    PyTextRank

    PyTextRank

    Python implementation of TextRank algorithms

    PyTextRank is a Python implementation of TextRank as a spaCy pipeline extension, for graph-based natural language work -- and related knowledge graph practices.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    PyTorch Book

    PyTorch Book

    PyTorch tutorials and fun projects including neural talk

    This is the corresponding code for the book "The Deep Learning Framework PyTorch: Getting Started and Practical", but it can also be used as a standalone PyTorch Getting Started Guide and Tutorial. The current version of the code is based on pytorch 1.0.1, if you want to use an older version please git checkout v0.4or git checkout v0.3. Legacy code has better python2/python3 compatibility, CPU/GPU compatibility test. The new version of the code has not been fully tested, it has been tested under GPU and python3. But in theory there shouldn't be too many problems on python2 and CPU. The basic part (the first five chapters) explains the content of PyTorch. This part introduces the main modules in PyTorch and some tools commonly used in deep learning. For this part of the content, Jupyter Notebook is used as a teaching tool here, and readers can modify and run with notebooks and repeat experiments.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 5
    PyTorch Forecasting

    PyTorch Forecasting

    Time series forecasting with PyTorch

    PyTorch Forecasting aims to ease state-of-the-art time series forecasting with neural networks for both real-world cases and research alike. The goal is to provide a high-level API with maximum flexibility for professionals and reasonable defaults for beginners. A time series dataset class that abstracts handling variable transformations, missing values, randomized subsampling, multiple history lengths, etc. A base model class that provides basic training of time series models along with logging in tensorboard and generic visualizations such actual vs predictions and dependency plots. Multiple neural network architectures for timeseries forecasting that have been enhanced for real-world deployment and come with in-built interpretation capabilities. The package is built on PyTorch Lightning to allow training on CPUs, single and multiple GPUs out-of-the-box.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    PyTorch Geometric Temporal

    PyTorch Geometric Temporal

    Spatiotemporal Signal Processing with Neural Machine Learning Models

    The library consists of various dynamic and temporal geometric deep learning, embedding, and Spatio-temporal regression methods from a variety of published research papers. Moreover, it comes with an easy-to-use dataset loader, train-test splitter and temporal snaphot iterator for dynamic and temporal graphs. The framework naturally provides GPU support. It also comes with a number of benchmark datasets from the epidemiological forecasting, sharing economy, energy production and web traffic management domains. Finally, you can also create your own datasets. The package interfaces well with Pytorch Lightning which allows training on CPUs, single and multiple GPUs out-of-the-box. PyTorch Geometric Temporal makes implementing Dynamic and Temporal Graph Neural Networks quite easy - see the accompanying tutorial. Head over to our documentation to find out more about installation, creation of datasets and a full list of implemented methods and available datasets.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    PyTorch Handbook

    PyTorch Handbook

    The pytorch handbook is an open source book

    PyTorch Handbook is an open-source educational project designed to help developers and researchers quickly learn deep learning using the PyTorch framework. The repository functions as an online handbook that explains how to build, train, and evaluate neural network models using PyTorch. It includes tutorials and examples that demonstrate common deep learning tasks such as image classification, neural network design, model training workflows, and evaluation techniques. The material is written with a practical focus so that readers can follow along and run the provided examples successfully. Each tutorial is tested to ensure that the code runs correctly, making the repository particularly useful for beginners who want reliable learning materials. The handbook emphasizes hands-on learning through real code examples rather than purely theoretical explanations.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    PyTorch Implementation of SDE Solvers

    PyTorch Implementation of SDE Solvers

    Differentiable SDE solvers with GPU support and efficient sensitivity

    This library provides stochastic differential equation (SDE) solvers with GPU support and efficient backpropagation. examples/demo.ipynb gives a short guide on how to solve SDEs, including subtle points such as fixing the randomness in the solver and the choice of noise types. examples/latent_sde.py learns a latent stochastic differential equation, as in Section 5 of [1]. The example fits an SDE to data, whilst regularizing it to be like an Ornstein-Uhlenbeck prior process. The model can be loosely viewed as a variational autoencoder with its prior and approximate posterior being SDEs. The program outputs figures to the path specified by <TRAIN_DIR>. Training should stabilize after 500 iterations with the default hyperparameters. examples/sde_gan.py learns an SDE as a GAN, as in [2], [3]. The example trains an SDE as the generator of a GAN, whilst using a neural CDE [4] as the discriminator.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    PyTorch Natural Language Processing

    PyTorch Natural Language Processing

    Basic Utilities for PyTorch Natural Language Processing (NLP)

    PyTorch-NLP is a library for Natural Language Processing (NLP) in Python. It’s built with the very latest research in mind, and was designed from day one to support rapid prototyping. PyTorch-NLP comes with pre-trained embeddings, samplers, dataset loaders, metrics, neural network modules and text encoders. It’s open-source software, released under the BSD3 license. With your batch in hand, you can use PyTorch to develop and train your model using gradient descent. For example, check out this example code for training on the Stanford Natural Language Inference (SNLI) Corpus. Now you've setup your pipeline, you may want to ensure that some functions run deterministically. Wrap any code that's random, with fork_rng and you'll be good to go. Now that you've computed your vocabulary, you may want to make use of pre-trained word vectors to set your embeddings.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Full-stack observability with actually useful AI | Grafana Cloud Icon
    Full-stack observability with actually useful AI | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 10
    PyTorch SimCLR

    PyTorch SimCLR

    PyTorch implementation of SimCLR: A Simple Framework

    For quite some time now, we know about the benefits of transfer learning in Computer Vision (CV) applications. Nowadays, pre-trained Deep Convolution Neural Networks (DCNNs) are the first go-to pre-solutions to learn a new task. These large models are trained on huge supervised corpora, like the ImageNet. And most important, their features are known to adapt well to new problems. This is particularly interesting when annotated training data is scarce. In situations like this, we take the models’ pre-trained weights, append a new classifier layer on top of it, and retrain the network. This is called transfer learning, and is one of the most used techniques in CV. Aside from a few tricks when performing fine-tuning (if the case), it has been shown (many times) that if training for a new task, models initialized with pre-trained weights tend to learn faster and be more accurate then training from scratch using random initialization.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    PyTorch/XLA

    PyTorch/XLA

    Enabling PyTorch on Google TPU

    PyTorch/XLA is a Python package that uses the XLA deep learning compiler to connect the PyTorch deep learning framework and Cloud TPUs. You can try it right now, for free, on a single Cloud TPU with Google Colab, and use it in production and on Cloud TPU Pods with Google Cloud. Take a look at one of our Colab notebooks to quickly try different PyTorch networks running on Cloud TPUs and learn how to use Cloud TPUs as PyTorch devices. We are also introducing new TPU VMs for more transparent and easier access to the TPU hardware. This is our recommedned way of running PyTorch/XLA on Cloud TPU. Please check out our Cloud TPU VM User Guide. Cloud TPU VM is currently on general availability and provides direct access to the TPU host. The recommended setup for running distributed training on TPU Pods uses the pairing of Compute VM Instance Groups and TPU Pods. Each of the Compute VM in the instance group drives 8 cores on the TPU Pod.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12

    PyVocabularyTree

    A vocabulary tree for image classification using OpenCV

    A vocabulary tree for image classification have been designed to be integrated in mobile robotic applications. It is a learning schema based on decission trees, bags of features and inverted files. The design provides training and optimization parameters that have been characterized using several detectors and descriptors for several input datasets. Evaluation tests performed on public image databases allow to compare obtained results with previously published literature. All the tools and resources used in this project are Open Source licensed.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    The Python Computer Vision Framework is an opened project deisgned for all those interested in computer vision. It aims at making computer vision more easy and structured and matlab-free. It may also be used for other artistic and scientific areas.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Python ML Jupyter Notebooks

    Python ML Jupyter Notebooks

    Practice and tutorial-style notebooks

    Python ML Jupyter Notebooks is an educational repository that demonstrates how to implement machine learning algorithms and data science workflows using Python. The project provides numerous examples and tutorials covering classical machine learning techniques such as regression, classification, clustering, and dimensionality reduction. It includes code implementations that show how to build models using popular libraries like scikit-learn, NumPy, pandas, and Matplotlib. The repository is designed to help learners understand both the theory and practical implementation of machine learning algorithms through step-by-step code examples. Many notebooks include explanations of algorithm behavior, data preparation techniques, and evaluation methods for machine learning models. The project also includes examples that demonstrate how to apply machine learning to real-world datasets and practical business problems.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Python Machine Learning

    Python Machine Learning

    The "Python Machine Learning (2nd edition)" book code repository

    This repository accompanies the well-known textbook “Python Machine Learning, 2nd Edition” by Sebastian Raschka and Vahid Mirjalili, serving as a complete codebase of examples, notebooks, scripts and supporting materials for the book. It covers a wide range of topics including supervised learning, unsupervised learning, dimensionality reduction, model evaluation, deep learning with TensorFlow, and embedding models into web apps. Each chapter has Jupyter notebooks and Python scripts that replicate the examples in the book, allowing readers to run, inspect, and tweak code directly as they follow material. The structure also includes errata documentation and assets (images) that appear in the printed edition, providing a rich supplement to learning. The repository is suitable both for classroom use and for self-study, as well as being a go-to reference for data scientists revisiting techniques.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Python Machine Learning book

    Python Machine Learning book

    The book code repository and info resource

    What you can expect are 400 pages rich in useful material just about everything you need to know to get started with machine learning. From theory to the actual code that you can directly put into action! This is not yet just another "this is how scikit-learn works" book. I aim to explain all the underlying concepts, tell you everything you need to know in terms of best practices and caveats, and we will put those concepts into action mainly using NumPy, scikit-learn, and Theano. This is not yet just another "this is how scikit-learn works" book. its aim is to explain how Machine Learning works, tell you everything you need to know in terms of best practices and caveats, and then we will learn how to put those concepts into action using NumPy, scikit-learn, Theano and so on. Many parts of this book will provide examples in scikit-learn, the most beautiful and practical machine learning library.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Pytorch Points 3D

    Pytorch Points 3D

    Pytorch framework for doing deep learning on point clouds

    Torch Points 3D is a framework for developing and testing common deep learning models to solve tasks related to unstructured 3D spatial data i.e. Point Clouds. The framework currently integrates some of the best-published architectures and it integrates the most common public datasets for ease of reproducibility. It heavily relies on Pytorch Geometric and Facebook Hydra library thanks for the great work! We aim to build a tool that can be used for benchmarking SOTA models, while also allowing practitioners to efficiently pursue research into point cloud analysis, with the end goal of building models which can be applied to real-life applications. Task driven implementation with dynamic model and dataset resolution from arguments. Core implementation of common components for point cloud deep learning - greatly simplifying the creation of new models. 4 Base Convolution base classes to simplify the implementation of new convolutions. Each base class supports a different data format.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    Pytorch-toolbelt

    Pytorch-toolbelt

    PyTorch extensions for fast R&D prototyping and Kaggle farming

    A pytorch-toolbelt is a Python library with a set of bells and whistles for PyTorch for fast R&D prototyping and Kaggle farming. Easy model building using flexible encoder-decoder architecture. Modules: CoordConv, SCSE, Hypercolumn, Depthwise separable convolution and more. GPU-friendly test-time augmentation TTA for segmentation and classification. GPU-friendly inference on huge (5000x5000) images. Every-day common routines (fix/restore random seed, filesystem utils, metrics). Losses: BinaryFocalLoss, Focal, ReducedFocal, Lovasz, Jaccard and Dice losses, Wing Loss and more. Extras for Catalyst library (Visualization of batch predictions, additional metrics). By design, both encoder and decoder produces a list of tensors, from fine (high-resolution, indexed 0) to coarse (low-resolution) feature maps. Access to all intermediate feature maps is beneficial if you want to apply deep supervision losses on them or encoder-decoder of object detection task.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    QSMM

    QSMM

    An adaptive state model development framework.

    QSMM, a recursive acronym for "QSMM State Machine Model", is a framework for learning finite automatons that perform goal-directed interaction with entities which exhibit deterministic or stochastic behavior. The learning process can be carried out in real time together with the interaction process. A basic building block for supporting state models of finite automatons is adaptive probabilistic mapping, which for an argument from its domain returns more often results that maximize or minimize values of one or more objective functions. Finite automatons can be represented by assembler programs with user-defined instructions that perform effective work. To assist in the learning of a finite automaton, a template for its state model can be provided as an assembler program with probabilistic jump instructions. The operating principle behind the framework resembles the Boltzmann machine.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    Qix

    Qix

    Machine Learning, Deep Learning, PostgreSQL, Distributed System

    Qix is an open-source repository that aggregates resources related to artificial intelligence, machine learning, and deep learning development. The project is designed as a curated index of useful tools, frameworks, research papers, and learning materials that developers can use to explore the AI ecosystem. It organizes information about libraries, tutorials, datasets, and software tools used in modern AI research and development. The repository helps developers quickly discover relevant technologies and educational resources without searching through multiple sources individually. It also includes references to frameworks, model implementations, and related research projects across various machine learning subfields. The project acts as a directory of resources that can guide learners toward useful tools and documentation in the AI domain.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    Qt Handwriting Recognizing it's a simple Qt GUI interface of a artificial neural network to provide handwrite recognition. This project use FANN (Fast Artificial Neural Network) on first approach.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    R-Portfolio
    R-Portfolio - breakeven optimally diversified investment portfolio. Оптимально диверсифицированный инвестиционный портфель.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    RBM-provisor
    An experimental unsupervised learning method for improvising jazz melodies, based on restricted Boltzmann machines (RBMs) layered into Deep Belief Networks (forms of neural networks).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    RL with PyTorch

    RL with PyTorch

    Clean, Robust, and Unified PyTorch implementation

    RL with PyTorch is a research-oriented repository that provides implementations of deep reinforcement learning algorithms using the PyTorch framework. The project focuses on helping developers and researchers understand reinforcement learning methods by providing clean and reproducible implementations of well-known algorithms. It includes code for popular deep reinforcement learning techniques such as Deep Q-Networks, policy gradient methods, actor-critic architectures, and other modern RL approaches. The repository is structured so that users can easily experiment with different algorithms and training environments. Many examples demonstrate how agents learn to interact with simulated environments through trial and error using reinforcement learning principles. The codebase emphasizes clarity and modular design so that researchers can extend the implementations or use them for experimentation and benchmarking.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Rainbow

    Rainbow

    Rainbow: Combining Improvements in Deep Reinforcement Learning

    Combining improvements in deep reinforcement learning. Results and pretrained models can be found in the releases. Data-efficient Rainbow can be run using several options (note that the "unbounded" memory is implemented here in practice by manually setting the memory capacity to be the same as the maximum number of timesteps).
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB