Open Source Linux Machine Learning Software - Page 8

Machine Learning Software for Linux

View 57 business solutions
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 1
    CVPR 2025

    CVPR 2025

    Collection of CVPR 2025 papers and open source projects

    CVPR 2025 curates accepted CVPR 2025 papers and pairs them with their corresponding code implementations when available, giving researchers and practitioners a fast way to move from reading to reproducing. It organizes entries by topic areas such as detection, segmentation, generative models, 3D vision, multi-modal learning, and efficiency, so you can navigate the year’s output efficiently. Each paper entry typically includes a title, author list, and links to the paper PDF and official or third-party code repositories. The list frequently highlights benchmarks, leaderboards, or notable results so readers can assess impact at a glance. Because conference content evolves rapidly, the repository is updated as authors release code or refine readme instructions, keeping the collection timely. For teams planning literature reviews, study groups, or rapid prototyping sprints, it acts as a central index to the year’s most relevant methods with working implementations.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 2
    CakeChat

    CakeChat

    CakeChat: Emotional Generative Dialog System

    CakeChat is a backend for chatbots that are able to express emotions via conversations. The code is flexible and allows to condition model's responses by an arbitrary categorical variable. For example, you can train your own persona-based neural conversational model or create an emotional chatting machine. Hierarchical Recurrent Encoder-Decoder (HRED) architecture for handling deep dialog context. Multilayer RNN with GRU cells. The first layer of the utterance-level encoder is always bidirectional. By default, CuDNNGRU implementation is used for ~25% acceleration during inference. Thought vector is fed into decoder on each decoding step. Decoder can be conditioned on any categorical label, for example, emotion label or persona id. May be initialized using w2v model trained on your corpus. Embedding layer may be either fixed or fine-tuned along with other weights of the network.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 3
    Catalyst

    Catalyst

    Accelerated deep learning R&D

    Catalyst is a PyTorch framework for accelerated Deep Learning research and development. It allows you to write compact but full-featured Deep Learning pipelines with just a few lines of code. With Catalyst you get a full set of features including a training loop with metrics, model checkpointing and more, all without the boilerplate. Catalyst is focused on reproducibility, rapid experimentation, and codebase reuse so you can break the cycle of writing another regular train loop and make something totally new. Catalyst is compatible with Python 3.6+. PyTorch 1.1+, and has been tested on Ubuntu 16.04/18.04/20.04, macOS 10.15, Windows 10 and Windows Subsystem for Linux. It's part of the PyTorch Ecosystem, as well as the Catalyst Ecosystem which includes Alchemy (experiments logging & visualization) and Reaction (convenient deep learning models serving).
    Downloads: 1 This Week
    Last Update:
    See Project
  • 4
    ClearML

    ClearML

    Streamline your ML workflow

    ClearML is an open source platform that automates and simplifies developing and managing machine learning solutions for thousands of data science teams all over the world. It is designed as an end-to-end MLOps suite allowing you to focus on developing your ML code & automation, while ClearML ensures your work is reproducible and scalable. The ClearML Python Package for integrating ClearML into your existing scripts by adding just two lines of code, and optionally extending your experiments and other workflows with ClearML powerful and versatile set of classes and methods. The ClearML Server storing experiment, model, and workflow data, and supports the Web UI experiment manager, and ML-Ops automation for reproducibility and tuning. It is available as a hosted service and open source for you to deploy your own ClearML Server. The ClearML Agent for ML-Ops orchestration, experiment and workflow reproducibility, and scalability.
    Downloads: 1 This Week
    Last Update:
    See Project
  • Fully Managed MySQL, PostgreSQL, and SQL Server Icon
    Fully Managed MySQL, PostgreSQL, and SQL Server

    Automatic backups, patching, replication, and failover. Focus on your app, not your database.

    Cloud SQL handles your database ops end to end, so you can focus on your app.
    Try Free
  • 5
    CodeContests

    CodeContests

    Large dataset of coding contests designed for AI and ML model training

    CodeContests, developed by Google DeepMind, is a large-scale competitive programming dataset designed for training and evaluating machine learning models on code generation and problem solving. This dataset played a central role in the development of AlphaCode, DeepMind’s model for solving programming problems at a human-competitive level, as published in Science. CodeContests aggregates problems and human-written solutions from multiple programming competition platforms, including AtCoder, Codeforces, CodeChef, Aizu, and HackerEarth. Each problem includes structured metadata, problem descriptions, paired input/output test cases, and multiple correct and incorrect solutions in various programming languages. The dataset is distributed in Riegeli format using Protocol Buffers, with separate training, validation, and test splits for reproducible machine learning experiments.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 6
    CodeSearchNet

    CodeSearchNet

    Datasets, tools, and benchmarks for representation learning of code

    CodeSearchNet is a large-scale dataset and research benchmark designed to advance the development of systems that retrieve source code using natural language queries. The project was created through collaboration between GitHub and Microsoft Research and aims to support research on semantic code search and program understanding. The dataset contains millions of pairs of source code functions and corresponding documentation comments extracted from open-source repositories. These pairs allow machine learning models to learn relationships between natural language descriptions and programming code. The dataset currently covers several widely used programming languages, including Python, JavaScript, Ruby, Go, Java, and PHP. In addition to the dataset itself, the repository includes baseline models, evaluation tools, and instructions for building code retrieval systems that can map user queries to relevant code snippets.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 7
    ComfyUI-3D-Pack

    ComfyUI-3D-Pack

    An extensive node suite that enables ComfyUI to process 3D inputs

    ComfyUI-3D-Pack is an extension package for the ComfyUI visual AI workflow environment that enables users to generate and manipulate 3D assets using advanced machine learning techniques. ComfyUI itself is a node-based interface for designing and executing generative AI pipelines, and this extension expands its capabilities by introducing nodes specifically designed for working with three-dimensional data. The package allows the platform to process inputs such as meshes and UV textures and integrate them into generative workflows similar to those used for image and video generation. It incorporates modern 3D generation technologies including neural radiance fields, Gaussian splatting, and other AI-driven reconstruction techniques. Through these nodes, users can convert images into 3D models, manipulate geometry, and experiment with generative 3D workflows inside the visual pipeline editor.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 8
    Computational Linear Algebra for Coders

    Computational Linear Algebra for Coders

    Free online textbook of Jupyter notebooks

    Computational Linear Algebra for Coders is an open-source educational repository created by the fast.ai community that serves as a free online textbook and course for computational linear algebra. The project presents linear algebra concepts from a practical perspective focused on how computers perform matrix operations efficiently and accurately. The course materials are organized as Jupyter notebooks that combine explanations, code demonstrations, and exercises. Instead of emphasizing purely theoretical mathematics, the project takes a programming-oriented approach that helps developers understand how linear algebra algorithms are implemented in real computational systems. The course explores topics such as matrix decomposition, numerical stability, and optimization techniques that are essential for machine learning and data science applications.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 9
    Core ML Tools

    Core ML Tools

    Core ML tools contain supporting tools for Core ML model conversion

    Use Core ML Tools (coremltools) to convert machine learning models from third-party libraries to the Core ML format. This Python package contains the supporting tools for converting models from training libraries. Core ML is an Apple framework to integrate machine learning models into your app. Core ML provides a unified representation for all models. Your app uses Core ML APIs and user data to make predictions, and to fine-tune models, all on the user’s device. Core ML optimizes on-device performance by leveraging the CPU, GPU, and Neural Engine while minimizing its memory footprint and power consumption. Running a model strictly on the user’s device removes any need for a network connection, which helps keep the user’s data private and your app responsive.
    Downloads: 1 This Week
    Last Update:
    See Project
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • 10
    DCVGAN

    DCVGAN

    DCVGAN: Depth Conditional Video Generation, ICIP 2019.

    This paper proposes a new GAN architecture for video generation with depth videos and color videos. The proposed model explicitly uses the information of depth in a video sequence as additional information for a GAN-based video generation scheme to make the model understands scene dynamics more accurately. The model uses pairs of color video and depth video for training and generates a video using the two steps. Generate the depth video to model the scene dynamics based on the geometrical information. To add appropriate color to the geometrical information of the scene, the domain translation from depth to color is performed for each image. This model has three networks in the generator. In addition, the model has two discriminators.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    Deep Learning course

    Deep Learning course

    Slides and Jupyter notebooks for the Deep Learning lectures

    Slides and Jupyter notebooks for the Deep Learning lectures at Master Year 2 Data Science from Institut Polytechnique de Paris. This course is being taught at as part of Master Year 2 Data Science IP-Paris. Note: press "P" to display the presenter's notes that include some comments and additional references. This lecture is built and maintained by Olivier Grisel and Charles Ollion.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 12
    DeepChem

    DeepChem

    Democratizing Deep-Learning for Drug Discovery, Quantum Chemistry, etc

    DeepChem aims to provide a high-quality open-source toolchain that democratizes the use of deep learning in drug discovery, materials science, quantum chemistry, and biology. DeepChem currently supports Python 3.7 through 3.9 and requires these packages on any condition. DeepChem has a number of "soft" requirements. If you face some errors like ImportError: This class requires XXXX, you may need to install some packages. Deepchem provides support for TensorFlow, PyTorch, JAX and each requires an individual pip Installation. The DeepChem project maintains an extensive collection of tutorials. All tutorials are designed to be run on Google collab (or locally if you prefer). Tutorials are arranged in a suggested learning sequence that will take you from beginner to proficient at molecular machine learning and computational biology more broadly.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 13
    DeepLearning

    DeepLearning

    Deep Learning (Flower Book) mathematical derivation

    " Deep Learning " is the only comprehensive book in the field of deep learning. The full name is also called the Deep Learning AI Bible (Deep Learning) . It is edited by three world-renowned experts, Ian Goodfellow, Yoshua Bengio, and Aaron Courville. Includes linear algebra, probability theory, information theory, numerical optimization, and related content in machine learning. At the same time, it also introduces deep learning techniques used by practitioners in the industry, including deep feedforward networks, regularization, optimization algorithms, convolutional networks, sequence modeling and practical methods, and investigates topics such as natural language processing, Applications in speech recognition, computer vision, online recommender systems, bioinformatics, and video games. Finally, the Deep Learning book provides research directions covering theoretical topics including linear factor models, autoencoders, representation learning, structured probabilistic models, etc.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 14
    DeepLearning Tutorial

    DeepLearning Tutorial

    Deep Learning Tutorial, Excellent Articles, Deep Learning Tutorial

    DeepLearning is an open-source repository that aggregates tutorials, articles, and educational resources related to deep learning and machine learning. The project is designed as a knowledge collection that helps beginners understand neural networks, deep learning architectures, and fundamental machine learning concepts. It contains curated learning materials covering topics such as feedforward neural networks, activation functions, backpropagation algorithms, optimization methods, and convolutional neural networks. The repository organizes these materials into structured tutorials and references that allow readers to explore deep learning concepts progressively. Many of the resources include explanations of common model architectures used in computer vision and artificial intelligence. The project also collects recommended articles and educational documents that expand on deep learning theory and practical implementation strategies.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 15
    Deepnote

    Deepnote

    Deepnote is a drop-in replacement for Jupyter

    Deepnote is an open-source collaborative data science notebook platform designed as a modern alternative to traditional Jupyter notebooks. The project provides an AI-first computational environment where users can write, analyze, and share code, data, and visualizations in a single integrated workspace. Built on top of the Jupyter kernel ecosystem, it maintains compatibility with existing notebook workflows while introducing additional features focused on collaboration and automation. The system supports programming languages such as Python, R, and SQL and allows users to execute and analyze data directly within interactive notebooks. Deepnote emphasizes team-based data science by enabling real-time collaboration similar to shared document editors, allowing multiple users to work simultaneously on the same notebook environment.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 16
    Diff Zoo

    Diff Zoo

    Differentiation for Hackers

    Diff-zoo is a learning-focused handbook designed to demystify algorithmic differentiation (AD), the core technique powering modern machine learning frameworks. The project introduces AD from a foundational calculus perspective and gradually builds towards toy implementations that resemble systems like PyTorch and TensorFlow. It clarifies the differences and connections between forward mode, reverse mode, symbolic, numeric, tracing, and source transformation approaches to differentiation. Unlike production-grade AD systems that are often obscured by complex implementation details, these examples are deliberately simple and coherent to highlight the fundamental ideas. The repository is organized as a set of Julia notebooks, allowing learners to explore concepts interactively and compare different methods side by side. By stripping away unnecessary complexity, diff-zoo serves as both an educational resource and a practical guide for anyone.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 17
    FEniCS.jl

    FEniCS.jl

    A scientific machine learning (SciML) wrapper for the FEniCS

    FEniCS.jl is a wrapper for the FEniCS library for finite element discretizations of PDEs. This wrapper includes three parts. Installation and direct access to FEniCS via a Conda installation. Alternatively one may use their current FEniCS installation. A low-level development API and provides some functionality to make directly dealing with the library a little bit easier, but still requires knowledge of FEniCS itself. Interfaces have been provided for the main functions and their attributes, and instructions to add further ones can be found here. A high-level API for usage with DifferentialEquations. An example can be seen in solving the heat equation with high-order adaptive time-stepping. Various gists/jupyter notebooks have been created to provide a brief overview of the overall functionality and of any differences between the pythonic FEniCS and the Julian wrapper. DifferentialEquations.jl ecosystem. Paraview can also be used to visualize various results just like in FEniCS.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 18

    Face Recognition

    World's simplest facial recognition api for Python & the command line

    Face Recognition is the world's simplest face recognition library. It allows you to recognize and manipulate faces from Python or from the command line using dlib's (a C++ toolkit containing machine learning algorithms and tools) state-of-the-art face recognition built with deep learning. Face Recognition is highly accurate and is able to do a number of things. It can find faces in pictures, manipulate facial features in pictures, identify faces in pictures, and do face recognition on a folder of images from the command line. It could even do real-time face recognition and blur faces on videos when used with other Python libraries.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 19
    FiftyOne

    FiftyOne

    The open-source tool for building high-quality datasets

    The open-source tool for building high-quality datasets and computer vision models. Nothing hinders the success of machine learning systems more than poor-quality data. And without the right tools, improving a model can be time-consuming and inefficient. FiftyOne supercharges your machine learning workflows by enabling you to visualize datasets and interpret models faster and more effectively. Improving data quality and understanding your model’s failure modes are the most impactful ways to boost the performance of your model. FiftyOne provides the building blocks for optimizing your dataset analysis pipeline. Use it to get hands-on with your data, including visualizing complex labels, evaluating your models, exploring scenarios of interest, identifying failure modes, finding annotation mistakes, and much more! Surveys show that machine learning engineers spend over half of their time wrangling data, but it doesn't have to be that way.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 20
    Finance

    Finance

    150+ quantitative finance Python programs

    Finance is a repository that compiles structured notes and educational material related to financial analysis, markets, and quantitative finance concepts. The project focuses on explaining key principles used in finance and investment analysis, including topics such as financial statements, valuation models, portfolio theory, and financial markets. The repository is designed as a study reference for students and professionals who want to understand financial systems and the analytical frameworks used in financial decision-making. It organizes concepts into structured documents that explain both theoretical principles and practical calculations used in finance. The materials often include definitions, formulas, conceptual explanations, and examples to help readers understand how financial models and instruments function in real markets.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 21
    FlexLLMGen

    FlexLLMGen

    Running large language models on a single GPU

    FlexLLMGen is an open-source inference engine designed to run large language models efficiently on limited hardware resources such as a single GPU. The system focuses on high-throughput generation workloads where large batches of text must be processed quickly, such as large-scale data extraction or document analysis tasks. Instead of requiring expensive multi-GPU systems, the framework uses techniques such as memory offloading, compression, and optimized batching to run large models on commodity hardware. The architecture distributes computation and memory usage across the GPU, CPU, and disk in order to maximize the number of tokens processed during inference. This design allows organizations to deploy powerful language models for high-volume tasks without the infrastructure costs typically associated with large-scale AI systems. The project is particularly useful for workloads that prioritize throughput over latency, including benchmarking experiments and large corpus analysis.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 22
    Hummingbird

    Hummingbird

    Hummingbird compiles trained ML models into tensor computation

    Hummingbird is a library for compiling trained traditional ML models into tensor computations. Hummingbird allows users to seamlessly leverage neural network frameworks (such as PyTorch) to accelerate traditional ML models. Thanks to Hummingbird, users can benefit from (1) all the current and future optimizations implemented in neural network frameworks; (2) native hardware acceleration; (3) having a unique platform to support both traditional and neural network models; and having all of this (4) without having to re-engineer their models.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 23
    Key-book

    Key-book

    Proofs, cases, concept supplements, and reference explanations

    The book "Introduction to Machine Learning Theory" (hereinafter referred to as "Introduction") written by Zhou Zhihua, Wang Wei, Gao Wei, and other teachers fills the regret of the lack of introductory works on machine learning theory in China. This book attempts to provide an introductory guide for readers interested in learning machine learning theory and researching machine learning theory in an easy-to-understand language. "Guide" mainly covers seven parts, corresponding to seven important concepts or theoretical tools in machine learning theory, namely: learnability, (hypothesis space) complexity, generalization bound, stability, consistency, convergence rate, regret circle. Daoyin is a highly theoretical book, involving a large number of mathematical theorems and various proofs. Although the writing team has reduced the difficulty as much as possible, due to the nature of machine learning theory, the book still places high demands on the reader's mathematical background.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 24

    LightGBM

    Gradient boosting framework based on decision tree algorithms

    LightGBM or Light Gradient Boosting Machine is a high-performance, open source gradient boosting framework based on decision tree algorithms. Compared to other boosting frameworks, LightGBM offers several advantages in terms of speed, efficiency and accuracy. Parallel experiments have shown that LightGBM can attain linear speed-up through multiple machines for training in specific settings, all while consuming less memory. LightGBM supports parallel and GPU learning, and can handle large-scale data. It’s become widely-used for ranking, classification and many other machine learning tasks.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 25
    Lightly

    Lightly

    A python library for self-supervised learning on images

    A python library for self-supervised learning on images. We, at Lightly, are passionate engineers who want to make deep learning more efficient. That's why - together with our community - we want to popularize the use of self-supervised methods to understand and curate raw image data. Our solution can be applied before any data annotation step and the learned representations can be used to visualize and analyze datasets. This allows selecting the best core set of samples for model training through advanced filtering. We provide PyTorch, PyTorch Lightning and PyTorch Lightning distributed examples for each of the models to kickstart your project. Lightly requires Python 3.6+ but we recommend using Python 3.7+. We recommend installing Lightly in a Linux or OSX environment. With lightly, you can use the latest self-supervised learning methods in a modular way using the full power of PyTorch. Experiment with different backbones, models, and loss functions.
    Downloads: 1 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB