Open Source Linux Machine Learning Software - Page 21

Machine Learning Software for Linux

View 57 business solutions
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 1
    Daft

    Daft

    Distributed DataFrame for Python designed for the cloud

    Daft is a framework for ETL, analytics and ML/AI at scale. Its familiar Python Dataframe API is built to outperform Spark in performance and ease of use. Daft plugs directly into your ML/AI stack through efficient zero-copy integrations with essential Python libraries such as Pytorch and Ray. It also allows requesting GPUs as a resource for running models. Daft runs locally with a lightweight multithreaded backend. When your local machine is no longer sufficient, it scales seamlessly to run out-of-core on a distributed cluster. Underneath its Python API, Daft is built in blazing fast Rust code. Rust powers Daft’s vectorized execution and async I/O, allowing Daft to outperform frameworks such as Spark.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    Data Science Articles from CodeCut

    Data Science Articles from CodeCut

    Collection of useful data science topics along with articles

    The Data-science repository from CodeCutTech is a curated collection of educational content focused on practical tools and workflows used in modern data science projects. Instead of providing a single software package, the repository aggregates articles, tutorials, and examples covering many topics within the data science ecosystem. The materials address areas such as MLOps, data management, project organization, testing practices, visualization techniques, and productivity tools used by data scientists. Each topic often includes references to code repositories, demonstrations, and video tutorials that show how the tools can be applied in real projects. The repository is intended to help practitioners stay updated with current best practices and technologies in the field of data science.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Data Science Collected Resources

    Data Science Collected Resources

    Carefully curated resource links for data science in one place

    Data Science Collected Resources is a curated collection of learning materials and reference links covering a wide range of topics in data science, artificial intelligence, and machine learning. The repository aggregates educational resources from research articles, technical blogs, tutorials, and documentation into a single organized knowledge hub. Its goal is to provide learners and practitioners with easy access to high-quality resources related to data science tools, programming languages, cloud platforms, and machine learning techniques. The repository includes links to materials discussing topics such as artificial intelligence research, AWS infrastructure, machine learning algorithms, and data analysis tools. It also contains supplementary documents like cheat sheets and machine learning notes that help readers review important concepts quickly.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Data Science Interviews

    Data Science Interviews

    Data science interview questions and answers

    Data Science Interviews is an open-source repository that collects common data science interview questions along with community-provided answers and explanations. The project serves as a preparation resource for students, job seekers, and professionals who want to review the technical knowledge required for data science roles. The repository organizes questions into different categories including theoretical machine learning concepts, technical programming questions, and probability or statistics problems. Many of the questions cover fundamental machine learning topics such as linear models, decision trees, neural networks, and evaluation metrics. In addition to theoretical questions, the repository also includes practical interview topics related to coding challenges, SQL queries, and algorithmic thinking.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 5
    Data-Science-Interview-Questions-Answers

    Data-Science-Interview-Questions-Answers

    Curated list of data science interview questions and answers

    Data-Science-Interview-Questions-Answers is a curated educational repository designed to help data science candidates prepare for technical interviews by organizing a large bank of questions and answers in one place. It began as a daily interview question initiative and was later consolidated into GitHub so learners could review the material more easily and revisit it over time. The repository focuses on core data science fundamentals rather than acting as a software framework, which makes it especially useful as a study and revision resource. Its content is organized into subject-specific documents that cover machine learning, deep learning, statistics, probability, Python, SQL and databases, and resume-based interview questions. That structure makes it practical for users who want to study by topic, strengthen weak areas, or simulate the range of questions they may encounter in interviews.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    DataDrivenDiffEq.jl

    DataDrivenDiffEq.jl

    Data driven modeling and automated discovery of dynamical systems

    DataDrivenDiffEq.jl is a package for finding systems of equations automatically from a dataset. The methods in this package take in data and return the model which generated the data. A known model is not required as input. These methods can estimate equation-free and equation-based models for discrete, continuous differential equations or direct mappings.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Datapipe

    Datapipe

    Real-time, incremental ETL library for ML with record-level depend

    Datapipe is a real-time, incremental ETL library for Python with record-level dependency tracking. Datapipe is designed to streamline the creation of data processing pipelines. It excels in scenarios where data is continuously changing, requiring pipelines to adapt and process only the modified data efficiently. This library tracks dependencies for each record in the pipeline, ensuring minimal and efficient data processing.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    DeeProtGO

    DeeProtGO

    DeeProtGO is a deep learning model for predicting GO terms of proteins

    This project contains the source code of DeeProtGO as well as an example of its use when predicting GO terms of the biological process sub-ontology for eukaryotic proteins.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    Deep Java Library (DJL)

    Deep Java Library (DJL)

    An engine-agnostic deep learning framework in Java

    Deep Java Library (DJL) is an open-source, high-level, engine-agnostic Java framework for deep learning. DJL is designed to be easy to get started with and simple to use for Java developers. DJL provides native Java development experience and functions like any other regular Java library. You don't have to be a machine learning/deep learning expert to get started. You can use your existing Java expertise as an on-ramp to learn and use machine learning and deep learning. You can use your favorite IDE to build, train, and deploy your models. DJL makes it easy to integrate these models with your Java applications. Because DJL is deep learning engine agnostic, you don't have to make a choice between engines when creating your projects. You can switch engines at any point. To ensure the best performance, DJL also provides automatic CPU/GPU choice based on hardware configuration.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8 Monitoring Tools in One APM. Install in 5 Minutes. Icon
    8 Monitoring Tools in One APM. Install in 5 Minutes.

    Errors, performance, logs, uptime, hosts, anomalies, dashboards, and check-ins. One interface.

    AppSignal works out of the box for Ruby, Elixir, Node.js, Python, and more. 30-day free trial, no credit card required.
    Start Free
  • 10
    Deep Learning Drizzle

    Deep Learning Drizzle

    Drench yourself in Deep Learning, Reinforcement Learning

    Drench yourself in Deep Learning, Reinforcement Learning, Machine Learning, Computer Vision, and NLP by learning from these exciting lectures! Optimization courses which form the foundation for ML, DL, RL. Computer Vision courses which are DL & ML heavy. Speech recognition courses which are DL heavy. Structured Courses on Geometric, Graph Neural Networks. Section on Autonomous Vehicles. Section on Computer Graphics with ML/DL focus.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Deep Learning Interviews book

    Deep Learning Interviews book

    Hundreds of fully solved job interview questions

    The interviews.ai repository hosts the open materials for the book Deep Learning Interviews, a comprehensive collection of technical questions and fully solved problems covering many aspects of artificial intelligence. The project was created to help students, researchers, and engineers prepare for machine learning and deep learning interviews by providing structured explanations of key concepts. The repository organizes problems across topics such as neural networks, optimization, probabilistic models, and mathematical foundations of machine learning. Each question is accompanied by detailed solutions that explain the reasoning behind the answers and the theoretical concepts involved. In addition to interview preparation, the material also serves as a condensed overview of many core topics taught in graduate-level machine learning programs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Deep Learning with Keras and Tensorflow

    Deep Learning with Keras and Tensorflow

    Introduction to Deep Neural Networks with Keras and Tensorflow

    Introduction to Deep Neural Networks with Keras and Tensorflow. To date tensorflow comes in two different packages, namely tensorflow and tensorflow-gpu, whether you want to install the framework with CPU-only or GPU support, respectively. NVIDIA Drivers and CuDNN must be installed and configured before hand. Please refer to the official Tensorflow documentation for further details. Since version 0.9 Theano introduced the libgpuarray in the stable release (it was previously only available in the development version). The goal of libgpuarray is (from the documentation) make a common GPU ndarray (n dimensions array) that can be reused by all projects that is as future proof as possible, while keeping it easy to use for simple need/quick test. The easiest way to get (most) these is to use an all-in-one installer such as Anaconda from Continuum. These are available for multiple architectures.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Deep Learning with PyTorch

    Deep Learning with PyTorch

    Latest techniques in deep learning and representation learning

    This course concerns the latest techniques in deep learning and representation learning, focusing on supervised and unsupervised deep learning, embedding methods, metric learning, convolutional and recurrent nets, with applications to computer vision, natural language understanding, and speech recognition. The prerequisites include DS-GA 1001 Intro to Data Science or a graduate-level machine learning course. To be able to follow the exercises, you are going to need a laptop with Miniconda (a minimal version of Anaconda) and several Python packages installed. The following instruction would work as is for Mac or Ubuntu Linux users, Windows users would need to install and work in the Git BASH terminal. JupyterLab has a built-in selectable dark theme, so you only need to install something if you want to use the classic notebook interface.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Deep Photo Style Transfer

    Deep Photo Style Transfer

    Code and data for paper "Deep Photo Style Transfer"

    Deep Photo Style Transfer is an implementation of the algorithm described in the paper “Deep Photo Style Transfer” (arXiv 1703.07511). The software allows users to transfer the style of one photograph to another while preserving photorealism and semantic consistency. It relies on semantic segmentation masks to guide style transfer (so that e.g. sky maps to sky, building maps to building), and uses a matting Laplacian regularization term to ensure smooth transitions. The repository provides code in Torch (Lua), MATLAB / Octave scripts for computing the Laplacian, and pre-trained models. Pretrained models and example scripts for ease of use. Compatibility with MATLAB / Octave for Laplacian computations.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Deep Reinforcement Learning for Keras

    Deep Reinforcement Learning for Keras

    Deep Reinforcement Learning for Keras.

    keras-rl implements some state-of-the-art deep reinforcement learning algorithms in Python and seamlessly integrates with the deep learning library Keras. Furthermore, keras-rl works with OpenAI Gym out of the box. This means that evaluating and playing around with different algorithms is easy. Of course, you can extend keras-rl according to your own needs. You can use built-in Keras callbacks and metrics or define your own. Even more so, it is easy to implement your own environments and even algorithms by simply extending some simple abstract classes. Documentation is available online.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Deep learning time series forecasting

    Deep learning time series forecasting

    Deep learning PyTorch library for time series forecasting

    Example image Flow Forecast (FF) is an open-source deep learning for time series forecasting framework. It provides all the latest state-of-the-art models (transformers, attention models, GRUs) and cutting-edge concepts with easy-to-understand interpretability metrics, cloud provider integration, and model serving capabilities. Flow Forecast was the first time series framework to feature support for transformer-based models and remains the only true end-to-end deep learning for time series forecasting framework. Currently, Task-TS from CoronaWhy primarily maintains this repository. Pull requests are welcome. Historically, this repository provided open-source benchmarks and codes for flash flood and river flow forecasting. Full transformer (SimpleTransformer in model_dict): The full original transformer with all 8 encoder and decoder blocks. Requires passing the target in at inference.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Deep-Learning-for-Recommendation-Systems

    Deep-Learning-for-Recommendation-Systems

    This repository contains Deep Learning based articles

    Deep-Learning-for-Recommendation-Systems is a curated repository that aggregates research papers, articles, and code related to deep learning methods for recommender systems. The project organizes influential academic work covering topics such as collaborative filtering, neural recommendation models, and deep feature learning. It includes references to papers describing architectures like collaborative deep learning, neural autoregressive models, and convolutional approaches to recommendation. The repository also provides links to implementations and external code repositories that demonstrate how these algorithms can be applied in real systems. By compiling research literature and practical resources in one location, the project helps researchers and engineers explore the evolving landscape of recommendation technologies. It highlights both theoretical innovations and applied engineering work used in modern recommendation engines.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    DeepCTR

    DeepCTR

    Package of deep-learning based CTR models

    DeepCTR is a Easy-to-use,Modular and Extendible package of deep-learning based CTR models along with lots of core components layers which can be used to easily build custom models. You can use any complex model with model.fit(), and model.predict(). Provide tf.keras.Model like interface for quick experiment. Provide tensorflow estimator interface for large scale data and distributed training. It is compatible with both tf 1.x and tf 2.x. With the great success of deep learning,DNN-based techniques have been widely used in CTR prediction task. The data in CTR estimation task usually includes high sparse,high cardinality categorical features and some dense numerical features. Since DNN are good at handling dense numerical features,we usually map the sparse categorical features to dense numerical through embedding technique.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    DeepCTR-Torch

    DeepCTR-Torch

    Easy-to-use,Modular and Extendible package of deep-learning models

    DeepCTR-Torch is an easy-to-use, Modular and Extendible package of deep-learning-based CTR models along with lots of core components layers that can be used to build your own custom model easily.It is compatible with PyTorch.You can use any complex model with model.fit() and model.predict(). With the great success of deep learning, DNN-based techniques have been widely used in CTR estimation tasks. The data in the CTR estimation task usually includes high sparse,high cardinality categorical features and some dense numerical features. Low-order Extractor learns feature interaction through product between vectors. Factorization-Machine and it’s variants are widely used to learn the low-order feature interaction. High-order Extractor learns feature combination through complex neural network functions like MLP, Cross Net, etc.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    DeepDetect

    DeepDetect

    Deep Learning API and Server in C++14 support for Caffe, PyTorch

    The core idea is to remove the error sources and difficulties of Deep Learning applications by providing a safe haven of commoditized practices, all available as a single core. While the Open Source Deep Learning Server is the core element, with REST API, and multi-platform support that allows training & inference everywhere, the Deep Learning Platform allows higher level management for training neural network models and using them as if they were simple code snippets. Ready for applications of image tagging, object detection, segmentation, OCR, Audio, Video, Text classification, CSV for tabular data and time series. Neural network templates for the most effective architectures for GPU, CPU, and Embedded devices. Training in a few hours and with small data thanks to 25+ pre-trained models. Full Open Source, with an ecosystem of tools (API clients, video, annotation, ...) Fast Server written in pure C++, a single codebase for Cloud, Desktop & Embedded.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    DeepLearn

    DeepLearn

    Implementation of research papers on Deep Learning+ NLP+ CV in Python

    Welcome to DeepLearn. This repository contains an implementation of the following research papers on NLP, CV, ML, and deep learning. The required dependencies are mentioned in requirement.txt. I will also use dl-text modules for preparing the datasets. If you haven't use it, please do have a quick look at it. CV, transfer learning, representation learning.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    DeepLearning

    DeepLearning

    Deep Learning (Flower Book) mathematical derivation

    " Deep Learning " is the only comprehensive book in the field of deep learning. The full name is also called the Deep Learning AI Bible (Deep Learning) . It is edited by three world-renowned experts, Ian Goodfellow, Yoshua Bengio, and Aaron Courville. Includes linear algebra, probability theory, information theory, numerical optimization, and related content in machine learning. At the same time, it also introduces deep learning techniques used by practitioners in the industry, including deep feedforward networks, regularization, optimization algorithms, convolutional networks, sequence modeling and practical methods, and investigates topics such as natural language processing, Applications in speech recognition, computer vision, online recommender systems, bioinformatics, and video games. Finally, the Deep Learning book provides research directions covering theoretical topics including linear factor models, autoencoders, representation learning, structured probabilistic models, etc.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    DeepLearningProject

    DeepLearningProject

    An in-depth machine learning tutorial

    This tutorial tries to do what most Most Machine Learning tutorials available online do not. It is not a 30 minute tutorial that teaches you how to "Train your own neural network" or "Learn deep learning in under 30 minutes". It's a full pipeline which you would need to do if you actually work with machine learning - introducing you to all the parts, and all the implementation decisions and details that need to be made. The dataset is not one of the standard sets like MNIST or CIFAR, you will make you very own dataset. Then you will go through a couple conventional machine learning algorithms, before finally getting to deep learning! In the fall of 2016, I was a Teaching Fellow (Harvard's version of TA) for the graduate class on "Advanced Topics in Data Science (CS209/109)" at Harvard University. I was in charge of designing the class project given to the students, and this tutorial has been built on top of the project I designed for the class.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    DeepPavlov

    DeepPavlov

    A library for deep learning end-to-end dialog systems and chatbots

    DeepPavlov makes it easy for beginners and experts to create dialogue systems. The best place to start is with user-friendly tutorials. They provide quick and convenient introduction on how to use DeepPavlov with complete, end-to-end examples. No installation needed. Guides explain the concepts and components of DeepPavlov. Follow step-by-step instructions to install, configure and extend DeepPavlov framework for your use case. DeepPavlov is an open-source framework for chatbots and virtual assistants development. It has comprehensive and flexible tools that let developers and NLP researchers create production-ready conversational skills and complex multi-skill conversational assistants. Use BERT and other state-of-the-art deep learning models to solve classification, NER, Q&A and other NLP tasks. DeepPavlov Agent allows building industrial solutions with multi-skill integration via API services.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    DeepTraffic

    DeepTraffic

    DeepTraffic is a deep reinforcement learning competition

    DeepTraffic is a deep reinforcement learning simulation designed to teach and evaluate autonomous driving algorithms in a dense highway environment. The system presents a simulated multi-lane highway where an AI-controlled vehicle must navigate traffic while maximizing speed and avoiding collisions. Participants design neural network policies that determine the vehicle’s actions, such as accelerating, decelerating, changing lanes, or maintaining speed. The project was created as part of an educational competition associated with MIT’s deep learning courses, encouraging students and researchers to experiment with reinforcement learning techniques. The environment provides a coding interface where users can design neural network architectures and tune hyperparameters while observing their agent’s performance in a visual simulation.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB