Open Source Machine Learning Software - Page 16

Machine Learning Software

View 446 business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    Let your crypto work for you

    Put idle assets to work with competitive interest rates, borrow without selling, and trade with precision. All in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 1
    PySINDy

    PySINDy

    A package for the sparse identification of nonlinear dynamical systems

    PySINDy is a Python library that implements the Sparse Identification of Nonlinear Dynamics (SINDy) method for discovering mathematical models of dynamical systems from data. The framework focuses on identifying governing equations that describe the behavior of complex physical systems by selecting sparse combinations of candidate functions. Instead of fitting a purely predictive machine learning model, PySINDy attempts to recover interpretable differential equations that explain how a system evolves over time. This approach is particularly valuable in scientific fields such as physics, engineering, and biology where researchers seek both predictive accuracy and theoretical insight. The library provides tools for constructing libraries of candidate functions, performing sparse regression, and validating discovered models against observed data. It integrates with standard Python scientific computing libraries, making it easy to apply to experimental datasets or simulated systems.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 2
    PyTorch Natural Language Processing

    PyTorch Natural Language Processing

    Basic Utilities for PyTorch Natural Language Processing (NLP)

    PyTorch-NLP is a library for Natural Language Processing (NLP) in Python. It’s built with the very latest research in mind, and was designed from day one to support rapid prototyping. PyTorch-NLP comes with pre-trained embeddings, samplers, dataset loaders, metrics, neural network modules and text encoders. It’s open-source software, released under the BSD3 license. With your batch in hand, you can use PyTorch to develop and train your model using gradient descent. For example, check out this example code for training on the Stanford Natural Language Inference (SNLI) Corpus. Now you've setup your pipeline, you may want to ensure that some functions run deterministically. Wrap any code that's random, with fork_rng and you'll be good to go. Now that you've computed your vocabulary, you may want to make use of pre-trained word vectors to set your embeddings.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 3
    Quantitative Trading System

    Quantitative Trading System

    A comprehensive quantitative trading system with AI-powered analysis

    Quantitative Trading System is a comprehensive quantitative trading platform that integrates artificial intelligence, financial data analysis, and automated strategy execution within a unified software system. The project is designed to provide an end-to-end infrastructure for building and operating algorithmic trading strategies in financial markets. It includes tools for collecting and processing market data from multiple sources, performing statistical and machine learning analysis, and generating trading signals based on quantitative models. The system supports real-time data streaming, allowing strategies to respond to market conditions as they evolve. QuantMuse also incorporates advanced risk management features, including portfolio monitoring, risk limits, and dynamic position sizing to control exposure.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 4
    RecBole

    RecBole

    A unified, comprehensive and efficient recommendation library

    A unified, comprehensive and efficient recommendation library. We design general and extensible data structures to unify the formatting and usage of various recommendation datasets. We implement more than 100 commonly used recommendation algorithms and provide formatted copies of 28 recommendation datasets. We support a series of widely adopted evaluation protocols or settings for testing and comparing recommendation algorithms. RecBole is developed based on Python and PyTorch for reproducing and developing recommendation algorithms in a unified, comprehensive and efficient framework for research purpose. It can be installed from pip, conda and source, and is easy to use. We have implemented more than 100 recommender system models, covering four common recommender system categories in RecBole and eight toolkits of RecBole2.0, including General Recommendation, Sequential Recommendation, Context-aware Recommendation, and Knowledge-based Recommendation and sub-packages.
    Downloads: 2 This Week
    Last Update:
    See Project
  • Full-stack observability with actually useful AI | Grafana Cloud Icon
    Full-stack observability with actually useful AI | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 5
    Seldon Core

    Seldon Core

    An MLOps framework to package, deploy, monitor and manage models

    The de facto standard open-source platform for rapidly deploying machine learning models on Kubernetes. Seldon Core, our open-source framework, makes it easier and faster to deploy your machine learning models and experiments at scale on Kubernetes. Seldon Core serves models built in any open-source or commercial model building framework. You can make use of powerful Kubernetes features like custom resource definitions to manage model graphs. And then connect your continuous integration and deployment (CI/CD) tools to scale and update your deployment. Built on Kubernetes, runs on any cloud and on-premises. Framework agnostic, supports top ML libraries, toolkits and languages. Advanced deployments with experiments, ensembles and transformers. Our open-source framework makes it easier and faster to deploy your machine learning models and experiments at scale on Kubernetes. The Kubeflow project is dedicated to making deployments of machine learning (ML) workflows on Kubernetes.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 6
    Simd Library

    Simd Library

    C++ image processing and machine learning library with using of SIMD

    The Simd Library is a free open-source image processing and machine learning library, designed for C and C++ programmers. It provides many useful high-performance algorithms for image processing such as pixel format conversion, image scaling and filtration, extraction of statistical information from images, motion detection, object detection and classification, neural networks. The algorithms are optimized with using of different SIMD CPU extensions. In particular, the library supports the following CPU extensions: SSE, AVX, AVX-512, and AMX for x86/x64, and NEON for ARM. The Simd Library has C API and also contains useful C++ classes and functions to facilitate access to C API. The library supports dynamic and static linking, 32-bit and 64-bit Windows and Linux, MSVS, G++ and Clang compilers, MSVS projects, and CMake build systems.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 7
    Spotlight

    Spotlight

    Deep recommender models using PyTorch

    Spotlight uses PyTorch to build both deep and shallow recommender models. By providing both a slew of building blocks for loss functions (various pointwise and pairwise ranking losses), representations (shallow factorization representations, deep sequence models), and utilities for fetching (or generating) recommendation datasets, it aims to be a tool for rapid exploration and prototyping of new recommender models. Spotlight offers a slew of popular datasets, including Movielens 100K, 1M, 10M, and 20M. It also incorporates utilities for creating synthetic datasets. For example, generate_sequential generates a Markov-chain-derived interaction dataset, where the next item a user chooses is a function of their previous interactions. Recommendations can be seen as a sequence prediction task: given the items a user has interacted with in the past, what will be the next item they will interact with? Spotlight provides a range of models.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 8
    The Hundred-Page Machine Learning Book

    The Hundred-Page Machine Learning Book

    The Python code to reproduce illustrations from Machine Learning Book

    The Hundred-Page Machine Learning Book is the official companion repository for The Hundred-Page Machine Learning Book written by machine learning researcher Andriy Burkov. The repository contains Python code used to generate the figures, visualizations, and illustrative examples presented in the book. Its purpose is to help readers better understand the concepts explained in the text by allowing them to run and experiment with the underlying code themselves. The book itself provides a concise overview of machine learning theory and practice, covering topics such as supervised learning, unsupervised learning, neural networks, and optimization algorithms. The repository complements these explanations by offering practical implementations that demonstrate how various algorithms behave when applied to data. Readers can explore the scripts to reproduce diagrams and observe how mathematical concepts translate into working code.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 9
    TorchRec

    TorchRec

    Pytorch domain library for recommendation systems

    TorchRec is a PyTorch domain library built to provide common sparsity & parallelism primitives needed for large-scale recommender systems (RecSys). It allows authors to train models with large embedding tables sharded across many GPUs. Parallelism primitives that enable easy authoring of large, performant multi-device/multi-node models using hybrid data-parallelism/model-parallelism. The TorchRec sharder can shard embedding tables with different sharding strategies including data-parallel, table-wise, row-wise, table-wise-row-wise, and column-wise sharding. The TorchRec planner can automatically generate optimized sharding plans for models. Pipelined training overlaps dataloading device transfer (copy to GPU), inter-device communications (input_dist), and computation (forward, backward) for increased performance. Optimized kernels for RecSys powered by FBGEMM. Quantization support for reduced precision training and inference. Common modules for RecSys.
    Downloads: 2 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 10
    TorchServe

    TorchServe

    Serve, optimize and scale PyTorch models in production

    TorchServe is a performant, flexible and easy-to-use tool for serving PyTorch eager mode and torschripted models. Multi-model management with the optimized worker to model allocation. REST and gRPC support for batched inference. Export your model for optimized inference. Torchscript out of the box, ORT, IPEX, TensorRT, FasterTransformer. Performance Guide: built-in support to optimize, benchmark and profile PyTorch and TorchServe performance. Expressive handlers: An expressive handler architecture that makes it trivial to support inferencing for your use case with many supported out of the box. Out-of-box support for system-level metrics with Prometheus exports, custom metrics and PyTorch profiler support.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 11
    Trax

    Trax

    Deep learning with clear code and speed

    Trax is an end-to-end library for deep learning that focuses on clear code and speed. It is actively used and maintained in the Google Brain team. Run a pre-trained Transformer, create a translator in a few lines of code. Features and resources, API docs, where to talk to us, how to open an issue and more. Walkthrough, how Trax works, how to make new models and train on your own data. Trax includes basic models (like ResNet, LSTM, Transformer) and RL algorithms (like REINFORCE, A2C, PPO). It is also actively used for research and includes new models like the Reformer and new RL algorithms like AWR. Trax has bindings to a large number of deep learning datasets, including Tensor2Tensor and TensorFlow datasets. You can use Trax either as a library from your own python scripts and notebooks or as a binary from the shell, which can be more convenient for training large models. It runs without any changes on CPUs, GPUs and TPUs.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 12
    Unity ML-Agents Toolkit

    Unity ML-Agents Toolkit

    Unity machine learning agents toolkit

    Train and embed intelligent agents by leveraging state-of-the-art deep learning technology. Creating responsive and intelligent virtual players and non-playable game characters is hard. Especially when the game is complex. To create intelligent behaviors, developers have had to resort to writing tons of code or using highly specialized tools. With Unity Machine Learning Agents (ML-Agents), you are no longer “coding” emergent behaviors, but rather teaching intelligent agents to “learn” through a combination of deep reinforcement learning and imitation learning. Using ML-Agents allows developers to create more compelling gameplay and an enhanced game experience. Advancement of artificial intelligence (AI) research depends on figuring out tough problems in existing environments using current benchmarks for training AI models. Using Unity and the ML-Agents toolkit, you can create AI environments that are physically, visually, and cognitively rich.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 13
    Velocity server

    Velocity server

    The modern, next-generation Minecraft server proxy

    Velocity is a high-performance Minecraft proxy server developed by the PaperMC project that allows multiple Minecraft servers to be linked together into a single network. Acting as an intermediary between players and backend servers, Velocity manages player connections and routes them to different game servers within a network. This architecture allows large Minecraft communities to run multiple servers for different game modes while presenting them as a unified system to players. The software is designed with a focus on performance, scalability, and modern architecture, allowing it to handle thousands of simultaneous players efficiently. Velocity also includes a plugin API that allows developers to extend the proxy with custom functionality and integrate it with existing server tools. Compared with older proxy solutions, the project emphasizes improved performance, reliability, and cleaner software design.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 14
    WaveFunctionCollapse

    WaveFunctionCollapse

    Bitmap & tilemap generation from a single example

    This program generates bitmaps that are locally similar to the input bitmap. WFC initializes output bitmap in a completely unobserved state, where each pixel value is in superposition of colors of the input bitmap (so if the input was black & white then the unobserved states are shown in different shades of grey). The coefficients in these superpositions are real numbers, not complex numbers, so it doesn't do the actual quantum mechanics, but it was inspired by QM. Then the program goes into the observation-propagation cycle. It may happen that during propagation all the coefficients for a certain pixel become zero. That means that the algorithm has run into a contradiction and can not continue. The problem of determining whether a certain bitmap allows other nontrivial bitmaps satisfying condition (C1) is NP-hard, so it's impossible to create a fast solution that always finishes. In practice, however, the algorithm runs into contradictions surprisingly rarely.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 15
    captcha_break

    captcha_break

    Identification codes

    This project will use Keras to build a deep convolutional neural network to identify the captcha verification code. It is recommended to use a graphics card to run the project. The following visualization codes are jupyter notebookall done in . If you want to write a python script, you can run it normally with a little modification. Of course, you can also remove these visualization codes. captcha is a library written in python to generate verification codes. It supports image verification codes and voice verification codes. We use its function of generating image verification codes. First, we set our verification code format to numbers and capital letters, and generate a string of verification codes. It is well known that tensorflow occupies all video memory by default, which is not conducive to us conducting multiple experiments at the same time, so we can use the following code when tensorflow uses the video memory it needs instead of directly occupying all video memory.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 16
    spaGO

    spaGO

    Self-contained Machine Learning and Natural Language Processing lib

    A Machine Learning library written in pure Go designed to support relevant neural architectures in Natural Language Processing. Spago is self-contained, in that it uses its own lightweight computational graph both for training and inference, easy to understand from start to finish. The core module of Spago relies only on testify for unit testing. In other words, it has "zero dependencies", and we are committed to keeping it that way as much as possible. Spago uses a multi-module workspace to ensure that additional dependencies are downloaded only when specific features (e.g. persistent embeddings) are used. A good place to start is by looking at the implementation of built-in neural models, such as the LSTM. Except for a few linear algebra operations written in assembly for optimal performance (a bit of copying from Gonum), it's straightforward Go code, so you don't have to worry.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 17
    talos

    talos

    Hyperparameter Optimization for TensorFlow, Keras and PyTorch

    Talos radically changes the ordinary Keras, TensorFlow (tf.keras), and PyTorch workflow by fully automating hyperparameter tuning and model evaluation. Talos exposes Keras and TensorFlow (tf.keras) and PyTorch functionality entirely and there is no new syntax or templates to learn. Talos is made for data scientists and data engineers that want to remain in complete control of their TensorFlow (tf.keras) and PyTorch models, but are tired of mindless parameter hopping and confusing optimization solutions that add complexity instead of reducing it. Within minutes, without learning any new syntax, Talos allows you to configure, perform, and evaluate hyperparameter optimization experiments that yield state-of-the-art results across a wide range of prediction tasks. Talos provides the simplest and yet most powerful available method for hyperparameter optimization with TensorFlow (tf.keras) and PyTorch.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 18
    jMIR

    jMIR

    Music research software

    jMIR is an open-source software suite implemented in Java for use in music information retrieval (MIR) research. It can be used to study music in the form of audio recordings, symbolic encodings and lyrical transcriptions, and can also mine cultural information from the Internet. It also includes tools for managing and profiling large music collections and for checking audio for production errors. jMIR includes software for extracting features, applying machine learning algorithms, applying heuristic error error checkers, mining metadata and analyzing metadata.
    Downloads: 29 This Week
    Last Update:
    See Project
  • 19
    This project contains weka packages of neural networks algorithms implementations like Learning Vector Quantizer (LVQ) and Self-organizing Maps (SOM). For more information about weka, please visit http://www.cs.waikato.ac.nz/~ml/weka/
    Leader badge
    Downloads: 46 This Week
    Last Update:
    See Project
  • 20
    Scikit Learn
    Machine Learning framework in Python
    Downloads: 12 This Week
    Last Update:
    See Project
  • 21
    OpenNN - Open Neural Networks Library

    OpenNN - Open Neural Networks Library

    Machine learning algorithms for advanced analytics

    OpenNN is a software library written in C++ for advanced analytics. It implements neural networks, the most successful machine learning method. Some typical applications of OpenNN are business intelligence (customer segmentation, churn prevention…), health care (early diagnosis, microarray analysis…) and engineering (performance optimization, predictive maitenance…). OpenNN does not deal with computer vision or natural language processing. The main advantage of OpenNN is its high performance. This library outstands in terms of execution speed and memory allocation. It is constantly optimized and parallelized in order to maximize its efficiency. The documentation is composed by tutorials and examples to offer a complete overview about the library. OpenNN is developed by Artelnics, a company specialized in artificial intelligence.
    Downloads: 10 This Week
    Last Update:
    See Project
  • 22
    MODLEM

    MODLEM

    rule-based, WEKA compatible, Machine Learning algorithm

    This project is a WEKA (Waikato Environment for Knowledge Analysis) compatible implementation of MODLEM - a Machine Learning algorithm which induces minimum set of rules. These rules can be adopted as a classifier (in terms of ML). It is a sequential covering algorithm, which was invented to cope with numeric data without discretization. Actually the nominal and numeric attributes are treated in the same way: attribute's space is being searched to find the best rule condition during rule induction. In result numeric attribute's conditions are more precise and closely describe the class. This algorithm contains some aspects of Rough Set Theory: the class definition can be described accordingly to its lower or upper approximation. For more information, see: Stefanowski, Jerzy. The rough set based rule induction technique for classification problems. In: Proc. 6th European Congress on Intelligent Techniques and Soft Computing, vol. 1. Aachen, 1998. s. 109-113.
    Leader badge
    Downloads: 17 This Week
    Last Update:
    See Project
  • 23

    sgmweka

    Weka wrapper for the SGM toolkit for text classification and modeling.

    Weka wrapper for the SGM toolkit for text classification and modeling. Provides Sparse Generative Models for scalable and accurate text classification and modeling for use in high-speed and large-scale text mining. Has lower time complexity of classification than comparable software due to inference based on sparse model representation and use of an inverted index. The provided .zip file is in the Weka package format, giving access to text classification. Other functions are usable through either Java command-line commands or class inclusion into Java projects.
    Leader badge
    Downloads: 29 This Week
    Last Update:
    See Project
  • 24
    openEAR is the Munich Open-Source Emotion and Affect Recognition Toolkit developed at the Technische Universität München (TUM). It provides efficient (audio) feature extraction algorithms implemented in C++, classfiers, and pre-trained models on well-known emotion databases. It is now maintained and supported by audEERING. Updates will follow soon.
    Downloads: 7 This Week
    Last Update:
    See Project
  • 25
    AI Cheatsheets

    AI Cheatsheets

    Essential Cheat Sheets for deep learning and machine learning research

    cheatsheets-ai is an open-source repository that collects essential cheat sheets covering many tools and concepts used in machine learning, deep learning, and data science. The project aims to provide quick-reference materials that help engineers, researchers, and students review key techniques and frameworks without reading extensive documentation. It compiles cheat sheets for widely used libraries and technologies such as TensorFlow, Keras, NumPy, Pandas, Scikit-learn, Matplotlib, and PySpark. These materials summarize common functions, workflows, and best practices in a concise visual format that makes them easy to consult during development or study sessions. The repository functions as a centralized library where users can quickly access reference materials for both machine learning theory and practical programming tools. Many of the cheat sheets are available as downloadable PDFs and images, allowing learners to keep them as quick references while working on projects.
    Downloads: 1 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB