Open Source Machine Learning Software - Page 18

Machine Learning Software

View 446 business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • 1
    Hugging Face Transformer

    Hugging Face Transformer

    CPU/GPU inference server for Hugging Face transformer models

    Optimize and deploy in production Hugging Face Transformer models in a single command line. At Lefebvre Dalloz we run in-production semantic search engines in the legal domain, in the non-marketing language it's a re-ranker, and we based ours on Transformer. In that setup, latency is key to providing a good user experience, and relevancy inference is done online for hundreds of snippets per user query. Most tutorials on Transformer deployment in production are built over Pytorch and FastAPI. Both are great tools but not very performant in inference. Then, if you spend some time, you can build something over ONNX Runtime and Triton inference server. You will usually get from 2X to 4X faster inference compared to vanilla Pytorch. It's cool! However, if you want the best in class performances on GPU, there is only a single possible combination: Nvidia TensorRT and Triton. You will usually get 5X faster inference compared to vanilla Pytorch.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 2
    Humanoid-Gym

    Humanoid-Gym

    Reinforcement Learning for Humanoid Robot with Zero-Shot Sim2Real

    Humanoid-Gym is a reinforcement learning framework designed to train locomotion and control policies for humanoid robots using high-performance simulation environments. The system is built on top of NVIDIA Isaac Gym, which allows large-scale parallel simulation of robotic environments directly on GPU hardware. Its primary goal is to enable efficient training of humanoid robots in simulation while enabling policies to transfer effectively to real-world hardware without additional training. The framework emphasizes the concept of zero-shot sim-to-real transfer, meaning that behaviors learned in simulation can be deployed directly on physical robots with minimal adjustment. To improve reliability and generalization, the framework also includes sim-to-sim validation pipelines that test trained policies across different physics engines.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 3
    Kodezi Chronos

    Kodezi Chronos

    Kodezi Chronos is a debugging-first language model

    Kodezi Chronos is a research project focused on developing a specialized language model designed specifically for debugging software and understanding large code repositories. Unlike general-purpose language models that focus primarily on code generation, Chronos is built to diagnose and repair bugs by analyzing complex relationships across files within a codebase. The project introduces architectural techniques such as Adaptive Graph-Guided Retrieval, which allows the system to navigate large repositories and retrieve relevant debugging information from multiple sources. Another component, Persistent Debug Memory, allows the system to learn patterns from past debugging sessions and apply that knowledge to future problems. The repository mainly contains research documentation, evaluation benchmarks, and experimental frameworks rather than the full proprietary model implementation.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 4
    LSTMs for Human Activity Recognition

    LSTMs for Human Activity Recognition

    Human Activity Recognition example using TensorFlow on smartphone

    LSTM-Human-Activity-Recognition is a machine learning project that demonstrates how recurrent neural networks can be used to recognize human activities from sensor data. The repository implements a deep learning model based on Long Short-Term Memory (LSTM) networks to classify physical activities using time-series data collected from wearable sensors. The project uses the well-known Human Activity Recognition dataset derived from smartphone accelerometer and gyroscope signals. Through the use of sequential neural network architectures, the system learns patterns in motion data that correspond to activities such as walking, sitting, standing, or climbing stairs. The repository includes data preprocessing scripts, neural network architecture definitions, and training pipelines that allow researchers to reproduce and modify the experiments. It serves as an educational example of how deep learning models can process temporal sensor signals for pattern recognition tasks.
    Downloads: 1 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 5
    Learn_Machine_Learning_in_3_Months

    Learn_Machine_Learning_in_3_Months

    This is the code for "Learn Machine Learning in 3 Months"

    This repository outlines an ambitious self-study curriculum for learning machine learning in roughly three months, emphasizing breadth, momentum, and hands-on practice. It sequences core topics—math foundations, classic ML, deep learning, and applied projects—so learners can pace themselves week by week. The plan mixes reading, lectures, coding assignments, and small build-it-yourself projects to reinforce understanding through repetition and implementation. Because ML is a wide field, the curriculum favors pragmatic coverage over academic completeness, pointing learners to widely used tools and approachable resources. It’s intended to help beginners overcome decision paralysis by giving a concrete schedule and a minimal set of action-oriented tasks.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 6
    Lip Reading

    Lip Reading

    Cross Audio-Visual Recognition using 3D Architectures

    The input pipeline must be prepared by the users. This code is aimed to provide the implementation for Coupled 3D Convolutional Neural Networks for audio-visual matching. Lip-reading can be a specific application for this work. Audio-visual recognition (AVR) has been considered as a solution for speech recognition tasks when the audio is corrupted, as well as a visual recognition method used for speaker verification in multi-speaker scenarios. The approach of AVR systems is to leverage the extracted information from one modality to improve the recognition ability of the other modality by complementing the missing information. The essential problem is to find the correspondence between the audio and visual streams, which is the goal of this work. We proposed the utilization of a coupled 3D Convolutional Neural Network (CNN) architecture that can map both modalities into a representation space to evaluate the correspondence of audio-visual streams using the learned multimodal features.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 7
    Lita

    Lita

    A robot companion for your company's chat room

    Lita is a chat bot written in Ruby that brings more fun and efficiency to your favorite chat service. Through its plugin system, Lita can be connected to different chat services and display new behavior preferred by those who use it. It's ideal for businesses that want a chat service that is not only efficient, but friendly and personalized as well. Lita can become your very own robot companion, tailor-made for your business. Lita can be customized according to your company's culture and needs. It can be used to automate various time-consuming and error-prone tasks, while also letting your company members have fun and create a sense of community.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 8
    MACE

    MACE

    Deep learning inference framework optimized for mobile platforms

    Mobile AI Compute Engine (or MACE for short) is a deep learning inference framework optimized for mobile heterogeneous computing on Android, iOS, Linux and Windows devices. Runtime is optimized with NEON, OpenCL and Hexagon, and Winograd algorithm is introduced to speed up convolution operations. The initialization is also optimized to be faster. Chip-dependent power options like big.LITTLE scheduling, Adreno GPU hints are included as advanced APIs. UI responsiveness guarantee is sometimes obligatory when running a model. Mechanism like automatically breaking OpenCL kernel into small units is introduced to allow better preemption for the UI rendering task. Graph level memory allocation optimization and buffer reuse are supported. The core library tries to keep minimum external dependencies to keep the library footprint small.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 9
    ML Visuals

    ML Visuals

    ML Visuals contains figures and templates which you can reuse

    ML Visuals is an open-source project that provides a collection of reusable diagrams, templates, and visual resources designed to improve communication in machine learning research and education. The repository contains professional-quality figures that illustrate machine learning concepts such as neural networks, optimization methods, model architectures, and common deep learning techniques. These visuals are intended to help researchers, educators, and students create clearer presentations, blog posts, and scientific papers. The project is maintained as a collaborative community effort where contributors can add new diagrams or visual components. Many of the visuals are designed using editable formats such as Google Slides, making it easy for users to customize them for their own work.
    Downloads: 1 This Week
    Last Update:
    See Project
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    Access competitive interest rates on your digital assets.

    Generate interest, borrow against your crypto, and trade a range of cryptocurrencies — all in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 10
    MLOps Zoomcamp

    MLOps Zoomcamp

    Free MLOps course from DataTalks.Club

    MLOps Zoomcamp is an open-source educational repository that contains the materials for a free course focused on machine learning operations and production machine learning systems. The course is designed to teach data scientists and engineers how to move machine learning models from experimentation environments into scalable production services. The repository provides lessons, code examples, and assignments that cover the entire MLOps lifecycle, including model training, experiment tracking, deployment, monitoring, and infrastructure management. Students learn to use widely adopted tools such as MLflow, orchestration frameworks, and cloud platforms to manage machine learning pipelines. The curriculum emphasizes hands-on projects so learners gain practical experience building automated ML pipelines and maintaining deployed models.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    MLPerf

    MLPerf

    Reference implementations of MLPerf™ training benchmarks

    This is a repository of reference implementations for the MLPerf training benchmarks. These implementations are valid as starting points for benchmark implementations but are not fully optimized and are not intended to be used for "real" performance measurements of software frameworks or hardware. Benchmarking the performance of training ML models on a wide variety of use cases, software, and hardware drives AI performance across the tech industry. The MLPerf Training working group draws on expertise in AI and the technology that powers AI from across the industry to design and create industry-standard benchmarks. Together, we create the reference implementations, rules, policies, and procedures to benchmark a wide variety of AI workloads.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 12
    MLX Engine

    MLX Engine

    LM Studio Apple MLX engine

    MLX Engine is the Apple MLX-based inference backend used by LM Studio to run large language models efficiently on Apple Silicon hardware. Built on top of the mlx-lm and mlx-vlm ecosystems, the engine provides a unified architecture capable of supporting both text-only and multimodal models. Its design focuses on high-performance on-device inference, leveraging Apple’s MLX stack to accelerate computation on M-series chips. The project introduces modular VisionAddOn components that allow image embeddings to be integrated seamlessly into language model workflows. It is bundled with newer versions of LM Studio but can also be used independently for experimentation and development. Overall, mlx-engine serves as a specialized high-efficiency runtime for local AI workloads on macOS systems.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 13
    Machine Learning Financial Laboratory

    Machine Learning Financial Laboratory

    MlFinLab helps portfolio managers and traders

    MlFinLab is a comprehensive Python library designed to support the development of machine learning strategies in quantitative finance and algorithmic trading. The project provides a large collection of tools that implement techniques from academic research on financial machine learning. It covers the full lifecycle of developing data-driven trading strategies, including data preprocessing, feature engineering, labeling techniques, model training, and performance evaluation. Many of the algorithms implemented in the library are based on concepts introduced in advanced quantitative finance literature and peer-reviewed research. The library also includes tools for constructing specialized financial data structures, generating predictive features, and evaluating trading strategies through backtesting. Its architecture emphasizes reproducibility, robust testing, and well-documented code so that researchers and practitioners can reliably experiment with financial machine learning models.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 14
    Machine Learning Foundations

    Machine Learning Foundations

    Machine Learning Foundations: Linear Algebra, Calculus, Statistics

    Machine Learning Foundations repository contains the code, notebooks, and teaching materials used in Jon Krohn’s Machine Learning Foundations curriculum. The project focuses on explaining the fundamental mathematical and computational concepts that underpin modern machine learning and artificial intelligence systems. The materials cover essential topics such as linear algebra, calculus, statistics, and probability, which form the theoretical basis of many machine learning algorithms. The repository includes Jupyter notebooks with explanations and examples that demonstrate how these mathematical principles relate to real machine learning applications. Each section introduces theoretical concepts and then illustrates them through practical coding examples to reinforce understanding. The project is designed for students and practitioners who want to strengthen their foundational knowledge before working with more advanced machine learning frameworks.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 15
    Machine Learning PyTorch Scikit-Learn

    Machine Learning PyTorch Scikit-Learn

    Code Repository for Machine Learning with PyTorch and Scikit-Learn

    Initially, this project started as the 4th edition of Python Machine Learning. However, after putting so much passion and hard work into the changes and new topics, we thought it deserved a new title. So, what’s new? There are many contents and additions, including the switch from TensorFlow to PyTorch, new chapters on graph neural networks and transformers, a new section on gradient boosting, and many more that I will detail in a separate blog post. For those who are interested in knowing what this book covers in general, I’d describe it as a comprehensive resource on the fundamental concepts of machine learning and deep learning. The first half of the book introduces readers to machine learning using scikit-learn, the defacto approach for working with tabular datasets. Then, the second half of this book focuses on deep learning, including applications to natural language processing and computer vision.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 16
    Machine learning basics

    Machine learning basics

    Plain python implementations of basic machine learning algorithms

    Machine learning basics repository is an educational project that provides plain Python implementations of fundamental machine learning algorithms designed to help learners understand how these methods work internally. Instead of relying on external machine learning libraries, the algorithms are implemented from scratch so that users can explore the mathematical logic and computational structure behind each technique. The repository includes notebooks that demonstrate classic algorithms such as linear regression, logistic regression, k-nearest neighbors, decision trees, support vector machines, and clustering techniques. Each notebook typically combines explanatory text, Python code, and visualizations to illustrate how the algorithm operates and how it can be applied to datasets.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 17
    Machine-Learning-Notes

    Machine-Learning-Notes

    Zhou Zhihua's "Machine Learning" push notes

    The Machine-Learning-Notes repository contains detailed handwritten-style study notes based on the popular machine learning textbook by Zhou Zhihua. The project focuses on deriving formulas and explaining algorithms step by step so that learners can understand the mathematical foundations behind machine learning methods. The notes span sixteen chapters that cover a wide range of topics, including model evaluation, linear models, decision trees, neural networks, support vector machines, Bayesian classifiers, ensemble methods, clustering, dimensionality reduction, and reinforcement learning. Each section explains the theoretical principles of the algorithms and walks through derivations to help readers understand why the methods work rather than simply how to use them. The repository organizes the material into printable chapters so that students can study the notes offline or use them as reference material while learning machine learning theory.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 18
    Made With ML

    Made With ML

    Learn how to develop, deploy and iterate on production-grade ML

    Made-With-ML is an open-source educational repository and course designed to teach developers how to build production-grade machine learning systems using modern MLOps practices. The project focuses on bridging the gap between experimental machine learning notebooks and real-world software systems that can be deployed, monitored, and maintained at scale. It provides structured lessons and practical code examples that demonstrate how to design machine learning workflows, manage datasets, train models, evaluate performance, and deploy inference services. The repository organizes these concepts into modular Python scripts that follow software engineering best practices such as testing, configuration management, logging, and version control. Through a combination of tutorials, notebooks, and production-ready scripts, the project demonstrates how machine learning applications should be developed as maintainable systems rather than isolated experiments.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 19
    NN-SVG

    NN-SVG

    Publication-ready NN-architecture schematics

    Illustrations of Neural Network architectures are often time-consuming to produce, and machine learning researchers all too often find themselves constructing these diagrams from scratch by hand. NN-SVG is a tool for creating Neural Network (NN) architecture drawings parametrically rather than manually. It also provides the ability to export those drawings to Scalable Vector Graphics (SVG) files, suitable for inclusion in academic papers or web pages. The tool provides the ability to generate figures of three kinds: classic Fully-Connected Neural Network (FCNN) figures, Convolutional Neural Network (CNN) figures of the sort introduced in the LeNet paper, and Deep Neural Network figures following the style introduced in the AlexNet paper. The former two are accomplished using the D3 javascript library and the latter with the javascript library Three.js. NN-SVG provides the ability to style the figure to the user's liking via many size, color, and layout parameters.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 20
    OpenPrompt

    OpenPrompt

    An Open-Source Framework for Prompt-Learning

    Prompt-learning is the latest paradigm to adapt pre-trained language models (PLMs) to downstream NLP tasks, which modifies the input text with a textual template and directly uses PLMs to conduct pre-trained tasks. OpenPrompt is a library built upon PyTorch and provides a standard, flexible and extensible framework to deploy the prompt-learning pipeline. OpenPrompt supports loading PLMs directly from huggingface transformers. In the future, we will also support PLMs implemented by other libraries. The template is one of the most important modules in prompt learning, which wraps the original input with textual or soft-encoding sequence. Use the implementations of current prompt-learning approaches.* We have implemented various of prompting methods, including templating, verbalizing and optimization strategies under a unified standard. You can easily call and understand these methods.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 21
    PandaOCR

    PandaOCR

    Multifunctional OCR Image and Text Recognition

    At present, the newly refactored PandaOCR.Pro professional version has been released. It is faster and more stable, with richer interfaces and easier operation. It is recommended for you to use it! The normal version will continue to be maintained, and all interfaces will be retained but no new functions will be added. The reason why the version number of the professional version starts from 5.x is that the normal version will be updated in the future, so a period of version number is reserved. You can continue to use the regular version for free as before, without worrying about deactivating the regular version after the launch of the professional version. If you have higher needs, you can try the professional version. You can also use the Baidu API interface without activation. Support shortcut keys and screen corner trigger screenshot recognition function, convenient and fast.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 22
    PennyLane

    PennyLane

    A cross-platform Python library for differentiable programming

    A cross-platform Python library for differentiable programming of quantum computers. Train a quantum computer the same way as a neural network. Built-in automatic differentiation of quantum circuits, using the near-term quantum devices directly. You can combine multiple quantum devices with classical processing arbitrarily! Support for hybrid quantum and classical models, and compatible with existing machine learning libraries. Quantum circuits can be set up to interface with either NumPy, PyTorch, JAX, or TensorFlow, allowing hybrid CPU-GPU-QPU computations. The same quantum circuit model can be run on different devices. Install plugins to run your computational circuits on more devices, including Strawberry Fields, Amazon Braket, Qiskit and IBM Q, Google Cirq, Rigetti Forest, and the Microsoft QDK.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 23
    Pipcook

    Pipcook

    Machine learning platform for Web developers

    A JavaScript application framework for machine learning and its engineering. With the mission of enabling JavaScript engineers to utilize the power of machine learning without any prerequisites and the vision to lead the front-end technical field to intelligence. Pipcook is to become the JavaScript application framework for the cross-cutting area of machine learning and front-end interaction. We are truly to design Pipcook's API for front-end and machine learning applications, and focusing on the front-end area and developed from the JavaScript engineers' view. With the principle of being friendly to JavaScript, we will push the whole area forward with the machine learning engineering. For this reason we opened an issue about machine-learning application APIs, and look forward to you get involved.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 24
    PixelAnnotationTool

    PixelAnnotationTool

    Annotate quickly images

    Software that allows you to manually and quickly annotate images in directories. The method is pseudo manual because it uses the algorithm watershed marked of OpenCV. The general idea is to manually provide the marker with brushes and then to launch the algorithm. If at first pass the segmentation needs to be corrected, the user can refine the markers by drawing new ones on the erroneous areas.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 25
    PyG

    PyG

    Graph Neural Network Library for PyTorch

    PyG (PyTorch Geometric) is a library built upon PyTorch to easily write and train Graph Neural Networks (GNNs) for a wide range of applications related to structured data. It consists of various methods for deep learning on graphs and other irregular structures, also known as geometric deep learning, from a variety of published papers. In addition, it consists of easy-to-use mini-batch loaders for operating on many small and single giant graphs, multi GPU-support, DataPipe support, distributed graph learning via Quiver, a large number of common benchmark datasets (based on simple interfaces to create your own), the GraphGym experiment manager, and helpful transforms, both for learning on arbitrary graphs as well as on 3D meshes or point clouds. All it takes is 10-20 lines of code to get started with training a GNN model (see the next section for a quick tour).
    Downloads: 1 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB