Open Source Machine Learning Software - Page 36

Machine Learning Software

View 446 business solutions
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • AI-powered service management for IT and enterprise teams Icon
    AI-powered service management for IT and enterprise teams

    Enterprise-grade ITSM, for every business

    Give your IT, operations, and business teams the ability to deliver exceptional services—without the complexity. Maximize operational efficiency with refreshingly simple, AI-powered Freshservice.
    Try it Free
  • 1
    Keras.js

    Keras.js

    Run Keras models in the browser, with GPU support using WebGL

    Run Keras models in the browser, with GPU support provided by WebGL 2. Models can be run in Node.js as well, but only in CPU mode. Because Keras abstracts away a number of frameworks as backends, the models can be trained in any backend, including TensorFlow, CNTK, etc. Check out the demos/ directory for real examples running Keras.js in VueJS. Library version compatibility, Keras 2.1.2.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    KerasTuner

    KerasTuner

    A Hyperparameter Tuning Library for Keras

    KerasTuner is an easy-to-use, scalable hyperparameter optimization framework that solves the pain points of hyperparameter search. Easily configure your search space with a define-by-run syntax, then leverage one of the available search algorithms to find the best hyperparameter values for your models. KerasTuner comes with Bayesian Optimization, Hyperband, and Random Search algorithms built-in, and is also designed to be easy for researchers to extend in order to experiment with new search algorithms.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Kernel Adaptive Filtering Toolbox

    Kernel Adaptive Filtering Toolbox

    a Matlab benchmarking toolbox for kernel adaptive filtering

    [Note: This project has moved. Visit https://github.com/steven2358/kafbox/ for the latest version.] A Matlab benchmarking toolbox for kernel adaptive filtering. Kernel adaptive filtering algorithms are online and adaptive regression algorithms based on kernels. They are suitable for nonlinear filtering, prediction, tracking and nonlinear regression in general. This toolbox includes algorithms, demos, and tools to compare their performance. See the included README file for a list of included algorithms and more details. If you use this toolbox in your research please cite: @inproceedings{vanvaerenbergh2013comparative, author = {Van Vaerenbergh, Steven and Santamar{\'i}a, Ignacio}, booktitle = {2013 IEEE Digital Signal Processing (DSP) Workshop and IEEE Signal Processing Education (SPE)}, title = {A Comparative Study of Kernel Adaptive Filtering Algorithms}, year = {2013}, note = {Software available at \url{https://github.com/steven2358/kafbox/}} }
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Key-book

    Key-book

    Proofs, cases, concept supplements, and reference explanations

    The book "Introduction to Machine Learning Theory" (hereinafter referred to as "Introduction") written by Zhou Zhihua, Wang Wei, Gao Wei, and other teachers fills the regret of the lack of introductory works on machine learning theory in China. This book attempts to provide an introductory guide for readers interested in learning machine learning theory and researching machine learning theory in an easy-to-understand language. "Guide" mainly covers seven parts, corresponding to seven important concepts or theoretical tools in machine learning theory, namely: learnability, (hypothesis space) complexity, generalization bound, stability, consistency, convergence rate, regret circle. Daoyin is a highly theoretical book, involving a large number of mathematical theorems and various proofs. Although the writing team has reduced the difficulty as much as possible, due to the nature of machine learning theory, the book still places high demands on the reader's mathematical background.
    Downloads: 0 This Week
    Last Update:
    See Project
  • AI-generated apps that pass security review Icon
    AI-generated apps that pass security review

    Stop waiting on engineering. Build production-ready internal tools with AI—on your company data, in your cloud.

    Retool lets you generate dashboards, admin panels, and workflows directly on your data. Type something like “Build me a revenue dashboard on my Stripe data” and get a working app with security, permissions, and compliance built in from day one. Whether on our cloud or self-hosted, create the internal software your team needs without compromising enterprise standards or control.
    Try Retool free
  • 5
    Spider that recollects data from MySpace Social Network. At now, it is only designed to extract information from native american people because it is used for a social science study in the UNAM (Universidad Nacional Autónoma de México).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    Knet

    Knet

    Koç University deep learning framework

    Knet.jl is a deep learning package implemented in Julia, so you should be able to run it on any machine that can run Julia. It has been extensively tested on Linux machines with NVIDIA GPUs and CUDA libraries, and it has been reported to work on OSX and Windows. If you would like to try it on your own computer, please follow the instructions on Installation. If you would like to try working with a GPU and do not have access to one, take a look at Using Amazon AWS or Using Microsoft Azure. If you find a bug, please open a GitHub issue. If you don't have access to a GPU machine, but would like to experiment with one, Amazon Web Services is a possible solution. I have prepared a machine image (AMI) with everything you need to run Knet. Here are step-by-step instructions for launching a GPU instance with a Knet image (the screens may have changed slightly since this writing).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Knock Knock

    Knock Knock

    Get notified when your training ends

    Knock Knock is a lightweight Python utility created by the Hugging Face team that allows developers to receive notifications when long-running machine learning tasks finish or fail. Training deep learning models often takes hours or even days, making it inconvenient for engineers to constantly monitor progress manually. The library solves this problem by adding simple decorators or command-line commands that automatically send notifications when a process completes or crashes. These alerts can be delivered through several communication platforms such as email, Slack, Telegram, or other messaging services. The goal of the project is to allow developers to monitor experiments remotely without needing to stay connected to the training environment. By adding only a few lines of code, the library can wrap around a training function and report execution status.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    Kodezi Chronos

    Kodezi Chronos

    Kodezi Chronos is a debugging-first language model

    Kodezi Chronos is a research project focused on developing a specialized language model designed specifically for debugging software and understanding large code repositories. Unlike general-purpose language models that focus primarily on code generation, Chronos is built to diagnose and repair bugs by analyzing complex relationships across files within a codebase. The project introduces architectural techniques such as Adaptive Graph-Guided Retrieval, which allows the system to navigate large repositories and retrieve relevant debugging information from multiple sources. Another component, Persistent Debug Memory, allows the system to learn patterns from past debugging sessions and apply that knowledge to future problems. The repository mainly contains research documentation, evaluation benchmarks, and experimental frameworks rather than the full proprietary model implementation.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    Koila

    Koila

    Prevent PyTorch's `CUDA error: out of memory` in just 1 line of code

    Koila is a lightweight Python library designed to help developers avoid memory errors when training deep learning models with PyTorch. The library introduces a lazy evaluation mechanism that delays computation until it is actually required, allowing the framework to better estimate the memory requirements of a model before execution. By building a computational graph first and executing operations only when necessary, koila reduces the risk of running out of GPU memory during the forward pass of neural network training. This approach enables developers to experiment with larger batch sizes and more complex architectures while maintaining stable training behavior. The system acts as a thin wrapper around PyTorch tensors and operations, meaning that it integrates easily into existing PyTorch code without requiring major changes to model implementations. It is particularly useful in environments where GPU resources are limited or where models frequently encounter CUDA memory errors.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • 10
    Kubeflow pipelines

    Kubeflow pipelines

    Machine Learning Pipelines for Kubeflow

    Kubeflow is a machine learning (ML) toolkit that is dedicated to making deployments of ML workflows on Kubernetes simple, portable, and scalable. A pipeline is a description of an ML workflow, including all of the components in the workflow and how they combine in the form of a graph. The pipeline includes the definition of the inputs (parameters) required to run the pipeline and the inputs and outputs of each component. A pipeline component is a self-contained set of user code, packaged as a Docker image, that performs one step in the pipeline. For example, a component can be responsible for data preprocessing, data transformation, model training, and so on.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11

    LAML:Linear Algebra and Machine Learning

    A stand-alone Java library for linear algebra and machine learning

    LAML is a stand-alone pure Java library for linear algebra and machine learning. The goal is to build efficient and easy-to-use linear algebra and machine learning libraries. The reason why linear algebra and machine learning are built together is that full control of the basic data structures for matrices and vectors is required to have fast implementation for machine learning methods. Additionally, LAML provides a lot of commonly used matrix functions in the same signature to MATLAB, thus can also be used to manually convert MATLAB code to Java code.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12

    LBP in multiple platforms

    LBP implementation in multiple computing platforms (ARM,GPU, DSP...)

    The Local Binary Pattern (LBP) is a texture operator that is used in several different computer vision applications and implemented in a variety of platforms. When selecting a suitable LBP implementation platform, the specific application and its requirements in terms of performance, size, energy efficiency, cost and developing time has to be carefully considered. This is a software toolbox that collects software implementations of the Local Binary Pattern operator in several platforms: - OpenCL for CPU & GPU - OpenCL for GPU (branchless) - C code optimized for ARM - OpenGL ES 2.0 shaders mobile GPUs - C code for TI C64x DSP core (branchless) - C code for TTA processor synthesis If you use the code somewhere, please cite: Bordallo López M., Nieto A., Boutellier J., Hannuksela J., and Silvén O. "Evaluation of real-time LBP computing in multiple architectures," Journal of Real Time Image Processing, 2014
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13

    LCS Open Repository

    LCS open repository in C++

    I am glad to present the Learning Classifier System Open Repository (LCSOR); a repository on LCSs implemented in C++ to foster practitioners the use of these machine learning techniques and to further extend LCSs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    LLM Applications

    LLM Applications

    A comprehensive guide to building RAG-based LLM applications

    LLM Applications is a practical reference repository that demonstrates how to build production-grade applications powered by large language models. The project focuses particularly on Retrieval-Augmented Generation architectures, which combine language models with external knowledge sources to improve accuracy and reliability. It provides step-by-step guidance for constructing systems that ingest documents, split them into chunks, generate embeddings, index them in vector databases, and retrieve relevant context during inference. The repository also shows how these components can be scaled and deployed using distributed computing frameworks such as Ray. In addition to development workflows, the project includes notebooks, datasets, and evaluation tools that help developers experiment with different retrieval strategies and model configurations.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    LPCforSOS is a machine learning framework with a special focus on structured output spaces and pairwise learning. It supports currently multiclass, ordinal, hierarchical, multi-label and label ranking classification settings.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    LRSLibrary

    LRSLibrary

    Low-Rank and Sparse Tools for Background Modeling and Subtraction

    LRSLibrary is a MATLAB library offering a broad collection of low-rank plus sparse decomposition algorithms, primarily aimed at background/foreground modeling from videos (background subtraction) and related computer vision tasks. Compatibility across MATLAB versions (tested in R2014–R2017) The library includes matrix and tensor methods (over 100 algorithms) and has been tested across MATLAB versions from R2014 onward. The algorithms can also be adapted to other computer vision or machine learning problems beyond video. Large algorithm collection: > 100 matrix- and tensor-based low-rank + sparse methods. Open-source license, documentation and references included.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    LSTMs for Human Activity Recognition

    LSTMs for Human Activity Recognition

    Human Activity Recognition example using TensorFlow on smartphone

    LSTM-Human-Activity-Recognition is a machine learning project that demonstrates how recurrent neural networks can be used to recognize human activities from sensor data. The repository implements a deep learning model based on Long Short-Term Memory (LSTM) networks to classify physical activities using time-series data collected from wearable sensors. The project uses the well-known Human Activity Recognition dataset derived from smartphone accelerometer and gyroscope signals. Through the use of sequential neural network architectures, the system learns patterns in motion data that correspond to activities such as walking, sitting, standing, or climbing stairs. The repository includes data preprocessing scripts, neural network architecture definitions, and training pipelines that allow researchers to reproduce and modify the experiments. It serves as an educational example of how deep learning models can process temporal sensor signals for pattern recognition tasks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    Linear Time Invariant (LTI) system identification using particle swarm optimization (PSO) algorithm. Creators : Vahid Kiani, Hadi Sadoghi Yazdi
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    LUMINOTH

    LUMINOTH

    Deep Learning toolkit for Computer Vision

    LUMINOTH is an open-source deep learning toolkit designed for computer vision tasks, particularly object detection. The framework is implemented in Python and built on top of TensorFlow and the Sonnet neural network library, providing a modular environment for training and deploying detection models. It was created to simplify the process of building and experimenting with deep learning models capable of identifying objects within images. Luminoth includes support for popular object detection architectures such as Faster R-CNN and SSD, enabling developers to train models on datasets like COCO and Pascal VOC. The toolkit provides command-line utilities for dataset management, training, and inference, making it easier to integrate into research workflows and production systems. Although the project is no longer actively maintained, it remains a useful educational and experimental platform for studying object detection pipelines and deep learning workflows.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20

    LWPR

    Locally Weighted Projection Regression (LWPR)

    Locally Weighted Projection Regression (LWPR) is a fully incremental, online algorithm for non-linear function approximation in high dimensional spaces, capable of handling redundant and irrelevant input dimensions. At its core, it uses locally linear models, spanned by a small number of univariate regressions in selected directions in input space. A locally weighted variant of Partial Least Squares (PLS) is employed for doing the dimensionality reduction. Please cite: [1] Sethu Vijayakumar, Aaron D'Souza and Stefan Schaal, Incremental Online Learning in High Dimensions, Neural Computation, vol. 17, no. 12, pp. 2602-2634 (2005). [2] Stefan Klanke, Sethu Vijayakumar and Stefan Schaal, A Library for Locally Weighted Projection Regression, Journal of Machine Learning Research (JMLR), vol. 9, pp. 623--626 (2008). More details and usage guidelines on the code website.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    Lambda Networks

    Lambda Networks

    Implementation of LambdaNetworks, a new approach to image recognition

    Implementation of λ Networks, a new approach to image recognition that reaches SOTA on ImageNet. The new method utilizes λ layer, which captures interactions by transforming contexts into linear functions, termed lambdas, and applying these linear functions to each input separately. Shinel94 has added a Keras implementation! It won't be officially supported in this repository, so either copy / paste the code under ./lambda_networks/tfkeras.py or make sure to install tensorflow and keras before running the provided commands.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    Leark is a Data Mining library developed in C#.NET. It contains several methods for ranking web documents described with a set of normalized features, and a feature selection algorithm. The methods are based on perceptron and clustering.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    Learn PyTorch for Deep Learning

    Learn PyTorch for Deep Learning

    Materials for the Learn PyTorch for Deep Learning

    Learn PyTorch for Deep Learning is an open-source educational repository that provides the full learning materials for the “Learn PyTorch for Deep Learning: Zero to Mastery” course created by Daniel Bourke. The project is designed to teach beginners how to build deep learning models using PyTorch through a hands-on, code-first learning approach. Instead of focusing heavily on theory alone, the repository encourages learners to experiment with code and develop practical machine learning skills through guided examples and exercises. The materials include Jupyter notebooks, explanations of core deep learning concepts, and step-by-step demonstrations of building and training neural networks. Throughout the lessons, users learn how to work with tensors, create neural network architectures, manage training workflows, and evaluate model performance.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    Learn_Data_Science_in_3_Months

    Learn_Data_Science_in_3_Months

    This is the Curriculum for "Learn Data Science in 3 Months"

    This project lays out a 12-week plan to go from basics to a portfolio-ready understanding of data science. It breaks the journey into clear stages: Python fundamentals, data wrangling, visualization, statistics, machine learning, and end-to-end projects. The schedule mixes learning and doing, encouraging you to build small deliverables each week—like notebooks, dashboards, and model demos—to reinforce skills. It also includes suggestions for datasets and problem domains so you aren’t stuck wondering what to analyze next. The plan is intentionally opinionated but flexible: you can swap resources while keeping the weekly objectives intact. By the end, you’re expected to have tangible artifacts to show employers or collaborators, not just notes and bookmarks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    LensOSX

    LensOSX

    LensOSX: the light, efficient network simulator

    Lens is the light, efficient network simulator, written by Doug Rohde. LensOSX is a native MacOSX port of Lens that runs on MacOSX 10.5 or higher, created by Harm Brouwer, Daniel de Kok and Hartmut Fitz.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB