Open Source Linux Machine Learning Software - Page 28

Machine Learning Software for Linux

View 58 business solutions
  • Go from Code to Production URL in Seconds Icon
    Go from Code to Production URL in Seconds

    Cloud Run deploys apps in any language instantly. Scales to zero. Pay only when code runs.

    Skip the Kubernetes configs. Cloud Run handles HTTPS, scaling, and infrastructure automatically. Two million requests free per month.
    Try it free
  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • 1

    JaCHMM

    Java Conditioned Hidden Markov Model library

    The JaCHMM - the Java Conditioned Hidden Markov Model library - is a complete implementation of a CHMM in Java ready to use either on command line or as a module. The JaCHMM is licenced under the BSD licence. It gives an implementation of the Viterbi, Forward-Backward, Baum-Welch and K-Means algorithms, all adapted for the CHMM. JaCHMM is based on the JaHMM and also designed to achieve reasonable performance without making the code unreadable. Consequently, it offers a good way of applying the Conditioned Hidden Markov Model in various tasks, e.g., for scientific or teaching purposes.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2

    Java Library for Machine Learning

    A pure Java library for machine learning

    JML is a machine learning library in Java, it is a pure Java package, and thus is cross-platform. The goal of JML is to make machine learning methods very easy to use and speed up code conversion from MATLAB to Java. Please be noted that JML has been replaced by LAML.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Program to performing the complete cycle of neural networks analysis: preparing data, choosing neural network (CasCor, MP, LogRegression, PNN), learning of network, monitoring learning state, ROC-analysis, optimization of network parameters using GA.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Jina

    Jina

    Build cross-modal and multimodal applications on the cloud

    Jina is a framework that empowers anyone to build cross-modal and multi-modal applications on the cloud. It uplifts a PoC into a production-ready service. Jina handles the infrastructure complexity, making advanced solution engineering and cloud-native technologies accessible to every developer. Build applications that deliver fresh insights from multiple data types such as text, image, audio, video, 3D mesh, PDF with Jina AI’s DocArray. Polyglot gateway that supports gRPC, Websockets, HTTP, GraphQL protocols with TLS. Intuitive design pattern for high-performance microservices. Seamless Docker container integration: sharing, exploring, sandboxing, versioning and dependency control via Jina Hub. Fast deployment to Kubernetes, Docker Compose and Jina Cloud. Improved engineering efficiency thanks to the Jina AI ecosystem, so you can focus on innovating with the data applications you build.
    Downloads: 0 This Week
    Last Update:
    See Project
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 5
    Jina-Serve

    Jina-Serve

    Build multimodal AI applications with cloud-native stack

    Jina Serve is an open-source framework designed for building, deploying, and scaling AI services and machine learning pipelines in production environments. The framework allows developers to create microservices that expose machine learning models through APIs that communicate using protocols such as HTTP, gRPC, and WebSockets. It is built with a cloud-native architecture that supports deployment on local machines, containerized environments, or large orchestration platforms such as Kubernetes. Jina Serve focuses on making it easier to turn machine learning models into production-ready services without forcing developers to manage complex infrastructure manually. The framework supports many major machine learning libraries and data types, making it suitable for multimodal AI systems that process text, images, audio, and other inputs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    Jittor

    Jittor

    Jittor is a high-performance deep learning framework

    Jittor is a high-performance deep learning framework based on JIT compiling and meta-operators. The whole framework and meta-operators are compiled just in time. A powerful op compiler and tuner are integrated into Jittor. It allowed us to generate high-performance code specialized for your model. Jittor also contains a wealth of high-performance model libraries, including image recognition, detection, segmentation, generation, differentiable rendering, geometric learning, reinforcement learning, etc. The front-end language is Python. Module Design and Dynamic Graph Execution is used in the front-end, which is the most popular design for deep learning framework interface. The back-end is implemented by high-performance languages, such as CUDA, C++. Jittor'op is similar to NumPy. Let's try some operations. We create Var a and b via operation jt.float32, and add them. Printing those variables shows they have the same shape and dtype.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7

    KMeansAniX

    Animation of kmeans clustering using X Window System

    Open source animation of kmeans clustering in X Window System using the C++ libplotter library. Supports Linux, Mac, and BSD. Includes common initialization methods such as Forgy, Macqueen, random, and angular. Sample videos are available through the Files Tab above. The SVN repo is accessible thorugh the Code Tab above. Requires a C++ compiler, libplot-dev, and libncurses5-dev Mac alternative to libplot-dev: macports plotutils +x11
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    The KReator project is a collection of software systems, tools, algorithms and data structures for logic-based knowledge representation. Currently, it includes the software systems KReator and MECore and the library Log4KR: - KReator is an integrated development environment (IDE) for relational probabilistic knowledge representation languages such as Bayesian Logic Programs (BLPs), Markov Logic Networks (MLNs), Relational Maximum Entropy (RME), First-Order Probabilistic Conditional Logic (FO-PCL), and others. - MECore is a shell-based system that allows the user to create propositional knowledge bases, to perform a variety of belief change operations, and to query a knowledge base with respect to the principle of optimum entropy. - Log4KR is a library providing data structures to represent knowledge bases in various logic formalisms (propositional, relational, conditional, probabilistic, ...) and providing algorithms to perform reasoning operations
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    Kaggle Solutions

    Kaggle Solutions

    Collection of Kaggle Solutions and Ideas

    Kaggle Solutions is an open-source repository that compiles winning solutions, insights, and educational resources from hundreds of Kaggle data science competitions. The repository acts as a knowledge base for competitive machine learning by collecting solution write-ups, discussion threads, code notebooks, and tutorial resources shared by top Kaggle participants. Each competition entry typically includes information about the dataset, evaluation metrics, modeling strategies, and techniques used by high-ranking competitors. The repository also highlights important machine learning concepts such as feature engineering, cross-validation strategies, ensemble modeling, and post-processing methods commonly used in winning solutions. Because the content is organized by competition categories such as computer vision, natural language processing, tabular data, and time-series forecasting, users can explore techniques relevant to specific problem types.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 10
    Karate Club

    Karate Club

    An API Oriented Open-source Python Framework for Unsupervised Learning

    Karate Club is an unsupervised machine learning extension library for NetworkX. Karate Club consists of state-of-the-art methods to do unsupervised learning on graph-structured data. To put it simply it is a Swiss Army knife for small-scale graph mining research. First, it provides network embedding techniques at the node and graph level. Second, it includes a variety of overlapping and non-overlapping community detection methods. Implemented methods cover a wide range of network science (NetSci, Complenet), data mining (ICDM, CIKM, KDD), artificial intelligence (AAAI, IJCAI) and machine learning (NeurIPS, ICML, ICLR) conferences, workshops, and pieces from prominent journals.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Karpathy

    Karpathy

    An agentic Machine Learning Engineer

    karpathy is an experimental agentic machine learning engineer framework designed to automate many aspects of the ML development workflow. The project sets up a sandboxed environment where an AI agent can access datasets, run experiments, and generate machine learning artifacts through a web interface. Its startup script automatically prepares the environment by creating a sandbox directory, installing key ML libraries, and launching the agent interface. The system is tightly integrated with the Claude Scientific Skills ecosystem, enabling the agent to leverage specialized scientific and machine learning tools. It is intended primarily for research and experimentation with autonomous ML workflows rather than as a polished production platform. Overall, karpathy represents an early step toward fully automated machine learning engineering driven by agentic AI systems.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Kashgari

    Kashgari

    Kashgari is a production-level NLP Transfer learning framework

    Kashgari is a simple and powerful NLP Transfer learning framework, build a state-of-art model in 5 minutes for named entity recognition (NER), part-of-speech tagging (PoS), and text classification tasks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Katib

    Katib

    Automated Machine Learning on Kubernetes

    Katib is a Kubernetes-native project for automated machine learning (AutoML). Katib supports Hyperparameter Tuning, Early Stopping and Neural Architecture Search. Katib is a project that is agnostic to machine learning (ML) frameworks. It can tune hyperparameters of applications written in any language of the users’ choice and natively supports many ML frameworks, such as TensorFlow, Apache MXNet, PyTorch, XGBoost, and others. Katib can perform training jobs using any Kubernetes Custom Resources with out-of-the-box support for Kubeflow Training Operator, Argo Workflows, Tekton Pipelines, and many more.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Keepsake

    Keepsake

    Version control for machine learning

    Keepsake is a Python library that uploads files and metadata (like hyperparameters) to Amazon S3 or Google Cloud Storage. You can get the data back out using the command-line interface or a notebook.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    KeplerWeka adds the functionality of the open-source machine learning and data mining workbench WEKA to the free and open-source, scientific workflow application, Kepler.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Keras Attention Mechanism

    Keras Attention Mechanism

    Attention mechanism Implementation for Keras

    Many-to-one attention mechanism for Keras. We demonstrate that using attention yields a higher accuracy on the IMDB dataset. We consider two LSTM networks: one with this attention layer and the other one with a fully connected layer. Both have the same number of parameters for a fair comparison (250K). The attention is expected to be the highest after the delimiters. An overview of the training is shown below, where the top represents the attention map and the bottom the ground truth. As the training progresses, the model learns the task and the attention map converges to the ground truth. We consider many 1D sequences of the same length. The task is to find the maximum of each sequence. We give the full sequence processed by the RNN layer to the attention layer. We expect the attention layer to focus on the maximum of each sequence.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Keras TCN

    Keras TCN

    Keras Temporal Convolutional Network

    TCNs exhibit longer memory than recurrent architectures with the same capacity. Performs better than LSTM/GRU on a vast range of tasks (Seq. MNIST, Adding Problem, Copy Memory, Word-level PTB...). Parallelism (convolutional layers), flexible receptive field size (possible to specify how far the model can see), stable gradients (backpropagation through time, vanishing gradients). The usual way is to import the TCN layer and use it inside a Keras model. The receptive field is defined as the maximum number of steps back in time from current sample at time T, that a filter from (block, layer, stack, TCN) can hit (effective history) + 1. The receptive field of the TCN can be calculated. Once keras-tcn is installed as a package, you can take a glimpse of what is possible to do with TCNs. Some tasks examples are available in the repository for this purpose.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    Keras.js

    Keras.js

    Run Keras models in the browser, with GPU support using WebGL

    Run Keras models in the browser, with GPU support provided by WebGL 2. Models can be run in Node.js as well, but only in CPU mode. Because Keras abstracts away a number of frameworks as backends, the models can be trained in any backend, including TensorFlow, CNTK, etc. Check out the demos/ directory for real examples running Keras.js in VueJS. Library version compatibility, Keras 2.1.2.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    KerasTuner

    KerasTuner

    A Hyperparameter Tuning Library for Keras

    KerasTuner is an easy-to-use, scalable hyperparameter optimization framework that solves the pain points of hyperparameter search. Easily configure your search space with a define-by-run syntax, then leverage one of the available search algorithms to find the best hyperparameter values for your models. KerasTuner comes with Bayesian Optimization, Hyperband, and Random Search algorithms built-in, and is also designed to be easy for researchers to extend in order to experiment with new search algorithms.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    Kernel Adaptive Filtering Toolbox

    Kernel Adaptive Filtering Toolbox

    a Matlab benchmarking toolbox for kernel adaptive filtering

    [Note: This project has moved. Visit https://github.com/steven2358/kafbox/ for the latest version.] A Matlab benchmarking toolbox for kernel adaptive filtering. Kernel adaptive filtering algorithms are online and adaptive regression algorithms based on kernels. They are suitable for nonlinear filtering, prediction, tracking and nonlinear regression in general. This toolbox includes algorithms, demos, and tools to compare their performance. See the included README file for a list of included algorithms and more details. If you use this toolbox in your research please cite: @inproceedings{vanvaerenbergh2013comparative, author = {Van Vaerenbergh, Steven and Santamar{\'i}a, Ignacio}, booktitle = {2013 IEEE Digital Signal Processing (DSP) Workshop and IEEE Signal Processing Education (SPE)}, title = {A Comparative Study of Kernel Adaptive Filtering Algorithms}, year = {2013}, note = {Software available at \url{https://github.com/steven2358/kafbox/}} }
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    Spider that recollects data from MySpace Social Network. At now, it is only designed to extract information from native american people because it is used for a social science study in the UNAM (Universidad Nacional Autónoma de México).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    Knet

    Knet

    Koç University deep learning framework

    Knet.jl is a deep learning package implemented in Julia, so you should be able to run it on any machine that can run Julia. It has been extensively tested on Linux machines with NVIDIA GPUs and CUDA libraries, and it has been reported to work on OSX and Windows. If you would like to try it on your own computer, please follow the instructions on Installation. If you would like to try working with a GPU and do not have access to one, take a look at Using Amazon AWS or Using Microsoft Azure. If you find a bug, please open a GitHub issue. If you don't have access to a GPU machine, but would like to experiment with one, Amazon Web Services is a possible solution. I have prepared a machine image (AMI) with everything you need to run Knet. Here are step-by-step instructions for launching a GPU instance with a Knet image (the screens may have changed slightly since this writing).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    Knock Knock

    Knock Knock

    Get notified when your training ends

    Knock Knock is a lightweight Python utility created by the Hugging Face team that allows developers to receive notifications when long-running machine learning tasks finish or fail. Training deep learning models often takes hours or even days, making it inconvenient for engineers to constantly monitor progress manually. The library solves this problem by adding simple decorators or command-line commands that automatically send notifications when a process completes or crashes. These alerts can be delivered through several communication platforms such as email, Slack, Telegram, or other messaging services. The goal of the project is to allow developers to monitor experiments remotely without needing to stay connected to the training environment. By adding only a few lines of code, the library can wrap around a training function and report execution status.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    Kodezi Chronos

    Kodezi Chronos

    Kodezi Chronos is a debugging-first language model

    Kodezi Chronos is a research project focused on developing a specialized language model designed specifically for debugging software and understanding large code repositories. Unlike general-purpose language models that focus primarily on code generation, Chronos is built to diagnose and repair bugs by analyzing complex relationships across files within a codebase. The project introduces architectural techniques such as Adaptive Graph-Guided Retrieval, which allows the system to navigate large repositories and retrieve relevant debugging information from multiple sources. Another component, Persistent Debug Memory, allows the system to learn patterns from past debugging sessions and apply that knowledge to future problems. The repository mainly contains research documentation, evaluation benchmarks, and experimental frameworks rather than the full proprietary model implementation.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Koila

    Koila

    Prevent PyTorch's `CUDA error: out of memory` in just 1 line of code

    Koila is a lightweight Python library designed to help developers avoid memory errors when training deep learning models with PyTorch. The library introduces a lazy evaluation mechanism that delays computation until it is actually required, allowing the framework to better estimate the memory requirements of a model before execution. By building a computational graph first and executing operations only when necessary, koila reduces the risk of running out of GPU memory during the forward pass of neural network training. This approach enables developers to experiment with larger batch sizes and more complex architectures while maintaining stable training behavior. The system acts as a thin wrapper around PyTorch tensors and operations, meaning that it integrates easily into existing PyTorch code without requiring major changes to model implementations. It is particularly useful in environments where GPU resources are limited or where models frequently encounter CUDA memory errors.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB