Open Source Linux Machine Learning Software - Page 51

Machine Learning Software for Linux

View 58 business solutions
  • Fully Managed MySQL, PostgreSQL, and SQL Server Icon
    Fully Managed MySQL, PostgreSQL, and SQL Server

    Automatic backups, patching, replication, and failover. Focus on your app, not your database.

    Cloud SQL handles your database ops end to end, so you can focus on your app.
    Try Free
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 1
    spaGO

    spaGO

    Self-contained Machine Learning and Natural Language Processing lib

    A Machine Learning library written in pure Go designed to support relevant neural architectures in Natural Language Processing. Spago is self-contained, in that it uses its own lightweight computational graph both for training and inference, easy to understand from start to finish. The core module of Spago relies only on testify for unit testing. In other words, it has "zero dependencies", and we are committed to keeping it that way as much as possible. Spago uses a multi-module workspace to ensure that additional dependencies are downloaded only when specific features (e.g. persistent embeddings) are used. A good place to start is by looking at the implementation of built-in neural models, such as the LSTM. Except for a few linear algebra operations written in assembly for optimal performance (a bit of copying from Gonum), it's straightforward Go code, so you don't have to worry.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    spark-ml-source-analysis

    spark-ml-source-analysis

    Spark ml algorithm principle analysis and specific source code

    spark-ml-source-analysis is a technical repository that analyzes the internal implementation of machine learning algorithms within Apache Spark’s MLlib library. The project aims to help developers and data scientists understand how distributed machine learning algorithms are implemented and optimized inside the Spark ecosystem. Instead of providing a runnable software system, the repository focuses on explaining algorithm principles and examining the underlying source code used in Spark’s machine learning package. The repository contains detailed analyses of various algorithms including classification, regression, clustering, dimensionality reduction, and recommendation systems. Each section discusses both the mathematical principles behind the algorithms and how Spark implements them in a distributed computing environment. By studying these implementations, readers gain insight into how large-scale machine learning pipelines operate across distributed data systems.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    This project relates to research work at Imperial College conducted by members of the SPIKE (Structured and Probabilistic Intelligent Knowledge Engineering), including in particular logic-based learning systems such as TAL, ASPAL and ILASP.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    stanford-tensorflow-tutorials

    stanford-tensorflow-tutorials

    This repository contains code examples for the Stanford's course

    This repository contains code examples for the course CS 20: TensorFlow for Deep Learning Research. It will be updated as the class progresses. Detailed syllabus and lecture notes can be found in the site. For this course, I use python3.6 and TensorFlow 1.4.1.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 5
    stkpp

    stkpp

    C++ Statistical ToolKit

    STK++ (http://www.stkpp.org) is a versatile, fast, reliable and elegant collection of C++ classes for statistics, clustering, linear algebra, arrays (with an Eigen-like API), regression, dimension reduction, etc. Some functionalities provided by the library are available in the R environment as R functions (http://cran.at.r-project.org/web/packages/rtkore/index.html). At a convenience, we propose the source packages on sourceforge. The library offers a dense set of (mostly) template classes in C++ and is suitable for projects ranging from small one-off projects to complete data mining application suites.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    supabase-py

    supabase-py

    Python Client for Supabase. Query Postgres from Flask, Django

    Python Client for Supabase. Query Postgres from Flask, Django, FastAPI. Python user authentication, security policies, edge functions, file storage, and realtime data streaming. Good first issue.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Class library for efficient large-scale text mining research on Linux.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    talos

    talos

    Hyperparameter Optimization for TensorFlow, Keras and PyTorch

    Talos radically changes the ordinary Keras, TensorFlow (tf.keras), and PyTorch workflow by fully automating hyperparameter tuning and model evaluation. Talos exposes Keras and TensorFlow (tf.keras) and PyTorch functionality entirely and there is no new syntax or templates to learn. Talos is made for data scientists and data engineers that want to remain in complete control of their TensorFlow (tf.keras) and PyTorch models, but are tired of mindless parameter hopping and confusing optimization solutions that add complexity instead of reducing it. Within minutes, without learning any new syntax, Talos allows you to configure, perform, and evaluate hyperparameter optimization experiments that yield state-of-the-art results across a wide range of prediction tasks. Talos provides the simplest and yet most powerful available method for hyperparameter optimization with TensorFlow (tf.keras) and PyTorch.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    tensorflow_template_application

    tensorflow_template_application

    TensorFlow template application for deep learning

    tensorflow_template_application is a template project that demonstrates how to structure scalable applications built with TensorFlow. The repository provides a standardized architecture that helps developers organize machine learning code into clear components such as data processing, model training, evaluation, and deployment. Instead of focusing on a specific algorithm, the project emphasizes software engineering practices that make machine learning systems easier to maintain and extend. The template includes configuration files, scripts, and project structures that help teams build reproducible experiments and production-ready pipelines. It is particularly useful for developers who want to transition from experimental notebooks to structured machine learning applications. By providing a reusable framework, the template reduces the time needed to set up new TensorFlow projects and encourages consistent development practices.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Go from Code to Production URL in Seconds Icon
    Go from Code to Production URL in Seconds

    Cloud Run deploys apps in any language instantly. Scales to zero. Pay only when code runs.

    Skip the Kubernetes configs. Cloud Run handles HTTPS, scaling, and infrastructure automatically. Two million requests free per month.
    Try it free
  • 10
    textgenrnn

    textgenrnn

    Easily train your own text-generating neural network

    With textgenrnn you can easily train your own text-generating neural network of any size and complexity on any text dataset with a few lines of code. A modern neural network architecture that utilizes new techniques as attention-weighting and skip-embedding to accelerate training and improve model quality. Train on and generate text at either the character-level or word-level. Configure RNN size, the number of RNN layers, and whether to use bidirectional RNNs. Train on any generic input text file, including large files. Train models on a GPU and then use them to generate text with a CPU. Utilize a powerful CuDNN implementation of RNNs when trained on the GPU, which massively speeds up training time as opposed to typical LSTM implementations. Train the model using contextual labels, allowing it to learn faster and produce better results in some cases.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    tf2onnx

    tf2onnx

    Convert TensorFlow, Keras, Tensorflow.js and Tflite models to ONNX

    tf2onnx converts TensorFlow (tf-1.x or tf-2.x), keras, tensorflow.js and tflite models to ONNX via command line or python API. Note: tensorflow.js support was just added. While we tested it with many tfjs models from tfhub, it should be considered experimental. TensorFlow has many more ops than ONNX and occasionally mapping a model to ONNX creates issues. tf2onnx will use the ONNX version installed on your system and installs the latest ONNX version if none is found. We support and test ONNX opset-13 to opset-17. opset-6 to opset-12 should work but we don't test them. If you want the graph to be generated with a specific opset, use --opset in the command line, for example --opset 13. When running under tf-2.x tf2onnx will use the tensorflow V2 controlflow.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    thorough-pytorch

    thorough-pytorch

    PyTorch Getting Started Tutorial, read online

    thorough-pytorch is an educational project designed to teach deep learning using the PyTorch framework through a structured learning series. The repository provides tutorials and practical exercises that guide learners from fundamental PyTorch concepts to more advanced deep learning techniques. It emphasizes a learning approach that combines theoretical explanations with hands-on coding exercises so that students can build and experiment with neural networks directly. The project encourages collaborative learning and often organizes materials in a step-by-step progression that gradually increases in complexity. Topics include neural network fundamentals, training procedures, model evaluation, and practical deep learning workflows. By combining structured lessons with programming projects, the repository aims to help learners develop both conceptual understanding and practical implementation skills.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    tika-python

    tika-python

    Python binding to the Apache Tika™ REST services

    A Python port of the Apache Tika library that makes Tika available using the Tika REST Server. This makes Apache Tika available as a Python library, installable via Setuptools, Pip and easy to install. To use this library, you need to have Java 7+ installed on your system as tika-python starts up the Tika REST server in the background. To get this working in a disconnected environment, download a tika server file (both tika-server.jar and tika-server.jar.md5, which can be found here) and set the TIKA_SERVER_JAR environment variable to TIKA_SERVER_JAR="file:////tika-server.jar" which successfully tells python-tika to "download" this file and move it to /tmp/tika-server.jar and run as a background process. This is the only way to run python-tika without internet access. Without this set, the default is to check the tika version and pull latest every time from Apache.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    tiny-dnn

    tiny-dnn

    header only, dependency-free deep learning framework in C++14

    tiny-dnn is a C++14 implementation of deep learning. It is suitable for deep learning on limited computational resources, embedded systems, and IoT devices.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    torchtext

    torchtext

    Data loaders and abstractions for text and NLP

    We recommend Anaconda as a Python package management system. Please refer to pytorch.org for the details of PyTorch installation. LTS versions are distributed through a different channel than the other versioned releases. Alternatively, you might want to use the Moses tokenizer port in SacreMoses (split from NLTK). You have to install SacreMoses. To build torchtext from source, you need git, CMake and C++11 compiler such as g++. When building from source, make sure that you have the same C++ compiler as the one used to build PyTorch. A simple way is to build PyTorch from source and use the same environment to build torchtext. If you are using the nightly build of PyTorch, check out the environment it was built with conda (here) and pip (here). Text classification: SST2, AG_NEWS, SogouNews, DBpedia, YelpReviewPolarity, YelpReviewFull, YahooAnswers, AmazonReviewPolarity, AmazonReviewFull, IMDB, etc.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    transferlearning-tutorial

    transferlearning-tutorial

    Tutorial on applied transfer learning

    A bilingual (English/Chinese) concise tutorial on applied transfer learning, hosted by Jindong Wang. It includes LaTeX source for a compact handbook covering theory, algorithms, surveys, and code in MATLAB and Python.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    tslearn

    tslearn

    The machine learning toolkit for time series analysis in Python

    The machine learning toolkit for time series analysis in Python. tslearn expects a time series dataset to be formatted as a 3D numpy array. The three dimensions correspond to the number of time series, the number of measurements per time series and the number of dimensions respectively (n_ts, max_sz, d). In order to get the data in the right format.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    tvm

    tvm

    Open deep learning compiler stack for cpu, gpu, etc.

    Apache TVM is an open source machine learning compiler framework for CPUs, GPUs, and machine learning accelerators. It aims to enable machine learning engineers to optimize and run computations efficiently on any hardware backend. The vision of the Apache TVM Project is to host a diverse community of experts and practitioners in machine learning, compilers, and systems architecture to build an accessible, extensible, and automated open-source framework that optimizes current and emerging machine learning models for any hardware platform. Compilation of deep learning models in Keras, MXNet, PyTorch, Tensorflow, CoreML, DarkNet and more. Start using TVM with Python today, build out production stacks using C++, Rust, or Java the next day.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    This project has been renamed to oooark. Old file releases will still be available here. uvsim is a project focused on enabling algorithm development for unmanned systems. It is being constructed to provide an identical interface to simulations and h
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20

    virgo

    32 bit VIRGO Linux Kernel

    Linux kernel fork-off with cloud and machine learning features
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    wav2letter++

    wav2letter++

    Facebook AI research's automatic speech recognition toolkit

    First, install Flashlight (using the 0.3 branch is required) with the ASR application. This repository includes recipes to reproduce the following research papers as well as pre-trained models. All results reproduction must use Flashlight <= 0.3.2 for exact reproducibility. At least one of LZMA, BZip2, or Z is required for LM compression with KenLM. It is highly recommended to build KenLM with position-independent code (-fPIC) enabled, to enable python compatibility. After installing, run export KENLM_ROOT_DIR=... so that wav2letter++ can find it. This is needed because KenLM doesn't support a make install step.wav2letter++ expects audio and transcription data to be prepared in a specific format so that they can be read from the pipelines. Each dataset (test/valid/train) needs to be in a separate file with one sample per line. A sample is specified using 4 columns separated by space (or tabs).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    weka outlier is an implementation of outlier detection algorithms for WEKA. CODB (Class Outliers: Distance-Based) Algorithm is the first algorithm developed using WEKA framework.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    whisper-timestamped

    whisper-timestamped

    Multilingual Automatic Speech Recognition with word-level timestamps

    Multilingual Automatic Speech Recognition with word-level timestamps and confidence. Whisper is a set of multi-lingual, robust speech recognition models trained by OpenAI that achieve state-of-the-art results in many languages. Whisper models were trained to predict approximate timestamps on speech segments (most of the time with 1-second accuracy), but they cannot originally predict word timestamps. This repository proposes an implementation to predict word timestamps and provide a more accurate estimation of speech segments when transcribing with Whisper models. Besides, a confidence score is assigned to each word and each segment.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    x-transformers

    x-transformers

    A simple but complete full-attention transformer

    A simple but complete full-attention transformer with a set of promising experimental features from various papers. Proposes adding learned memory key/values prior to attending. They were able to remove feedforwards altogether and attain a similar performance to the original transformers. I have found that keeping the feedforwards and adding the memory key/values leads to even better performance. Proposes adding learned tokens, akin to CLS tokens, named memory tokens, that is passed through the attention layers alongside the input tokens. You can also use the l2 normalized embeddings proposed as part of fixnorm. I have found it leads to improved convergence when paired with small initialization (proposed by BlinkDL). The small initialization will be taken care of as long as l2norm_embed is set to True.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    yacaree
    yacaree: Yet Another Closure-based Association Rules Experimentation Environment (Development stalled here, check out under Wiki to find where it went.)
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB