Open Source Linux Deep Learning Frameworks - Page 9

Deep Learning Frameworks for Linux

View 20 business solutions
  • Fully Managed MySQL, PostgreSQL, and SQL Server Icon
    Fully Managed MySQL, PostgreSQL, and SQL Server

    Automatic backups, patching, replication, and failover. Focus on your app, not your database.

    Cloud SQL handles your database ops end to end, so you can focus on your app.
    Try Free
  • Full-stack observability with actually useful AI | Grafana Cloud Icon
    Full-stack observability with actually useful AI | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 1
    PyTorch Book

    PyTorch Book

    PyTorch tutorials and fun projects including neural talk

    This is the corresponding code for the book "The Deep Learning Framework PyTorch: Getting Started and Practical", but it can also be used as a standalone PyTorch Getting Started Guide and Tutorial. The current version of the code is based on pytorch 1.0.1, if you want to use an older version please git checkout v0.4or git checkout v0.3. Legacy code has better python2/python3 compatibility, CPU/GPU compatibility test. The new version of the code has not been fully tested, it has been tested under GPU and python3. But in theory there shouldn't be too many problems on python2 and CPU. The basic part (the first five chapters) explains the content of PyTorch. This part introduces the main modules in PyTorch and some tools commonly used in deep learning. For this part of the content, Jupyter Notebook is used as a teaching tool here, and readers can modify and run with notebooks and repeat experiments.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    PyTorch Geometric

    PyTorch Geometric

    Geometric deep learning extension library for PyTorch

    It consists of various methods for deep learning on graphs and other irregular structures, also known as geometric deep learning, from a variety of published papers. In addition, it consists of an easy-to-use mini-batch loader for many small and single giant graphs, a large number of common benchmark datasets (based on simple interfaces to create your own), and helpful transforms, both for learning on arbitrary graphs as well as on 3D meshes or point clouds. We have outsourced a lot of functionality of PyTorch Geometric to other packages, which needs to be additionally installed. These packages come with their own CPU and GPU kernel implementations based on C++/CUDA extensions. We do not recommend installation as root user on your system python. Please setup an Anaconda/Miniconda environment or create a Docker image. We provide pip wheels for all major OS/PyTorch/CUDA combinations.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    PyTorch Geometric Temporal

    PyTorch Geometric Temporal

    Spatiotemporal Signal Processing with Neural Machine Learning Models

    The library consists of various dynamic and temporal geometric deep learning, embedding, and Spatio-temporal regression methods from a variety of published research papers. Moreover, it comes with an easy-to-use dataset loader, train-test splitter and temporal snaphot iterator for dynamic and temporal graphs. The framework naturally provides GPU support. It also comes with a number of benchmark datasets from the epidemiological forecasting, sharing economy, energy production and web traffic management domains. Finally, you can also create your own datasets. The package interfaces well with Pytorch Lightning which allows training on CPUs, single and multiple GPUs out-of-the-box. PyTorch Geometric Temporal makes implementing Dynamic and Temporal Graph Neural Networks quite easy - see the accompanying tutorial. Head over to our documentation to find out more about installation, creation of datasets and a full list of implemented methods and available datasets.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    PyTorch/XLA

    PyTorch/XLA

    Enabling PyTorch on Google TPU

    PyTorch/XLA is a Python package that uses the XLA deep learning compiler to connect the PyTorch deep learning framework and Cloud TPUs. You can try it right now, for free, on a single Cloud TPU with Google Colab, and use it in production and on Cloud TPU Pods with Google Cloud. Take a look at one of our Colab notebooks to quickly try different PyTorch networks running on Cloud TPUs and learn how to use Cloud TPUs as PyTorch devices. We are also introducing new TPU VMs for more transparent and easier access to the TPU hardware. This is our recommedned way of running PyTorch/XLA on Cloud TPU. Please check out our Cloud TPU VM User Guide. Cloud TPU VM is currently on general availability and provides direct access to the TPU host. The recommended setup for running distributed training on TPU Pods uses the pairing of Compute VM Instance Groups and TPU Pods. Each of the Compute VM in the instance group drives 8 cores on the TPU Pod.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 5
    Pytorch Points 3D

    Pytorch Points 3D

    Pytorch framework for doing deep learning on point clouds

    Torch Points 3D is a framework for developing and testing common deep learning models to solve tasks related to unstructured 3D spatial data i.e. Point Clouds. The framework currently integrates some of the best-published architectures and it integrates the most common public datasets for ease of reproducibility. It heavily relies on Pytorch Geometric and Facebook Hydra library thanks for the great work! We aim to build a tool that can be used for benchmarking SOTA models, while also allowing practitioners to efficiently pursue research into point cloud analysis, with the end goal of building models which can be applied to real-life applications. Task driven implementation with dynamic model and dataset resolution from arguments. Core implementation of common components for point cloud deep learning - greatly simplifying the creation of new models. 4 Base Convolution base classes to simplify the implementation of new convolutions. Each base class supports a different data format.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    Raster Vision

    Raster Vision

    Open source framework for deep learning satellite and aerial imagery

    Raster Vision is an open source framework for Python developers building computer vision models on satellite, aerial, and other large imagery sets (including oblique drone imagery). There is built-in support for chip classification, object detection, and semantic segmentation using PyTorch. Raster Vision allows engineers to quickly and repeatably configure pipelines that go through core components of a machine learning workflow: analyzing training data, creating training chips, training models, creating predictions, evaluating models, and bundling the model files and configuration for easy deployment. The input to a Raster Vision pipeline is a set of images and training data, optionally with Areas of Interest (AOIs) that describe where the images are labeled. The output of a Raster Vision pipeline is a model bundle that allows you to easily utilize models in various deployment scenarios.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Recommenders

    Recommenders

    Best practices on recommendation systems

    The Recommenders repository provides examples and best practices for building recommendation systems, provided as Jupyter notebooks. The module reco_utils contains functions to simplify common tasks used when developing and evaluating recommender systems. Several utilities are provided in reco_utils to support common tasks such as loading datasets in the format expected by different algorithms, evaluating model outputs, and splitting training/test data. Implementations of several state-of-the-art algorithms are included for self-study and customization in your own applications. Please see the setup guide for more details on setting up your machine locally, on a data science virtual machine (DSVM) or on Azure Databricks. Independent or incubating algorithms and utilities are candidates for the contrib folder. This will house contributions which may not easily fit into the core repository or need time to refactor or mature the code and add necessary tests.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    Rhino

    Rhino

    On-device Speech-to-Intent engine powered by deep learning

    Rhino is Picovoice's Speech-to-Intent engine. It directly infers intent from spoken commands within a given context of interest, in real-time. The end-to-end platform for embedding private voice AI into any software in a few lines of code. Design with no limits on top of a modular platform. Create use-case-specific voice AI models in seconds. Develop voice features with a few lines of code using intuitive and cross-platform SDKs. Deliver voice AI everywhere: on-device, mobile, web browsers, on-premise, or cloud. Measure adoption, learn, and iterate. Continuously re-design and re-train to optimize engagement. Building accurate, responsive, and private voice technology is difficult. We learned the hard way, so you don’t have to. Picovoice heavily invests in R&D to offer superior voice AI that surpasses even Big Tech in accuracy and efficiency. Picovoice researchers do not follow recent frameworks and techniques but build them.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    SINGA

    SINGA

    A distributed deep learning platform

    Apache SINGA is an Apache Top Level Project, focusing on distributed training of deep learning and machine learning models. Various example deep learning models are provided in SINGA repo on Github and on Google Colab. SINGA supports data parallel training across multiple GPUs (on a single node or across different nodes). SINGA supports various popular optimizers including stochastic gradient descent with momentum, Adam, RMSProp, and AdaGrad, etc. SINGA records the computation graph and applies the backward propagation automatically after forward propagation. The optimization of memory are implemented in the Device class. SINGA supports loading ONNX format models and saving models defined using SINGA APIs into ONNX format, which enables AI developers to use models across different libraries and tools. SINGA supports the time profiling of each of the operators buffered in the graph. Half precision is supported to bring benefits.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 10
    SOD

    SOD

    An Embedded Computer Vision & Machine Learning Library

    SOD is an embedded, modern cross-platform computer vision and machine learning software library that expose a set of APIs for deep-learning, advanced media analysis & processing including real-time, multi-class object detection and model training on embedded systems with limited computational resource and IoT devices. SOD was built to provide a common infrastructure for computer vision applications and to accelerate the use of machine perception in open source as well as commercial products. SOD implements state-of-the-art computer vision algorithms found to be mandatory in real-world application areas. Sobel operator, Otsu's binarization and over 100 image/frame processing & analysis interfaces. Designed for computational efficiency and with a strong focus on real-time applications. SOD includes a comprehensive set of both classic and state-of-the-art deep-neural networks with their pre-trained models.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    SageMaker MXNet Inference Toolkit

    SageMaker MXNet Inference Toolkit

    Toolkit for allowing inference and serving with MXNet in SageMaker

    SageMaker MXNet Inference Toolkit is an open-source library for serving MXNet models on Amazon SageMaker. This library provides default pre-processing, predict and postprocessing for certain MXNet model types and utilizes the SageMaker Inference Toolkit for starting up the model server, which is responsible for handling inference requests. AWS Deep Learning Containers (DLCs) are a set of Docker images for training and serving models in TensorFlow, TensorFlow 2, PyTorch, and MXNet. Deep Learning Containers provide optimized environments with TensorFlow and MXNet, Nvidia CUDA (for GPU instances), and Intel MKL (for CPU instances) libraries and are available in the Amazon Elastic Container Registry (Amazon ECR). The AWS DLCs are used in Amazon SageMaker as the default vehicles for your SageMaker jobs such as training, inference, transforms etc. They've been tested for machine learning workloads on Amazon EC2, Amazon ECS and Amazon EKS services as well.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    SageMaker MXNet Training Toolkit

    SageMaker MXNet Training Toolkit

    Toolkit for running MXNet training scripts on SageMaker

    SageMaker MXNet Training Toolkit is an open-source library for using MXNet to train models on Amazon SageMaker. For inference, see SageMaker MXNet Inference Toolkit. For the Dockerfiles used for building SageMaker MXNet Containers, see AWS Deep Learning Containers. For information on running MXNet jobs on Amazon SageMaker, please refer to the SageMaker Python SDK documentation. With the SDK, you can train and deploy models using popular deep learning frameworks Apache MXNet and TensorFlow. You can also train and deploy models with Amazon algorithms, which are scalable implementations of core machine learning algorithms that are optimized for SageMaker and GPU training. If you have your own algorithms built into SageMaker compatible Docker containers, you can train and host models using these as well.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Seldon Server

    Seldon Server

    Machine learning platform and recommendation engine on Kubernetes

    Seldon Server is a machine learning platform and recommendation engine built on Kubernetes. Seldon reduces time-to-value so models can get to work faster. Scale with confidence and minimize risk through interpretable results and transparent model performance. Seldon Core focuses purely on deploying a wide range of ML models on Kubernetes, allowing complex runtime serving graphs to be managed in production. Seldon Core is a progression of the goals of the Seldon-Server project but also a more restricted focus to solving the final step in a machine learning project which is serving models in production. Seldon Server is a machine learning platform that helps your data science team deploy models into production. It provides an open-source data science stack that runs within a Kubernetes Cluster. You can use Seldon to deploy machine learning and deep learning models into production on-premise or in the cloud (e.g. GCP, AWS, Azure).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Spektral

    Spektral

    Graph Neural Networks with Keras and Tensorflow 2

    Spektral is a Python library for graph deep learning, based on the Keras API and TensorFlow 2. The main goal of this project is to provide a simple but flexible framework for creating graph neural networks (GNNs). You can use Spektral for classifying the users of a social network, predicting molecular properties, generating new graphs with GANs, clustering nodes, predicting links, and any other task where data is described by graphs. Spektral implements some of the most popular layers for graph deep learning. Spektral also includes lots of utilities for representing, manipulating, and transforming graphs in your graph deep learning projects. Spektral is compatible with Python 3.6 and above, and is tested on the latest versions of Ubuntu, MacOS, and Windows. Other Linux distros should work as well. The 1.0 release of Spektral is an important milestone for the library and brings many new features and improvements.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Surface Defect Detection Dataset Papers

    Surface Defect Detection Dataset Papers

    Constantly summarizing open source dataset and critical papers

    At present, surface defect equipment based on machine vision has widely replaced artificial visual inspection in various industrial fields, including 3C, automobiles, home appliances, machinery manufacturing, semiconductors and electronics, chemical, pharmaceutical, aerospace, light industry and other industries. Traditional surface defect detection methods based on machine vision often use conventional image processing algorithms or artificially designed features plus classifiers. Generally speaking, imaging schemes are usually designed by using the different properties of the inspected surface or defects. A reasonable imaging scheme helps to obtain images with uniform illumination and clearly reflect the surface defects of the object. In recent years, many defect detection methods based on deep learning have also been widely used in various industrial scenarios.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Swift AI

    Swift AI

    The Swift machine learning library

    Swift AI is a high-performance deep learning library written entirely in Swift. We currently offer support for all Apple platforms, with Linux support coming soon. Swift AI includes a collection of common tools used for artificial intelligence and scientific applications. A flexible, fully-connected neural network with support for deep learning. Optimized specifically for Apple hardware, using advanced parallel processing techniques. We've created some example projects to demonstrate the usage of Swift AI. Each resides in their own repository and can be built with little or no configuration. Each module now contains its own documentation. We recommend that you read the docs carefully for detailed instructions on using the various components of Swift AI. The example projects are another great resource for seeing real-world usage of these tools. Swift AI currently depends on Apple's Accelerate framework for vector/matrix calculations and digital signal processing.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Synapse Machine Learning

    Synapse Machine Learning

    Simple and distributed Machine Learning

    SynapseML (previously MMLSpark) is an open source library to simplify the creation of scalable machine learning pipelines. SynapseML builds on Apache Spark and SparkML to enable new kinds of machine learning, analytics, and model deployment workflows. SynapseML adds many deep learning and data science tools to the Spark ecosystem, including seamless integration of Spark Machine Learning pipelines with the Open Neural Network Exchange (ONNX), LightGBM, The Cognitive Services, Vowpal Wabbit, and OpenCV. These tools enable powerful and highly-scalable predictive and analytical models for a variety of data sources. SynapseML also brings new networking capabilities to the Spark Ecosystem. With the HTTP on Spark project, users can embed any web service into their SparkML models. For production-grade deployment, the Spark Serving project enables high throughput, sub-millisecond latency web services, backed by your Spark cluster.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    T81 558

    T81 558

    Applications of Deep Neural Networks

    Deep learning is a group of exciting new technologies for neural networks. Through a combination of advanced training techniques and neural network architectural components, it is now possible to create neural networks that can handle tabular data, images, text, and audio as both input and output. Deep learning allows a neural network to learn hierarchies of information in a way that is like the function of the human brain. This course will introduce the student to classic neural network structures, Convolution Neural Networks (CNN), Long Short-Term Memory (LSTM), Gated Recurrent Neural Networks (GRU), General Adversarial Networks (GAN) and reinforcement learning. Application of these architectures to computer vision, time series, security, natural language processing (NLP), and data generation will be covered. High-Performance Computing (HPC) aspects will demonstrate how deep learning can be leveraged both on graphical processing units (GPUs), as well as grids.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    TFLearn

    TFLearn

    Deep learning library featuring a higher-level API for TensorFlow

    TFlearn is a modular and transparent deep learning library built on top of Tensorflow. It was designed to provide a higher-level API to TensorFlow in order to facilitate and speed up experimentations while remaining fully transparent and compatible with it. Easy-to-use and understand high-level API for implementing deep neural networks, with tutorials and examples. Fast prototyping through highly modular built-in neural network layers, regularizers, optimizers, and metrics. Full transparency over Tensorflow. All functions are built over tensors and can be used independently of TFLearn. Powerful helper functions to train any TensorFlow graph, with support of multiple inputs, outputs, and optimizers. Easy and beautiful graph visualization, with details about weights, gradients, activations, and more. Effortless device placement for using multiple CPU/GPU. The high-level API currently supports the most of the recent deep learning models, such as Convolutions, LSTM, BiRNN, BatchNorm, etc.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    TNN

    TNN

    Uniform deep learning inference framework for mobile

    TNN, a high-performance, lightweight neural network inference framework open sourced by Tencent Youtu Lab. It also has many outstanding advantages such as cross-platform, high performance, model compression, and code tailoring. The TNN framework further strengthens the support and performance optimization of mobile devices on the basis of the original Rapidnet and ncnn frameworks. At the same time, it refers to the high performance and good scalability characteristics of the industry's mainstream open source frameworks, and expands the support for X86 and NV GPUs. On the mobile phone, TNN has been used by many applications such as mobile QQ, weishi, and Pitu. As a basic acceleration framework for Tencent Cloud AI, TNN has provided acceleration support for the implementation of many businesses. Everyone is welcome to participate in the collaborative construction to promote the further improvement of the TNN inference framework.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    Tensor2Tensor

    Tensor2Tensor

    Library of deep learning models and datasets

    Deep Learning (DL) has enabled the rapid advancement of many useful technologies, such as machine translation, speech recognition and object detection. In the research community, one can find code open-sourced by the authors to help in replicating their results and further advancing deep learning. However, most of these DL systems use unique setups that require significant engineering effort and may only work for a specific problem or architecture, making it hard to run new experiments and compare the results. Tensor2Tensor, or T2T for short, is a library of deep learning models and datasets designed to make deep learning more accessible and accelerate ML research. T2T was developed by researchers and engineers in the Google Brain team and a community of users. It is now deprecated, we keep it running and welcome bug-fixes, but encourage users to use the successor library Trax.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    TensorFlow Course

    TensorFlow Course

    Simple and ready-to-use tutorials for TensorFlow

    This repository houses a highly popular (~16k stars) set of TensorFlow tutorials and example code aimed at beginners and intermediate users. It includes Jupyter notebooks and scripts that cover neural network fundamentals, model training, deployment, and more, with support for Google Colab.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    TensorFlow Probability

    TensorFlow Probability

    Probabilistic reasoning and statistical analysis in TensorFlow

    TensorFlow Probability is a library for probabilistic reasoning and statistical analysis. TensorFlow Probability (TFP) is a Python library built on TensorFlow that makes it easy to combine probabilistic models and deep learning on modern hardware (TPU, GPU). It's for data scientists, statisticians, ML researchers, and practitioners who want to encode domain knowledge to understand data and make predictions. Since TFP inherits the benefits of TensorFlow, you can build, fit, and deploy a model using a single language throughout the lifecycle of model exploration and production. TFP is open source and available on GitHub. Tools to build deep probabilistic models, including probabilistic layers and a `JointDistribution` abstraction. Variational inference and Markov chain Monte Carlo. A wide selection of probability distributions and bijectors. Optimizers such as Nelder-Mead, BFGS, and SGLD.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    TensorFlow Ranking

    TensorFlow Ranking

    Learning to rank in TensorFlow

    TensorFlow Ranking is a library for Learning-to-Rank (LTR) techniques on the TensorFlow platform. Commonly used loss functions including pointwise, pairwise, and listwise losses. Commonly used ranking metrics like Mean Reciprocal Rank (MRR) and Normalized Discounted Cumulative Gain (NDCG). Multi-item (also known as groupwise) scoring functions. LambdaLoss implementation for direct ranking metric optimization. Unbiased Learning-to-Rank from biased feedback data. We envision that this library will provide a convenient open platform for hosting and advancing state-of-the-art ranking models based on deep learning techniques, and thus facilitate both academic research and industrial applications. We provide a demo, with no installation required, to get started on using TF-Ranking. This demo runs on a colaboratory notebook, an interactive Python environment. Using sparse features and embeddings in TF-Ranking.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    TensorFlow World

    TensorFlow World

    Simple and ready-to-use tutorials for TensorFlow

    This repository aims to provide simple and ready-to-use tutorials for TensorFlow. The explanations are present in the wiki associated with this repository. There are different motivations for this open source project. TensorFlow (as we write this document) is one of / the best deep learning frameworks available. The question that should be asked is why has this repository been created when there are so many other tutorials about TensorFlow available on the web? Deep Learning is in very high interest these days - there's a crucial need for rapid and optimized implementations of the algorithms and architectures. TensorFlow is designed to facilitate this goal. The strong advantage of TensorFlow is it flexibility in designing highly modular models which can also be a disadvantage for beginners since a lot of the pieces must be considered together when creating the model.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB