Showing 75 open source projects for "inference"

View related business solutions
  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    Let your crypto work for you

    Put idle assets to work with competitive interest rates, borrow without selling, and trade with precision. All in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 1
    MMDeploy

    MMDeploy

    OpenMMLab Model Deployment Framework

    ...Models can be exported and run in several backends, and more will be compatible. All kinds of modules in the SDK can be extended, such as Transform for image processing, Net for Neural Network inference, Module for postprocessing and so on. Install and build your target backend. ONNX Runtime is a cross-platform inference and training accelerator compatible with many popular ML/DNN frameworks. Please read getting_started for the basic usage of MMDeploy.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    YoloV3 Implemented in TensorFlow 2.0

    YoloV3 Implemented in TensorFlow 2.0

    YoloV3 Implemented in Tensorflow 2.0

    ...YOLOv3 works by dividing an image into grid regions and predicting bounding boxes and class probabilities simultaneously, allowing objects to be detected quickly and efficiently. The repository includes training scripts, inference tools, and configuration files that make it possible to train custom object detection models on user-defined datasets. It also demonstrates how to integrate the model with TensorFlow’s high-level APIs such as Keras for easier experimentation and model development. The project supports both pretrained models and full training pipelines, enabling researchers and developers to adapt YOLOv3 for tasks such as surveillance, robotics, autonomous driving, and image analysis.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    LLM Applications

    LLM Applications

    A comprehensive guide to building RAG-based LLM applications

    ...It provides step-by-step guidance for constructing systems that ingest documents, split them into chunks, generate embeddings, index them in vector databases, and retrieve relevant context during inference. The repository also shows how these components can be scaled and deployed using distributed computing frameworks such as Ray. In addition to development workflows, the project includes notebooks, datasets, and evaluation tools that help developers experiment with different retrieval strategies and model configurations.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Lightning Bolts

    Lightning Bolts

    Toolbox of models, callbacks, and datasets for AI/ML researchers

    Bolts package provides a variety of components to extend PyTorch Lightning, such as callbacks & datasets, for applied research and production. Torch ORT converts your model into an optimized ONNX graph, speeding up training & inference when using NVIDIA or AMD GPUs. We can introduce sparsity during fine-tuning with SparseML, which ultimately allows us to leverage the DeepSparse engine to see performance improvements at inference time.
    Downloads: 0 This Week
    Last Update:
    See Project
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 5
    NanoDet-Plus

    NanoDet-Plus

    Lightweight anchor-free object detection model

    ...NanoDet provide multi-backend C++ demo including ncnn, OpenVINO and MNN. There is also an Android demo based on ncnn library. Supports various backends including ncnn, MNN and OpenVINO. Also provide Android demo based on ncnn inference framework.
    Downloads: 10 This Week
    Last Update:
    See Project
  • 6
    Sockeye

    Sockeye

    Sequence-to-sequence framework, focused on Neural Machine Translation

    Sockeye is an open-source sequence-to-sequence framework for Neural Machine Translation built on PyTorch. It implements distributed training and optimized inference for state-of-the-art models, powering Amazon Translate and other MT applications. For a quickstart guide to training a standard NMT model on any size of data, see the WMT 2014 English-German tutorial. If you are interested in collaborating or have any questions, please submit a pull request or issue. You can also send questions to sockeye-dev-at-amazon-dot-com. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Catalyst

    Catalyst

    Accelerated deep learning R&D

    Catalyst is a PyTorch framework for accelerated Deep Learning research and development. It allows you to write compact but full-featured Deep Learning pipelines with just a few lines of code. With Catalyst you get a full set of features including a training loop with metrics, model checkpointing and more, all without the boilerplate. Catalyst is focused on reproducibility, rapid experimentation, and codebase reuse so you can break the cycle of writing another regular train loop and make...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 8
    Deep learning time series forecasting

    Deep learning time series forecasting

    Deep learning PyTorch library for time series forecasting

    ...Historically, this repository provided open-source benchmarks and codes for flash flood and river flow forecasting. Full transformer (SimpleTransformer in model_dict): The full original transformer with all 8 encoder and decoder blocks. Requires passing the target in at inference.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    EasyNLP

    EasyNLP

    EasyNLP: A Comprehensive and Easy-to-use NLP Toolkit

    ...EasyNLP integrates knowledge distillation and few-shot learning for landing large pre-trained models, together with various popular multi-modality pre-trained models. It provides a unified framework of model training, inference, and deployment for real-world applications. It has powered more than 10 BUs and more than 20 business scenarios within the Alibaba group. It is seamlessly integrated to Platform of AI (PAI) products, including PAI-DSW for development, PAI-DLC for cloud-native training, PAI-EAS for serving, and PAI-Designer for zero-code model training.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Enterprise-grade ITSM, for every business Icon
    Enterprise-grade ITSM, for every business

    Give your IT, operations, and business teams the ability to deliver exceptional services—without the complexity.

    Freshservice is an intuitive, AI-powered platform that helps IT, operations, and business teams deliver exceptional service without the usual complexity. Automate repetitive tasks, resolve issues faster, and provide seamless support across the organization. From managing incidents and assets to driving smarter decisions, Freshservice makes it easy to stay efficient and scale with confidence.
    Try it Free
  • 10
    Hugging Face Transformer

    Hugging Face Transformer

    CPU/GPU inference server for Hugging Face transformer models

    ...At Lefebvre Dalloz we run in-production semantic search engines in the legal domain, in the non-marketing language it's a re-ranker, and we based ours on Transformer. In that setup, latency is key to providing a good user experience, and relevancy inference is done online for hundreds of snippets per user query. Most tutorials on Transformer deployment in production are built over Pytorch and FastAPI. Both are great tools but not very performant in inference. Then, if you spend some time, you can build something over ONNX Runtime and Triton inference server. You will usually get from 2X to 4X faster inference compared to vanilla Pytorch. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    TensorFlow Backend for ONNX

    TensorFlow Backend for ONNX

    Tensorflow Backend for ONNX

    Open Neural Network Exchange (ONNX) is an open standard format for representing machine learning models. ONNX is supported by a community of partners who have implemented it in many frameworks and tools. TensorFlow Backend for ONNX makes it possible to use ONNX models as input for TensorFlow. The ONNX model is first converted to a TensorFlow model and then delegated for execution on TensorFlow to produce the output. This is one of the two TensorFlow converter projects which serve different...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    DeepDanbooru

    DeepDanbooru

    AI based multi-label girl image classification system

    DeepDanbooru is a deep learning system designed to automatically tag anime-style images using neural networks trained on datasets derived from the Danbooru imageboard. The project focuses on multi-label image classification, where a model predicts multiple descriptive tags that represent visual elements in an image. These tags may include characters, styles, clothing, emotions, or other attributes associated with anime artwork. The system uses convolutional neural networks trained on large...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 13
    Minkowski Engine

    Minkowski Engine

    Auto-diff neural network library for high-dimensional sparse tensors

    ...We list a few popular network architectures and applications here. To run the examples, please install the package and run the command in the package root directory. Compressing a neural network to speed up inference and minimize memory footprint has been studied widely. One of the popular techniques for model compression is pruning the weights in convnets, is also known as sparse convolutional networks. Such parameter-space sparsity used for model compression compresses networks that operate on dense tensors and all intermediate activations of these networks are also dense tensors.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    BudgetML

    BudgetML

    Deploy a ML inference service on a budget in 10 lines of code

    Deploy a ML inference service on a budget in less than 10 lines of code. BudgetML is perfect for practitioners who would like to quickly deploy their models to an endpoint, but not waste a lot of time, money, and effort trying to figure out how to do this end-to-end. We built BudgetML because it's hard to find a simple way to get a model in production fast and cheaply.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    NLP Architect

    NLP Architect

    A model library for exploring state-of-the-art deep learning

    NLP Architect is an open-source Python library for exploring state-of-the-art deep learning topologies and techniques for optimizing Natural Language Processing and Natural Language Understanding neural networks. The library includes our past and ongoing NLP research and development efforts as part of Intel AI Lab. NLP Architect is designed to be flexible for adding new models, neural network components, data handling methods, and for easy training and running models. NLP Architect is a...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    MMdnn

    MMdnn

    Tools to help users inter-operate among deep learning frameworks

    ...We implement a universal converter to convert DL models between frameworks, which means you can train a model with one framework and deploy it with another. During the model conversion, we generate some code snippets to simplify later retraining or inference. We provide a model collection to help you find some popular models. We provide a model visualizer to display the network architecture more intuitively. We provide some guidelines to help you deploy DL models to another hardware platform.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    NLP-progress

    NLP-progress

    Repository to track the progress in Natural Language Processing (NLP)

    ...It aims to cover both traditional and core NLP tasks such as dependency parsing and part-of-speech tagging as well as more recent ones such as reading comprehension and natural language inference. The main objective is to provide the reader with a quick overview of benchmark datasets and the state-of-the-art for their task of interest, which serves as a stepping stone for further research. To this end, if there is a place where results for a task are already published and regularly maintained, such as a public leaderboard, the reader will be pointed there.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    CrypTen

    CrypTen

    A framework for Privacy Preserving Machine Learning

    ...The framework supports both encryption and decryption of tensors and operations such as addition and multiplication over encrypted values. Although not yet production-ready, CrypTen focuses on advancing real-world secure ML applications, such as training and inference over private datasets, without exposing sensitive data.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    Image Quality Assessment

    Image Quality Assessment

    Convolutional Neural Networks to predict aesthetic quality of images

    ...Instead of relying on simple image statistics, the system learns patterns that correlate with human judgments about image aesthetics and technical quality. The repository includes code for training models, performing inference, and evaluating predicted scores against labeled datasets. It also provides utilities for image preprocessing and data management that help prepare datasets for training deep learning models.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    PyTorch Natural Language Processing

    PyTorch Natural Language Processing

    Basic Utilities for PyTorch Natural Language Processing (NLP)

    ...With your batch in hand, you can use PyTorch to develop and train your model using gradient descent. For example, check out this example code for training on the Stanford Natural Language Inference (SNLI) Corpus. Now you've setup your pipeline, you may want to ensure that some functions run deterministically. Wrap any code that's random, with fork_rng and you'll be good to go. Now that you've computed your vocabulary, you may want to make use of pre-trained word vectors to set your embeddings.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 21
    CakeChat

    CakeChat

    CakeChat: Emotional Generative Dialog System

    ...Multilayer RNN with GRU cells. The first layer of the utterance-level encoder is always bidirectional. By default, CuDNNGRU implementation is used for ~25% acceleration during inference. Thought vector is fed into decoder on each decoding step. Decoder can be conditioned on any categorical label, for example, emotion label or persona id. May be initialized using w2v model trained on your corpus. Embedding layer may be either fixed or fine-tuned along with other weights of the network.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    LUMINOTH

    LUMINOTH

    Deep Learning toolkit for Computer Vision

    ...Luminoth includes support for popular object detection architectures such as Faster R-CNN and SSD, enabling developers to train models on datasets like COCO and Pascal VOC. The toolkit provides command-line utilities for dataset management, training, and inference, making it easier to integrate into research workflows and production systems. Although the project is no longer actively maintained, it remains a useful educational and experimental platform for studying object detection pipelines and deep learning workflows.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    Skater

    Skater

    Python library for model interpretation/explanations

    Skater is a unified framework to enable Model Interpretation for all forms of the model to help one build an Interpretable machine learning system often needed for real-world use-cases(** we are actively working towards to enabling faithful interpretability for all forms models). It is an open-source python library designed to demystify the learned structures of a black box model both globally(inference on the basis of a complete data set) and locally(inference about an individual prediction). The concept of model interpretability in the field of machine learning is still new, largely subjective, and, at times, controversial. Model interpretation is the ability to explain and validate the decisions of a predictive model to enable fairness, accountability, and transparency in algorithmic decision-making. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    The Deep Review

    The Deep Review

    A collaboratively written review paper on deep learning, genomics, etc

    This repository is home to the Deep Review, a review article on deep learning in precision medicine. The Deep Review is collaboratively written on GitHub using a tool called Manubot (see below). The project operates on an open contribution model, welcoming contributions from anyone. To see what's incoming, check the open pull requests. For project discussion and planning see the Issues. As of writing, we are aiming to publish an update of the deep review. We will continue to make project...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25

    Savant

    Python Computer Vision & Video Analytics Framework With Batteries Incl

    Savant is an open-source, high-level framework for building real-time, streaming, highly efficient multimedia AI applications on the Nvidia stack. It helps to develop dynamic, fault-tolerant inference pipelines that utilize the best Nvidia approaches for data center and edge accelerators. Savant is built on DeepStream and provides a high-level abstraction layer for building inference pipelines. It is designed to be easy to use, flexible, and scalable. It is a great choice for building smart CV and video analytics applications for cities, retail, manufacturing, and more.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB