Showing 36 open source projects for "network performance"

View related business solutions
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 1
    TensorRT

    TensorRT

    C++ library for high performance inference on NVIDIA GPUs

    NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference. It includes a deep learning inference optimizer and runtime that delivers low latency and high throughput for deep learning inference applications. TensorRT-based applications perform up to 40X faster than CPU-only platforms during inference. With TensorRT, you can optimize neural network models trained in all major frameworks, calibrate for lower precision with high accuracy, and deploy to hyperscale data centers, embedded, or automotive product platforms. ...
    Downloads: 41 This Week
    Last Update:
    See Project
  • 2
    Velocity server

    Velocity server

    The modern, next-generation Minecraft server proxy

    Velocity is a high-performance Minecraft proxy server developed by the PaperMC project that allows multiple Minecraft servers to be linked together into a single network. Acting as an intermediary between players and backend servers, Velocity manages player connections and routes them to different game servers within a network. This architecture allows large Minecraft communities to run multiple servers for different game modes while presenting them as a unified system to players. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 3
    imgclsmob Deep learning networks

    imgclsmob Deep learning networks

    Sandbox for training deep learning networks

    ...The project is frequently used by developers who want to study modern convolutional neural network designs and compare their performance across datasets.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 4
    oneDNN

    oneDNN

    oneAPI Deep Neural Network Library (oneDNN)

    This software was previously known as Intel(R) Math Kernel Library for Deep Neural Networks (Intel(R) MKL-DNN) and Deep Neural Network Library (DNNL). oneAPI Deep Neural Network Library (oneDNN) is an open-source cross-platform performance library of basic building blocks for deep learning applications. oneDNN is part of oneAPI. The library is optimized for Intel(R) Architecture Processors, Intel Processor Graphics and Xe Architecture graphics. oneDNN has experimental support for the following architectures: Arm* 64-bit Architecture (AArch64), NVIDIA* GPU, OpenPOWER* Power ISA (PPC64), IBMz* (s390x), and RISC-V. oneDNN is intended for deep learning applications and framework developers interested in improving application performance on Intel CPUs and GPUs. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • Powerful App Monitoring Without Surprise Bills Icon
    Powerful App Monitoring Without Surprise Bills

    AppSignal starts at $23/month with all features included. No overages, no hidden fees. 30-day free trial.

    Tired of monitoring tools that punish you for scaling? AppSignal offers transparent, predictable pricing with every feature unlocked on every plan. Track errors, monitor performance, detect anomalies, and manage logs across Ruby, Python, Node.js, and more. Trusted by developers since 2012 with free dev-to-dev support. No credit card required to start your 30-day trial.
    Try AppSignal Free
  • 5
    face.evoLVe

    face.evoLVe

    High-Performance Face Recognition Library on PaddlePaddle & PyTorch

    ...The repository supports multiple neural network backbones such as ResNet, DenseNet, MobileNet, and ShuffleNet, enabling experimentation with different architectures depending on performance requirements. It also implements a wide range of loss functions commonly used in face recognition research, including ArcFace, CosFace, Triplet loss, and Softmax variants. To improve scalability, the library introduces distributed training techniques that allow large models to be trained efficiently across multiple GPUs.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 6
    Gen.jl

    Gen.jl

    A general-purpose probabilistic programming system

    ...But it also lets users migrate parts of their model or inference algorithm to specialized modeling languages for which it can generate especially fast code. Users can also hand-code parts of their models that demand better performance. Neural network inference is fast, but can be inaccurate on out-of-distribution data, and requires expensive training.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 7
    TensorFlow Quantum

    TensorFlow Quantum

    Open-source Python framework for hybrid quantum-classical ml learning

    TensorFlow Quantum is an open-source software framework designed for building and training hybrid quantum-classical machine learning models within the TensorFlow ecosystem. The framework enables researchers and developers to represent quantum circuits as data and integrate them directly into machine learning workflows. By combining classical deep learning techniques with quantum algorithms, the platform allows experimentation with quantum machine learning methods that may offer advantages...
    Downloads: 10 This Week
    Last Update:
    See Project
  • 8
    Intel Extension for PyTorch

    Intel Extension for PyTorch

    A Python package for extending the official PyTorch

    Intel® Extension for PyTorch* extends PyTorch* with up-to-date features optimizations for an extra performance boost on Intel hardware. Optimizations take advantage of Intel® Advanced Vector Extensions 512 (Intel® AVX-512) Vector Neural Network Instructions (VNNI) and Intel® Advanced Matrix Extensions (Intel® AMX) on Intel CPUs as well as Intel Xe Matrix Extensions (XMX) AI engines on Intel discrete GPUs. Moreover, Intel® Extension for PyTorch* provides easy GPU acceleration for Intel discrete GPUs through the PyTorch* xpu device.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 9
    Core ML Tools

    Core ML Tools

    Core ML tools contain supporting tools for Core ML model conversion

    ...Core ML provides a unified representation for all models. Your app uses Core ML APIs and user data to make predictions, and to fine-tune models, all on the user’s device. Core ML optimizes on-device performance by leveraging the CPU, GPU, and Neural Engine while minimizing its memory footprint and power consumption. Running a model strictly on the user’s device removes any need for a network connection, which helps keep the user’s data private and your app responsive.
    Downloads: 7 This Week
    Last Update:
    See Project
  • Auth0 B2B Essentials: SSO, MFA, and RBAC Built In Icon
    Auth0 B2B Essentials: SSO, MFA, and RBAC Built In

    Unlimited organizations, 3 enterprise SSO connections, role-based access control, and pro MFA included. Dev and prod tenants out of the box.

    Auth0's B2B Essentials plan gives you everything you need to ship secure multi-tenant apps. Unlimited orgs, enterprise SSO, RBAC, audit log streaming, and higher auth and API limits included. Add on M2M tokens, enterprise MFA, or additional SSO connections as you scale.
    Sign Up Free
  • 10
    ANE Training

    ANE Training

    Training neural networks on Apple Neural Engine via APIs

    ANE Training is an experimental research project that demonstrates how to train neural networks directly on Apple’s Neural Engine by leveraging reverse-engineered private APIs that are normally inaccessible to developers. The repository implements a from-scratch transformer training pipeline capable of running both forward and backward passes on ANE hardware without relying on CoreML, Metal, or GPU acceleration. It explores the internal software stack of the Apple Neural Engine by...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 11
    NeuralNote

    NeuralNote

    Audio Plugin for Audio to MIDI transcription using deep learning

    ...The software functions as an audio plugin that can be used inside digital audio workstations as well as a standalone application for music production and analysis. Its main purpose is to perform audio-to-MIDI transcription, allowing musicians to record a performance and automatically transform it into editable MIDI notes. NeuralNote supports polyphonic transcription, meaning it can detect multiple notes played simultaneously, making it useful for instruments such as piano or guitar. The system relies on neural network models to analyze audio signals and infer pitch, timing, and other musical attributes that can be represented as MIDI data. ...
    Downloads: 37 This Week
    Last Update:
    See Project
  • 12
    Learn PyTorch for Deep Learning

    Learn PyTorch for Deep Learning

    Materials for the Learn PyTorch for Deep Learning

    ...The materials include Jupyter notebooks, explanations of core deep learning concepts, and step-by-step demonstrations of building and training neural networks. Throughout the lessons, users learn how to work with tensors, create neural network architectures, manage training workflows, and evaluate model performance.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 13
    Opacus

    Opacus

    Training PyTorch models with differential privacy

    Opacus is a library that enables training PyTorch models with differential privacy. It supports training with minimal code changes required on the client, has little impact on training performance, and allows the client to online track the privacy budget expended at any given moment. Vectorized per-sample gradient computation that is 10x faster than micro batching. Supports most types of PyTorch models and can be used with minimal modification to the original neural network. Open source, modular API for differential privacy research. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    TensorFlow Model Optimization Toolkit

    TensorFlow Model Optimization Toolkit

    A toolkit to optimize ML models for deployment for Keras & TensorFlow

    The TensorFlow Model Optimization Toolkit is a suite of tools for optimizing ML models for deployment and execution. Among many uses, the toolkit supports techniques used to reduce latency and inference costs for cloud and edge devices (e.g. mobile, IoT). Deploy models to edge devices with restrictions on processing, memory, power consumption, network usage, and model storage space. Enable execution on and optimize for existing hardware or new special purpose accelerators. Choose the model...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    HPCC Systems

    HPCC Systems

    End-to-end big data in a massively scalable supercomputing platform.

    HPCC Systems® (www.hpccsystems.com) from LexisNexis® Risk Solutions is a proven, open source solution for Big Data insights that can be implemented by businesses of all sizes. With HPCC Systems, developers can design applications with Big Data at their core, enabling businesses to better analyze and understand data at scale, improving business time to results and decisions. HPCC Systems offers a consistent data-centric programming language, two processing platforms and a single, complete...
    Downloads: 9 This Week
    Last Update:
    See Project
  • 16
    picoGPT

    picoGPT

    An unnecessarily tiny implementation of GPT-2 in NumPy

    picoGPT is a minimal implementation of the GPT-2 language model designed to demonstrate how transformer-based language models work at a conceptual level. The repository focuses on educational clarity rather than production performance, implementing the core components of the GPT architecture in a concise and readable way. It allows users to understand how tokenization, transformer layers, attention mechanisms, and autoregressive text generation operate in modern large language models. The project uses a small amount of code to illustrate the essential mathematical operations involved in training and running a transformer-based neural network.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 17
    T81 558

    T81 558

    Applications of Deep Neural Networks

    Deep learning is a group of exciting new technologies for neural networks. Through a combination of advanced training techniques and neural network architectural components, it is now possible to create neural networks that can handle tabular data, images, text, and audio as both input and output. Deep learning allows a neural network to learn hierarchies of information in a way that is like the function of the human brain. This course will introduce the student to classic neural network structures, Convolution Neural Networks (CNN), Long Short-Term Memory (LSTM), Gated Recurrent Neural Networks (GRU), General Adversarial Networks (GAN) and reinforcement learning. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    LSTMs for Human Activity Recognition

    LSTMs for Human Activity Recognition

    Human Activity Recognition example using TensorFlow on smartphone

    LSTM-Human-Activity-Recognition is a machine learning project that demonstrates how recurrent neural networks can be used to recognize human activities from sensor data. The repository implements a deep learning model based on Long Short-Term Memory (LSTM) networks to classify physical activities using time-series data collected from wearable sensors. The project uses the well-known Human Activity Recognition dataset derived from smartphone accelerometer and gyroscope signals. Through the...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 19
    nlpaug

    nlpaug

    Data augmentation for NLP

    This Python library helps you with augmenting nlp for your machine learning projects. Visit this introduction to understand Data Augmentation in NLP. Augmenter is the basic element of augmentation while Flow is a pipeline to orchestra multi augmenters together.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    YOLOv4-large

    YOLOv4-large

    Scaled-YOLOv4: Scaling Cross Stage Partial Network

    ...The project provides a PyTorch implementation of the Scaled-YOLOv4 framework, which extends the original YOLOv4 architecture using Cross Stage Partial (CSP) networks and new scaling techniques. Unlike earlier object detection systems that only scale depth or width, this architecture scales multiple aspects of the neural network including structure, resolution, and channel configuration. This scaling strategy enables the model to adapt to different hardware environments while maintaining a strong balance between speed and detection accuracy. The repository includes multiple model variants such as YOLOv4-tiny, YOLOv4-CSP, and large-scale configurations designed for high-performance detection tasks.
    Downloads: 10 This Week
    Last Update:
    See Project
  • 21
    TNN

    TNN

    Uniform deep learning inference framework for mobile

    TNN, a high-performance, lightweight neural network inference framework open sourced by Tencent Youtu Lab. It also has many outstanding advantages such as cross-platform, high performance, model compression, and code tailoring. The TNN framework further strengthens the support and performance optimization of mobile devices on the basis of the original Rapidnet and ncnn frameworks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    CapsGNN

    CapsGNN

    A PyTorch implementation of "Capsule Graph Neural Network"

    A PyTorch implementation of "Capsule Graph Neural Network" (ICLR 2019). The high-quality node embeddings learned from the Graph Neural Networks (GNNs) have been applied to a wide range of node-based applications and some of them have achieved state-of-the-art (SOTA) performance. However, when applying node embeddings learned from GNNs to generate graph embeddings, the scalar node representation may not suffice to preserve the node/graph properties efficiently, resulting in sub-optimal graph embeddings. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    Effective TensorFlow 2

    Effective TensorFlow 2

    TensorFlow tutorials and best practices

    ...The project focuses on helping developers write efficient, maintainable, and reliable TensorFlow code when building deep learning systems. It includes practical guidelines that explain common pitfalls in neural network training, such as numerical instability and gradient-related issues. The repository also demonstrates techniques for improving model performance, optimizing training loops, and debugging TensorFlow programs. Through examples and explanations, the project highlights how developers can structure machine learning code to improve readability and maintainability. ...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 24
    EfficientNet Keras

    EfficientNet Keras

    Implementation of EfficientNet model. Keras and TensorFlow Keras

    ...In this paper, we systematically study model scaling and identify that carefully balancing network depth, width, and resolution can lead to better performance. Based on this observation, we propose a new scaling method that uniformly scales all dimensions of depth/width/resolution using a simple yet highly effective compound coefficient. We demonstrate the effectiveness of this method on scaling up MobileNets and ResNet.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    NLP-Models-Tensorflow

    NLP-Models-Tensorflow

    Gathers machine learning and Tensorflow deep learning models for NLP

    ...The project includes scripts for preparing datasets, training models, and evaluating performance on various text analysis tasks. Many implementations are designed for experimentation, allowing developers to adjust parameters, swap architectures, and test different preprocessing techniques.
    Downloads: 2 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • Next
MongoDB Logo MongoDB