Showing 20 open source projects for "network performance"

View related business solutions
  • Our Free Plans just got better! | Auth0 Icon
    Our Free Plans just got better! | Auth0

    With up to 25k MAUs and unlimited Okta connections, our Free Plan lets you focus on what you do best—building great apps.

    You asked, we delivered! Auth0 is excited to expand our Free and Paid plans to include more options so you can focus on building, deploying, and scaling applications without having to worry about your security. Auth0 now, thank yourself later.
    Try free now
  • 8 Monitoring Tools in One APM. Install in 5 Minutes. Icon
    8 Monitoring Tools in One APM. Install in 5 Minutes.

    Errors, performance, logs, uptime, hosts, anomalies, dashboards, and check-ins. One interface.

    AppSignal works out of the box for Ruby, Elixir, Node.js, Python, and more. 30-day free trial, no credit card required.
    Start Free
  • 1
    imgclsmob Deep learning networks

    imgclsmob Deep learning networks

    Sandbox for training deep learning networks

    ...The project is frequently used by developers who want to study modern convolutional neural network designs and compare their performance across datasets.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    face.evoLVe

    face.evoLVe

    High-Performance Face Recognition Library on PaddlePaddle & PyTorch

    ...The repository supports multiple neural network backbones such as ResNet, DenseNet, MobileNet, and ShuffleNet, enabling experimentation with different architectures depending on performance requirements. It also implements a wide range of loss functions commonly used in face recognition research, including ArcFace, CosFace, Triplet loss, and Softmax variants. To improve scalability, the library introduces distributed training techniques that allow large models to be trained efficiently across multiple GPUs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Intel Extension for PyTorch

    Intel Extension for PyTorch

    A Python package for extending the official PyTorch

    Intel® Extension for PyTorch* extends PyTorch* with up-to-date features optimizations for an extra performance boost on Intel hardware. Optimizations take advantage of Intel® Advanced Vector Extensions 512 (Intel® AVX-512) Vector Neural Network Instructions (VNNI) and Intel® Advanced Matrix Extensions (Intel® AMX) on Intel CPUs as well as Intel Xe Matrix Extensions (XMX) AI engines on Intel discrete GPUs. Moreover, Intel® Extension for PyTorch* provides easy GPU acceleration for Intel discrete GPUs through the PyTorch* xpu device.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 4
    Opacus

    Opacus

    Training PyTorch models with differential privacy

    Opacus is a library that enables training PyTorch models with differential privacy. It supports training with minimal code changes required on the client, has little impact on training performance, and allows the client to online track the privacy budget expended at any given moment. Vectorized per-sample gradient computation that is 10x faster than micro batching. Supports most types of PyTorch models and can be used with minimal modification to the original neural network. Open source, modular API for differential privacy research. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 5
    TensorFlow Quantum

    TensorFlow Quantum

    Open-source Python framework for hybrid quantum-classical ml learning

    TensorFlow Quantum is an open-source software framework designed for building and training hybrid quantum-classical machine learning models within the TensorFlow ecosystem. The framework enables researchers and developers to represent quantum circuits as data and integrate them directly into machine learning workflows. By combining classical deep learning techniques with quantum algorithms, the platform allows experimentation with quantum machine learning methods that may offer advantages...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    Core ML Tools

    Core ML Tools

    Core ML tools contain supporting tools for Core ML model conversion

    ...Core ML provides a unified representation for all models. Your app uses Core ML APIs and user data to make predictions, and to fine-tune models, all on the user’s device. Core ML optimizes on-device performance by leveraging the CPU, GPU, and Neural Engine while minimizing its memory footprint and power consumption. Running a model strictly on the user’s device removes any need for a network connection, which helps keep the user’s data private and your app responsive.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    TensorFlow Model Optimization Toolkit

    TensorFlow Model Optimization Toolkit

    A toolkit to optimize ML models for deployment for Keras & TensorFlow

    The TensorFlow Model Optimization Toolkit is a suite of tools for optimizing ML models for deployment and execution. Among many uses, the toolkit supports techniques used to reduce latency and inference costs for cloud and edge devices (e.g. mobile, IoT). Deploy models to edge devices with restrictions on processing, memory, power consumption, network usage, and model storage space. Enable execution on and optimize for existing hardware or new special purpose accelerators. Choose the model...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    picoGPT

    picoGPT

    An unnecessarily tiny implementation of GPT-2 in NumPy

    picoGPT is a minimal implementation of the GPT-2 language model designed to demonstrate how transformer-based language models work at a conceptual level. The repository focuses on educational clarity rather than production performance, implementing the core components of the GPT architecture in a concise and readable way. It allows users to understand how tokenization, transformer layers, attention mechanisms, and autoregressive text generation operate in modern large language models. The project uses a small amount of code to illustrate the essential mathematical operations involved in training and running a transformer-based neural network.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    LSTMs for Human Activity Recognition

    LSTMs for Human Activity Recognition

    Human Activity Recognition example using TensorFlow on smartphone

    LSTM-Human-Activity-Recognition is a machine learning project that demonstrates how recurrent neural networks can be used to recognize human activities from sensor data. The repository implements a deep learning model based on Long Short-Term Memory (LSTM) networks to classify physical activities using time-series data collected from wearable sensors. The project uses the well-known Human Activity Recognition dataset derived from smartphone accelerometer and gyroscope signals. Through the...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 10
    nlpaug

    nlpaug

    Data augmentation for NLP

    This Python library helps you with augmenting nlp for your machine learning projects. Visit this introduction to understand Data Augmentation in NLP. Augmenter is the basic element of augmentation while Flow is a pipeline to orchestra multi augmenters together.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    YOLOv4-large

    YOLOv4-large

    Scaled-YOLOv4: Scaling Cross Stage Partial Network

    ...The project provides a PyTorch implementation of the Scaled-YOLOv4 framework, which extends the original YOLOv4 architecture using Cross Stage Partial (CSP) networks and new scaling techniques. Unlike earlier object detection systems that only scale depth or width, this architecture scales multiple aspects of the neural network including structure, resolution, and channel configuration. This scaling strategy enables the model to adapt to different hardware environments while maintaining a strong balance between speed and detection accuracy. The repository includes multiple model variants such as YOLOv4-tiny, YOLOv4-CSP, and large-scale configurations designed for high-performance detection tasks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    CapsGNN

    CapsGNN

    A PyTorch implementation of "Capsule Graph Neural Network"

    A PyTorch implementation of "Capsule Graph Neural Network" (ICLR 2019). The high-quality node embeddings learned from the Graph Neural Networks (GNNs) have been applied to a wide range of node-based applications and some of them have achieved state-of-the-art (SOTA) performance. However, when applying node embeddings learned from GNNs to generate graph embeddings, the scalar node representation may not suffice to preserve the node/graph properties efficiently, resulting in sub-optimal graph embeddings. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    EfficientNet Keras

    EfficientNet Keras

    Implementation of EfficientNet model. Keras and TensorFlow Keras

    ...In this paper, we systematically study model scaling and identify that carefully balancing network depth, width, and resolution can lead to better performance. Based on this observation, we propose a new scaling method that uniformly scales all dimensions of depth/width/resolution using a simple yet highly effective compound coefficient. We demonstrate the effectiveness of this method on scaling up MobileNets and ResNet.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    NLP-Models-Tensorflow

    NLP-Models-Tensorflow

    Gathers machine learning and Tensorflow deep learning models for NLP

    ...The project includes scripts for preparing datasets, training models, and evaluating performance on various text analysis tasks. Many implementations are designed for experimentation, allowing developers to adjust parameters, swap architectures, and test different preprocessing techniques.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    BytePS

    BytePS

    A high performance and generic framework for distributed DNN training

    ...We use Tesla V100 32GB GPUs and set batch size equal to 64 per GPU. Each machine has 8 V100 GPUs (32GB memory) with NVLink-enabled. Machines are inter-connected with 100 Gbps RDMA network. This is the same hardware setup you can get on AWS.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    NeuroNER

    NeuroNER

    Named-entity recognition using neural networks

    ..."deep learning") Is cross-platform, open source, freely available, and straightforward to use. Enables the users to create or modify annotations for a new or existing corpus. Train the neural network that performs the NER. During the training, NeuroNER allows monitoring of the network. Evaluate the quality of the predictions made by NeuroNER. The performance metrics can be calculated and plotted by comparing the predicted labels with the gold labels.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Easy-TensorFlow

    Easy-TensorFlow

    Simple and comprehensive tutorials in TensorFlow

    ...Furthermore, since most of the developers are using TensorFlow for code development, having hands-on on TensorFlow is a necessity these days. Tensorboard is a powerful visualization suite that is developed to track both the network topology and performance, making debugging even simpler.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    Tensorpack

    Tensorpack

    A Neural Net Training Interface on TensorFlow, with focus on speed

    ...Squeeze the best data loading performance of Python with tensorpack.dataflow. Symbolic programming (e.g. tf.data) does not offer the data processing flexibility needed in research. Tensorpack squeezes the most performance out of pure Python with various auto parallelization strategies. There are too many symbolic function wrappers already. Tensorpack includes only a few common layers.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    DIGITS

    DIGITS

    Deep Learning GPU training system

    The NVIDIA Deep Learning GPU Training System (DIGITS) puts the power of deep learning into the hands of engineers and data scientists. DIGITS can be used to rapidly train the highly accurate deep neural network (DNNs) for image classification, segmentation and object detection tasks. DIGITS simplifies common deep learning tasks such as managing data, designing and training neural networks on multi-GPU systems, monitoring performance in real-time with advanced visualizations, and selecting the best performing model from the results browser for deployment. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    BitNet

    BitNet

    BitNet: Scaling 1-bit Transformers for Large Language Models

    BitNet is a machine learning research implementation that explores extremely low-precision neural network architectures designed to dramatically reduce the computational cost of large language models. The project implements the BitNet architecture described in research on scaling transformer models using extremely low-bit quantization techniques. In this approach, neural network weights are quantized to approximately one bit per parameter, allowing models to operate with far lower memory usage than traditional 16-bit or 32-bit neural networks. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB