C++ Neural Network Libraries

View 51 business solutions

Browse free open source C++ Neural Network Libraries and projects below. Use the toggles on the left to filter open source C++ Neural Network Libraries by OS, license, language, programming language, and project status.

  • Forever Free Full-Stack Observability | Grafana Cloud Icon
    Forever Free Full-Stack Observability | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • 1
    TensorRT

    TensorRT

    C++ library for high performance inference on NVIDIA GPUs

    NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference. It includes a deep learning inference optimizer and runtime that delivers low latency and high throughput for deep learning inference applications. TensorRT-based applications perform up to 40X faster than CPU-only platforms during inference. With TensorRT, you can optimize neural network models trained in all major frameworks, calibrate for lower precision with high accuracy, and deploy to hyperscale data centers, embedded, or automotive product platforms. TensorRT is built on CUDA®, NVIDIA’s parallel programming model, and enables you to optimize inference leveraging libraries, development tools, and technologies in CUDA-X™ for artificial intelligence, autonomous machines, high-performance computing, and graphics. With new NVIDIA Ampere Architecture GPUs, TensorRT also leverages sparse tensor cores providing an additional performance boost.
    Downloads: 22 This Week
    Last Update:
    See Project
  • 2
    ncnn

    ncnn

    High-performance neural network inference framework for mobile

    ncnn is a high-performance neural network inference computing framework designed specifically for mobile platforms. It brings artificial intelligence right at your fingertips with no third-party dependencies, and speeds faster than all other known open source frameworks for mobile phone cpu. ncnn allows developers to easily deploy deep learning algorithm models to the mobile platform and create intelligent APPs. It is cross-platform and supports most commonly used CNN networks, including Classical CNN (VGG AlexNet GoogleNet Inception), Face Detection (MTCNN RetinaFace), Segmentation (FCN PSPNet UNet YOLACT), and more. ncnn is currently being used in a number of Tencent applications, namely: QQ, Qzone, WeChat, and Pitu.
    Downloads: 17 This Week
    Last Update:
    See Project
  • 3
    Fast Artificial Neural Network Library is a free open source neural network library, which implements multilayer artificial neural networks in C with support for both fully connected and sparsely connected networks. Cross-platform execution in both fixed and floating point are supported. It includes a framework for easy handling of training data sets. It is easy to use, versatile, well documented, and fast. Bindings to more than 15 programming languages are available. An easy to read introduction article and a reference manual accompanies the library with examples and recommendations on how to use the library. Several graphical user interfaces are also available for the library.
    Downloads: 40 This Week
    Last Update:
    See Project
  • 4
    XNNPACK

    XNNPACK

    High-efficiency floating-point neural network inference operators

    XNNPACK is a highly optimized, low-level neural network inference library developed by Google for accelerating deep learning workloads across a variety of hardware architectures, including ARM, x86, WebAssembly, and RISC-V. Rather than serving as a standalone ML framework, XNNPACK provides high-performance computational primitives—such as convolutions, pooling, activation functions, and arithmetic operations—that are integrated into higher-level frameworks like TensorFlow Lite, PyTorch Mobile, ONNX Runtime, TensorFlow.js, and MediaPipe. The library is written in C/C++ and designed for maximum portability, efficiency, and performance, leveraging platform-specific instruction sets (e.g., NEON, AVX, SIMD) for optimized execution. It supports NHWC tensor layouts and allows flexible striding along the channel dimension to efficiently handle channel-split and concatenation operations without additional cost.
    Downloads: 4 This Week
    Last Update:
    See Project
  • AI-powered service management for IT and enterprise teams Icon
    AI-powered service management for IT and enterprise teams

    Enterprise-grade ITSM, for every business

    Give your IT, operations, and business teams the ability to deliver exceptional services—without the complexity. Maximize operational efficiency with refreshingly simple, AI-powered Freshservice.
    Try it Free
  • 5
    Simd

    Simd

    High performance image processing library in C++

    The Simd Library is a free open source image processing library, designed for C and C++ programmers. It provides many useful high performance algorithms for image processing such as: pixel format conversion, image scaling and filtration, extraction of statistic information from images, motion detection, object detection (HAAR and LBP classifier cascades) and classification, neural network. The algorithms are optimized with using of different SIMD CPU extensions. In particular the library supports following CPU extensions: SSE, SSE2, SSE3, SSSE3, SSE4.1, SSE4.2, AVX, AVX2 and AVX-512 for x86/x64, VMX(Altivec) and VSX(Power7) for PowerPC, NEON for ARM. The Simd Library has C API and also contains useful C++ classes and functions to facilitate access to C API. The library supports dynamic and static linking, 32-bit and 64-bit Windows, Android and Linux, MSVS, G++ and Clang compilers, MSVS project and CMake build systems.
    Leader badge
    Downloads: 16 This Week
    Last Update:
    See Project
  • 6
    Tiny CUDA Neural Networks

    Tiny CUDA Neural Networks

    Lightning fast C++/CUDA neural network framework

    This is a small, self-contained framework for training and querying neural networks. Most notably, it contains a lightning-fast "fully fused" multi-layer perceptron (technical paper), a versatile multiresolution hash encoding (technical paper), as well as support for various other input encodings, losses, and optimizers. We provide a sample application where an image function (x,y) -> (R,G,B) is learned. The fully fused MLP component of this framework requires a very large amount of shared memory in its default configuration. It will likely only work on an RTX 3090, an RTX 2080 Ti, or high-end enterprise GPUs. Lower-end cards must reduce the n_neurons parameter or use the CutlassMLP (better compatibility but slower) instead. tiny-cuda-nn comes with a PyTorch extension that allows using the fast MLPs and input encodings from within a Python context. These bindings can be significantly faster than full Python implementations; in particular for the multiresolution hash encoding.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 7
    oneDNN

    oneDNN

    oneAPI Deep Neural Network Library (oneDNN)

    This software was previously known as Intel(R) Math Kernel Library for Deep Neural Networks (Intel(R) MKL-DNN) and Deep Neural Network Library (DNNL). oneAPI Deep Neural Network Library (oneDNN) is an open-source cross-platform performance library of basic building blocks for deep learning applications. oneDNN is part of oneAPI. The library is optimized for Intel(R) Architecture Processors, Intel Processor Graphics and Xe Architecture graphics. oneDNN has experimental support for the following architectures: Arm* 64-bit Architecture (AArch64), NVIDIA* GPU, OpenPOWER* Power ISA (PPC64), IBMz* (s390x), and RISC-V. oneDNN is intended for deep learning applications and framework developers interested in improving application performance on Intel CPUs and GPUs. Deep learning practitioners should use one of the applications enabled with oneDNN.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 8

    CURRENNT

    CUDA-enabled machine learning library for recurrent neural networks

    CURRENNT is a machine learning library for Recurrent Neural Networks (RNNs) which uses NVIDIA graphics cards to accelerate the computations. The library implements uni- and bidirectional Long Short-Term Memory (LSTM) architectures and supports deep networks as well as very large data sets that do not fit into main memory.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 9
    The DAMNED simulator is a Distributed And Multithreaded Neural Event-Driven simulation framework designed for Distributed Discrete Event Simulations, actually well suited for Large Scale Spiking Neural Network simulations.
    Downloads: 1 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 10
    It's an object-oriented library written in C++ for creating arbitrary kind of neural networks. The user can use the classes provided to create neural network with arbitrary topology and mixed type of neurons. It's very easy add custom neurons.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    Do you want a neural network OO class? With documentation? Do not go further... You found it! you have here a complete library of different neural network in an OO encapsulation. Starting from adaline, back propagation, Kohonen
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12

    ANNGD

    Artificial neural network gender detection

    ANNGD is a artificial neural network gender detection application. The detection is made in real time images captured by webcam by opencv library. The package include applications to image pre-processing and artificial neural network back-propagation training.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Amygdala is a C++ spiking neural network library. It includes several neuron models, SMP support and facilities for developing SNNs with genetic algorithms. Support for running Amygdala neural networks on workstation clusters and MPPs is also under way
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Kohonen neural network library is a set of classes and functions for design, train and use Kohonen network (self organizing map) which is one of AI algorithms and useful tool for data mining and discovery knowledge in data (http://knnl.sf.net).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    CNN for Image Retrieval
    cnn-for-image-retrieval is a research-oriented project that demonstrates the use of convolutional neural networks (CNNs) for image retrieval tasks. The repository provides implementations of CNN-based methods to extract feature representations from images and use them for similarity-based retrieval. It focuses on applying deep learning techniques to improve upon traditional handcrafted descriptors by learning features directly from data. The code includes training and evaluation scripts that can be adapted for custom datasets, making it useful for experimenting with retrieval systems in computer vision. By leveraging CNN architectures, the project showcases how learned embeddings can capture semantic similarity across varied images. This resource serves as both an educational reference and a foundation for further exploration in image retrieval research.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Cluster Networks are a new style of neural simulation / neural network modeling, that models networks of neural populations ("clusters") that transform and transmit information using precisely-timed, graded bursts ("pulses" or "volleys") of firing.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    DaNNet

    DaNNet

    Deep Artificial Neural Network framework using Armadillo

    DaNNet is a C++ deep neural network library using the Armadillo library as a base. It is intended to be a small and easy to use framework with no other dependencies than Armadillo. It uses independent layer-wise optimization giving you full flexibility to train your network.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    Euler

    Euler

    A distributed graph deep learning framework.

    As a general data structure with strong expressive ability, graphs can be used to describe many problems in the real world, such as user networks in social scenarios, user and commodity networks in e-commerce scenarios, communication networks in telecom scenarios, and transaction networks in financial scenarios. and drug molecule networks in medical scenarios, etc. Data in the fields of text, speech, and images is easier to process into a grid-like type of Euclidean space, which is suitable for processing by existing deep learning models. Graph is a data type in non-Euclidean space and cannot be directly applied to existing methods, requiring a specially designed graph neural network system. Graph-based learning methods such as graph neural networks combine end-to-end learning with inductive reasoning, and are expected to solve a series of problems such as relational reasoning and interpretability that deep learning cannot handle.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19

    FeedForwardNeuralNetworkC++

    Feedforward Neural Network writen in C++

    Feedforward Neural Network writen in C++ serial and parallelized in TBB library. Also using Autotune library for best parallel performance.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    Foad (EKG Processing)
    Foad is an open source software which receive an EKG Signal from scanner, WFDB database or heart sensors. Finding patient disease started by taking Fourier transform (FFT) from input signal and extract a single cycle. Based on some heuristic algorithm the most important feature like P , Q , R , S , T captured and feed to trained neural network. and so the final decision made by CNN library. As mentioned before this software also capable do some image processing on scanned paper to lower the final costs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    Inanna is an object oriented artificial neural network C++ library. Libraries are based on MagiClib base class library (included).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22

    LightSpeedANN

    Generator for optimized, vectorized neural net code

    This Ruby program takes in a topology specification for an artificial neural network and emits optimized C code (using SSE intrinsics) that implements fast forward and backward propagation for that specific topology.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    Lightweight backpropagation neural network in C++. The project provides a class implementing a feedforward neural network, and a class for easily train it. It is highly customizable to manage your problem and comes with a simple graphical interface.
    Leader badge
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    MCPN
    Multi-Core optimized Perceptron Network is a high-performance artificial neural network specially designed for workstations with multi-core CPUs, implemented as a shared library and coded in C++.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    MIVisionX

    MIVisionX

    Set of comprehensive computer vision & machine intelligence libraries

    MIVisionX toolkit is a set of comprehensive computer vision and machine intelligence libraries, utilities, and applications bundled into a single toolkit. AMD MIVisionX delivers highly optimized open-source implementation of the Khronos OpenVX™ and OpenVX™ Extensions along with Convolution Neural Net Model Compiler & Optimizer supporting ONNX, and Khronos NNEF™ exchange formats. The toolkit allows for rapid prototyping and deployment of optimized computer vision and machine learning inference workloads on a wide range of computer hardware, including small embedded x86 CPUs, APUs, discrete GPUs, and heterogeneous servers. AMD OpenVX is a highly optimized open-source implementation of the Khronos OpenVX™ 1.3 computer vision specification. It allows for rapid prototyping as well as fast execution on a wide range of computer hardware, including small embedded x86 CPUs and large workstation discrete GPUs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • Next
MongoDB Logo MongoDB