Luminal
Luminal is a machine-learning framework built for speed, simplicity, and composability, focusing on static graphs and compiler-based optimization to deliver high performance even for complex neural networks. It compiles models into minimal “primops” (only 12 primitive operations) and then applies compiler passes to replace those with device-specific optimized kernels, enabling efficient execution on GPU or other backends. It supports modules (building blocks of networks with a standard forward API) and the GraphTensor interface (typed tensors and graphs at compile time) for model definition and execution. Luminal’s core remains intentionally small and hackable, with extensibility via external compilers for datatypes, devices, training, quantization, and more. Quick-start guidance shows how to clone the repo, build a “Hello World” example, or run a larger model like LLaMA 3 using GPU features.
Learn more
Zebra by Mipsology
Zebra by Mipsology is the ideal Deep Learning compute engine for neural network inference. Zebra seamlessly replaces or complements CPUs/GPUs, allowing any neural network to compute faster, with lower power consumption, at a lower cost. Zebra deploys swiftly, seamlessly, and painlessly without knowledge of underlying hardware technology, use of specific compilation tools, or changes to the neural network, the training, the framework, and the application. Zebra computes neural networks at world-class speed, setting a new standard for performance. Zebra runs on highest-throughput boards all the way to the smallest boards. The scaling provides the required throughput, in data centers, at the edge, or in the cloud. Zebra accelerates any neural network, including user-defined neural networks. Zebra processes the same CPU/GPU-based trained neural network with the same accuracy without any change.
Learn more
DeePhi Quantization Tool
This is a model quantization tool for convolution neural networks(CNN). This tool could quantize both weights/biases and activations from 32-bit floating-point (FP32) format to 8-bit integer(INT8) format or any other bit depths. With this tool, you can boost the inference performance and efficiency significantly, while maintaining the accuracy. This tool supports common layer types in neural networks, including convolution, pooling, fully-connected, batch normalization and so on. The quantization tool does not need the retraining of the network or labeled datasets, only one batch of pictures are needed. The process time ranges from a few seconds to several minutes depending on the size of neural network, which makes rapid model update possible. This tool is collaborative optimized for DeePhi DPU and could generate INT8 format model files required by DNNC.
Learn more
ConvNetJS
ConvNetJS is a Javascript library for training deep learning models (neural networks) entirely in your browser. Open a tab and you're training. No software requirements, no compilers, no installations, no GPUs, no sweat. The library allows you to formulate and solve neural networks in Javascript, and was originally written by @karpathy. However, the library has since been extended by contributions from the community and more are warmly welcome. The fastest way to obtain the library in a plug-and-play way if you don't care about developing is through this link to convnet-min.js, which contains the minified library. Alternatively, you can also choose to download the latest release of the library from Github. The file you are probably most interested in is build/convnet-min.js, which contains the entire library. To use it, create a bare-bones index.html file in some folder and copy build/convnet-min.js to the same folder.
Learn more