Showing 215 open source projects for "computation c#"

View related business solutions
  • Enterprise-grade ITSM, for every business Icon
    Enterprise-grade ITSM, for every business

    Give your IT, operations, and business teams the ability to deliver exceptional services—without the complexity.

    Freshservice is an intuitive, AI-powered platform that helps IT, operations, and business teams deliver exceptional service without the usual complexity. Automate repetitive tasks, resolve issues faster, and provide seamless support across the organization. From managing incidents and assets to driving smarter decisions, Freshservice makes it easy to stay efficient and scale with confidence.
    Try it Free
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 1
    ggml

    ggml

    Tensor library for machine learning

    ggml is an open-source tensor library designed for efficient machine learning computation with a focus on running models locally and with minimal dependencies. Written primarily in C and C++, the library provides low-level tensor operations and automatic differentiation that allow developers to implement machine learning algorithms and neural networks efficiently. The project emphasizes portability and performance, enabling machine learning inference across a wide range of hardware environments including CPUs and specialized accelerators. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 2
    Qulacs

    Qulacs

    Variational Quantum Circuit Simulator for Quantum Computation Research

    Variational Quantum Circuit Simulator for Quantum Computation Research. Qulacs is a Python/C++ library for fast simulation of large, noisy, or parametric quantum circuits. Qulacs is developed at QunaSys, Osaka University, NTT, and Fujitsu.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 3
    Cactus

    Cactus

    Low-latency AI inference engine optimized for mobile devices

    Cactus is a low-latency, energy-efficient AI inference framework designed specifically for mobile devices and wearables, enabling advanced machine learning capabilities directly on-device. It provides a full-stack architecture composed of an inference engine, a computation graph system, and highly optimized hardware kernels tailored for ARM-based processors. Cactus emphasizes efficient memory usage through techniques such as zero-copy computation graphs and quantized model formats, allowing...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    MuJoCo

    MuJoCo

    Multi-Joint dynamics with Contact. A general purpose physics simulator

    ...It is widely used in research fields such as robotics, biomechanics, computer graphics, animation, and machine learning, where fast and accurate physics simulations are essential. The engine provides a robust C API optimized for real-time computation, making it suitable for scientific research and advanced simulation environments. MuJoCo’s core architecture is performance-tuned and utilizes preallocated data structures created through an XML-based compiler. The platform includes built-in interactive visualization using OpenGL and a native graphical interface for analyzing and testing simulations. ...
    Downloads: 6 This Week
    Last Update:
    See Project
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 5

    C-BerryTrans

    Calculation of Berry-curvature-driven AHC or ANC using WIEN2k

    C-BerryTrans is a C++ based code designed for first-principles calculation of Berry-curvature-driven anomalous transverse transport properties, namely the anomalous Hall conductivity (AHC) and anomalous Nernst conductivity (ANC). The code directly extracts eigenvalues and momentum matrix elements from the WIEN2k package and evaluates the Berry curvature using a Kubo-like formalism, thereby avoiding interpolation errors associated with Wannier-based methods. To ensure efficiency, the code...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 6
    Evolutionary Computation Framework

    Evolutionary Computation Framework

    C++ framework for application of any type of evolutionary computation.

    ECF is a framework intended for application of any type of evolutionary computation (GA/GP, DE, Clonalg, ES, PSO, ABC, GAn, local search...). It offers simplicity for the end-user (parameterless usage, tutorial) and customization for experienced EC practicioners.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7

    Halide

    A language for fast, portable data-parallel computation

    Halide is a programming language for fast, portable data-parallel computation. It was designed to make writing high-performance image and array processing code much easier on modern machines. It works on all major operating systems and with several CPU architectures (X86, ARM, MIPS, Hexagon, PowerPC) and GPU Compute APIs (CUDA, OpenCL, OpenGL, among others). It isn't a standalone programming language however; rather it is embedded in C++ which means that you write C++ code, building an in-memory representation of a Halide pipeline using Halide's C++ API. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 8
    PyTorch

    PyTorch

    Open source machine learning framework

    PyTorch is a Python package that offers Tensor computation (like NumPy) with strong GPU acceleration and deep neural networks built on tape-based autograd system. This project allows for fast, flexible experimentation and efficient production. PyTorch consists of torch (Tensor library), torch.autograd (tape-based automatic differentiation library), torch.jit (a compilation stack [TorchScript]), torch.nn (neural networks library), torch.multiprocessing (Python multiprocessing), and...
    Downloads: 124 This Week
    Last Update:
    See Project
  • 9
    Flash-MoE

    Flash-MoE

    Running a big model on a small laptop

    Flash-MoE is a high-performance implementation of mixture-of-experts (MoE) architectures designed to optimize the efficiency and scalability of large AI models. It focuses on accelerating routing and computation by leveraging optimized kernels and memory management techniques, allowing models to dynamically select specialized sub-networks during inference. The project aims to reduce the computational cost typically associated with MoE systems while maintaining or improving performance. It...
    Downloads: 1 This Week
    Last Update:
    See Project
  • Go From AI Idea to AI App Fast Icon
    Go From AI Idea to AI App Fast

    One platform to build, fine-tune, and deploy ML models. No MLOps team required.

    Access Gemini 3 and 200+ models. Build chatbots, agents, or custom models with built-in monitoring and scaling.
    Try Free
  • 10
    Shumai

    Shumai

    Fast Differentiable Tensor Library in JavaScript & TypeScript with Bun

    Shumai is an experimental differentiable tensor library for TypeScript and JavaScript, developed by Facebook Research. It provides a high-performance framework for numerical computing and machine learning within modern JavaScript runtimes. Built on Bun and Flashlight, with ArrayFire as its numerical backend, Shumai brings GPU-accelerated tensor operations, automatic differentiation, and scientific computing tools directly to JavaScript developers. It allows seamless integration of machine...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    Hasktorch

    Hasktorch

    Tensors and neural networks in Haskell

    ...Contributions/PR are encouraged. Hasktorch is a library for tensors and neural networks in Haskell. It is an independent open source community project which leverages the core C++ libraries shared by PyTorch.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    S2 Geometry

    S2 Geometry

    Computational geometry and spatial indexing on the sphere

    s2geometry is Google’s open source geometry library designed for representing, analyzing, and manipulating geometric shapes on a sphere rather than a flat plane. This makes it particularly suited for applications involving geospatial data, such as mapping, spatial indexing, and geographic information systems (GIS). The library provides a robust mathematical framework for spherical geometry, allowing developers to work with polygons, points, and regions on the Earth’s surface using consistent...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 13
    Symbolics.jl

    Symbolics.jl

    Symbolic programming for the next generation of numerical software

    Symbolics.jl is a high-performance symbolic computation library for the Julia programming language. It enables users to define, manipulate, and analyze mathematical expressions symbolically, with strong support for symbolic differentiation, simplification, equation solving, and code generation. Designed for use in scientific computing, machine learning, and engineering, Symbolics.jl integrates smoothly with Julia’s numerical ecosystem, allowing symbolic expressions to be compiled and...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 14

    Microsoft SEAL

    Easy-to-use and powerful homomorphic encryption library

    Microsoft SEAL is an open-source homomorphic encryption library that allows computations to be performed directly on encrypted data, with no need for any decryption. Developed by the Cryptography and Privacy Research group at Microsoft, it enables software engineers to build end-to-end encrypted data storage and computation services that never have to procure the customer's key. Microsoft SEAL is very easy to use, compile and run in many different environments. Homomorphic encryption is...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 15
    ONNX

    ONNX

    Open standard for machine learning interoperability

    ONNX is an open format built to represent machine learning models. ONNX defines a common set of operators - the building blocks of machine learning and deep learning models - and a common file format to enable AI developers to use models with a variety of frameworks, tools, runtimes, and compilers. Open Neural Network Exchange (ONNX) is an open ecosystem that empowers AI developers to choose the right tools as their project evolves. ONNX provides an open source format for AI models, both...
    Downloads: 9 This Week
    Last Update:
    See Project
  • 16

    C++ Eigenvectors

    C++ matrix class template with eigenvectors

    All code is in a single .h file except for a .cpp file with the main() demo. Computation time scales as N^4. Also has class template Polynomial. Has extensive operators and methods. Algorithm summary: Any polynomial of degree m is uniquely and simply determined by its values at m+1 points. In particular the characteristic polynomial of an NxN matrix A can be determined by evaluating the determinant |x * I - A| at N+1 distinct values of x. Householder reduction is used to find ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    collapse

    collapse

    Advanced and Fast Data Transformation in R

    collapse is a high-performance R package designed for fast and efficient data transformation, aggregation, reshaping, and statistical computation. Built to offer a more performant alternative to dplyr and data.table, it is particularly well-suited for large datasets and econometric applications. It operates on base R data structures like data frames and vectors and uses highly optimized C++ code under the hood to deliver significant speed improvements. collapse also includes tools for grouped operations, weighted statistics, and time series manipulation, making it a compact yet powerful utility for data scientists and researchers working in R.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    CTranslate2

    CTranslate2

    Fast inference engine for Transformer models

    ...The model serialization and computation support weights with reduced precision: 16-bit floating points (FP16), 16-bit integers (INT16), and 8-bit integers (INT8). The project supports x86-64 and AArch64/ARM64 processors and integrates multiple backends that are optimized for these platforms: Intel MKL, oneDNN, OpenBLAS, Ruy, and Apple Accelerate.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 19
    Bolt NLP

    Bolt NLP

    Bolt is a deep learning library with high performance

    Bolt is a high-performance deep learning inference framework developed by Huawei Noah's Ark Lab. It is designed to optimize and accelerate the deployment of deep learning models across various hardware platforms. Bolt is a light-weight library for deep learning. Bolt, as a universal deployment tool for all kinds of neural networks, aims to automate the deployment pipeline and achieve extreme acceleration. Bolt has been widely deployed and used in many departments of HUAWEI company, such as...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 20
    Double Conversion

    Double Conversion

    Efficient binary-decimal & decimal-binary conversion routines for IEEE

    ...The library ensures consistent and accurate results for converting between double values and their string representations, avoiding rounding errors and performance bottlenecks common in standard conversion routines. It is optimized for both speed and correctness, making it ideal for numerical computation libraries, serialization systems, and scripting engines. The codebase includes detailed documentation and comprehensive unit tests to validate correctness across various platforms. With flexible build options using SCons, CMake, or Bazel, Double Conversion integrates seamlessly into modern C++ development workflows.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    RunAnywhere

    RunAnywhere

    Production ready toolkit to run AI locally

    RunAnywhere SDKs are a set of cross-platform development tools that enable applications to run artificial intelligence models directly on user devices instead of relying on cloud infrastructure. The toolkit allows developers to integrate language models, speech recognition, and voice synthesis capabilities into mobile or desktop applications while keeping all computation local. By running models entirely on device, the platform eliminates network latency and protects user data because...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 22
    Apache Doris

    Apache Doris

    MPP-based interactive SQL data warehousing for reporting and analysis

    Apache Doris is a modern MPP analytical database product. It can provide sub-second queries and efficient real-time data analysis. With it's distributed architecture, up to 10PB level datasets will be well supported and easy to operate. Apache Doris can meet various data analysis demands, including history data reports, real-time data analysis, interactive data analysis, and exploratory data analysis. Make your data analysis easier! Support standard SQL language, compatible with MySQL...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 23
    PicoLM

    PicoLM

    Run a 1-billion parameter LLM on a $10 board with 256MB RAM

    PicoLM is an open-source inference framework designed to run large language models on extremely constrained hardware environments such as inexpensive single-board computers and embedded systems. The project focuses on enabling efficient local inference by optimizing memory usage, computation, and system dependencies so that relatively large models can operate on devices with minimal RAM. It is written primarily in C and designed with a minimalist architecture that removes unnecessary dependencies and external libraries. The runtime is capable of running language models with billions of parameters on devices with only a few hundred megabytes of memory, which is significantly lower than typical LLM infrastructure requirements. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    VibeTensor

    VibeTensor

    Our first fully AI generated deep learning system

    VibeTensor is a groundbreaking open-source research system software stack for deep learning that was uniquely generated almost entirely by AI coding agents under guided human supervision, demonstrating a new frontier in AI-assisted software engineering. It implements a PyTorch-style eager tensor library with a modern C++20 core that supports both CPU and CUDA backends, giving it the ability to manage tensors, automatic differentiation (autograd), and complex computation flows similar to mainstream frameworks. What makes VibeTensor remarkable is that every major component, from core libraries and dispatch systems to CUDA runtime support, caching allocators, and language bindings, was created and validated by coding agents using automated builds and tests rather than manual line-by-line human coding. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    ImplicitGlobalGrid.jl

    ImplicitGlobalGrid.jl

    Distributed parallelization of stencil-based GPU and CPU applications

    ...ImplicitGlobalGrid relies on the Julia MPI wrapper (MPI.jl) to perform halo updates close to hardware limit and leverages CUDA-aware or ROCm-aware MPI for GPU-applications. The communication can straightforwardly be hidden behind computation [1, 3] (how this can be done automatically when using ParallelStencil.jl is shown in; a general approach particularly suited for CUDA C applications is explained in.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • 4
  • 5
  • Next
MongoDB Logo MongoDB