Search Results for "high performance computing"

Showing 2783 open source projects for "high performance computing"

View related business solutions
  • AI-generated apps that pass security review Icon
    AI-generated apps that pass security review

    Stop waiting on engineering. Build production-ready internal tools with AI—on your company data, in your cloud.

    Retool lets you generate dashboards, admin panels, and workflows directly on your data. Type something like “Build me a revenue dashboard on my Stripe data” and get a working app with security, permissions, and compliance built in from day one. Whether on our cloud or self-hosted, create the internal software your team needs without compromising enterprise standards or control.
    Try Retool free
  • Go From AI Idea to AI App Fast Icon
    Go From AI Idea to AI App Fast

    One platform to build, fine-tune, and deploy ML models. No MLOps team required.

    Access Gemini 3 and 200+ models. Build chatbots, agents, or custom models with built-in monitoring and scaling.
    Try Free
  • 1
    Minecraft High-Performance Redstone

    Minecraft High-Performance Redstone

    A multithreaded Minecraft server built for redstone

    A Minecraft 1.20.4 creative server built for Redstone. Each 256x256 plot runs on a separate thread, allowing for less lag, more concurrency, and many awesome extra features. MCHPRS is very different from traditional servers. Because this server is tailored to the use of computation redstone, many things that are a part of Vanilla Minecraft servers don't exist here. That being said, MCHPRS comes with many of its own unique features.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 2
    The Julia Programming Language

    The Julia Programming Language

    High-level, high-performance dynamic language for technical computing

    Julia is a fast, open source high-performance dynamic language for technical computing. It can be used for data visualization and plotting, deep learning, machine learning, scientific computing, parallel computing and so much more. Having a high level syntax, Julia is easy to use for programmers of every level and background. Julia has more than 2,800 community-registered packages including various mathematical libraries, data manipulation tools, and packages for general purpose computing. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 3
    Sogou C++ Workflow

    Sogou C++ Workflow

    C++ parallel computing and asynchronous networking engine

    As Sogou`s C++ server engine, Sogou C++ Workflow supports almost all back-end C++ online services of Sogou, including all search services, cloud input method, online advertisements, etc., handling more than 10 billion requests every day. This is an enterprise-level programming engine in light and elegant design which can satisfy most C++ back-end development requirements.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    NVIDIA Warp

    NVIDIA Warp

    A Python framework for accelerated simulation, data generation

    NVIDIA Warp is a high-performance Python framework developed by NVIDIA for building and accelerating simulation, graphics, and physics-based workloads using GPU computing. It enables developers to write kernel-level code in Python that is automatically compiled into efficient CUDA kernels, combining ease of use with near-native performance. The framework is designed for applications such as robotics, reinforcement learning, physical simulation, and differentiable computing, where performance and flexibility are critical. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 5
    LinDB

    LinDB

    LinDB is a scalable, high performance, high availability database

    LinDB is a scalable, high-performance, high-availability distributed time series database. A single server could easily support more than one million write TPS; With fundamental techniques like efficient compression storage and parallel computing, LinDB delivers highly optimized query performance. The multi-channel replication protocol supports any amount of nodes, and ensures the system's availability.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    TensorRT

    TensorRT

    C++ library for high performance inference on NVIDIA GPUs

    ...TensorRT is built on CUDA®, NVIDIA’s parallel programming model, and enables you to optimize inference leveraging libraries, development tools, and technologies in CUDA-X™ for artificial intelligence, autonomous machines, high-performance computing, and graphics. With new NVIDIA Ampere Architecture GPUs, TensorRT also leverages sparse tensor cores providing an additional performance boost.
    Downloads: 10 This Week
    Last Update:
    See Project
  • 7
    Quantum++

    Quantum++

    Modern C++ quantum computing library

    ...The main design factors taken in consideration were ease of use, high portability, and high performance. The library's simulation capabilities are only restricted by the amount of available physical memory. On a typical machine (Intel i5 8Gb RAM) Quantum++ can successfully simulate the evolution of 25 qubits in a pure state or of 12 qubits in a mixed state reasonably fast.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    NumPy

    NumPy

    The fundamental package for scientific computing with Python

    Fast and versatile, the NumPy vectorization, indexing, and broadcasting concepts are the de-facto standards of array computing today. NumPy offers comprehensive mathematical functions, random number generators, linear algebra routines, Fourier transforms, and more. NumPy supports a wide range of hardware and computing platforms, and plays well with distributed, GPU, and sparse array libraries. The core of NumPy is well-optimized C code. Enjoy the flexibility of Python with the speed of compiled code. ...
    Downloads: 96 This Week
    Last Update:
    See Project
  • 9
    ObjectBox Go Database

    ObjectBox Go Database

    Embedded Go Database, the fast alternative to SQLite, gorm, etc.

    ObjectBox Go is a lightweight and high-performance embedded NoSQL database designed for Go applications, providing efficient data storage and retrieval with minimal overhead.
    Downloads: 0 This Week
    Last Update:
    See Project
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • 10
    HDF5

    HDF5

    Official HDF5® Library Repository

    HDF5 (Hierarchical Data Format v5) is a widely-used data management library and file format for storing large and complex scientific data sets efficiently.
    Downloads: 22 This Week
    Last Update:
    See Project
  • 11
    ncnn

    ncnn

    High-performance neural network inference framework for mobile

    ncnn is a high-performance neural network inference computing framework designed specifically for mobile platforms. It brings artificial intelligence right at your fingertips with no third-party dependencies, and speeds faster than all other known open source frameworks for mobile phone cpu. ncnn allows developers to easily deploy deep learning algorithm models to the mobile platform and create intelligent APPs.
    Downloads: 92 This Week
    Last Update:
    See Project
  • 12
    ModelingToolkit.jl

    ModelingToolkit.jl

    Modeling framework for automatically parallelized scientific ML

    ModelingToolkit.jl is a modeling language for high-performance symbolic-numeric computation in scientific computing and scientific machine learning. It then mixes ideas from symbolic computational algebra systems with causal and acausal equation-based modeling frameworks to give an extendable and parallel modeling system. It allows for users to give a high-level description of a model for symbolic preprocessing to analyze and enhance the model.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 13
    SIMD

    SIMD

    C++ wrappers for SIMD intrinsics

    SIMD is a C++ library that provides portable abstractions over SIMD (Single Instruction, Multiple Data) instructions, enabling developers to write high-performance vectorized code without dealing directly with architecture-specific intrinsics. SIMD instructions allow a single operation to be applied to multiple data elements simultaneously, significantly accelerating numerical and data-parallel computations. However, differences across CPU architectures and compilers make direct usage complex, which xsimd addresses by offering a unified API that maps efficiently to underlying hardware capabilities. ...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 14
    Apache SeaTunnel

    Apache SeaTunnel

    SeaTunnel is a distributed, high-performance data integration platform

    SeaTunnel is a very easy-to-use ultra-high-performance distributed data integration platform that supports real-time synchronization of massive data. It can synchronize tens of billions of data stably and efficiently every day, and has been used in the production of nearly 100 companies. There are hundreds of commonly-used data sources of which versions are incompatible. With the emergence of new technologies, more data sources are appearing.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 15
    Chitu

    Chitu

    High-performance inference framework for large language models

    Chitu is a high-performance inference engine designed to deploy and run large language models efficiently in production environments. The framework focuses on improving efficiency, flexibility, and scalability for organizations that need to run LLM inference workloads across different hardware platforms. It supports heterogeneous computing environments, including CPUs, GPUs, and various specialized AI accelerators, allowing models to run across a wide range of infrastructure configurations. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    AssemblyScript

    AssemblyScript

    A TypeScript-like language for WebAssembly

    ...Its design bridges the gap between high-level scripting and low-level execution, allowing developers to achieve near-native performance without switching to languages like C or Rust. AssemblyScript integrates with the WebAssembly ecosystem, enabling seamless interaction with JavaScript through standard APIs. It is widely used in scenarios where performance and portability are essential, such as gaming, data processing, and edge computing.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 17
    PowerJob

    PowerJob

    Enterprise job scheduling middleware with distributed computing

    PowerJob is an open-source distributed computing and job scheduling framework which allows developers to easily schedule tasks in their own application. Front-end page is provided and developers can manage their task, monitor the status, check the logs online, etc. Four timing strategies are supported, including CRON expression, fixed rate, fixed delay and OpenAPI which allows you to define your own scheduling policies, such as delaying execution. Four execution modes are supported,...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    CubeCL

    CubeCL

    Multi-platform high-performance compute language extension for Rust

    CubeCL is a low-level compute language and compiler framework designed to simplify and optimize GPU programming for high-performance workloads, particularly in machine learning and numerical computing. It provides an abstraction layer that allows developers to write portable, hardware-efficient compute kernels without directly dealing with complex GPU APIs such as CUDA or OpenCL. CubeCL focuses on delivering predictable performance and composability by exposing explicit control over memory layouts, parallelism, and execution patterns while still maintaining a developer-friendly syntax. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 19
    Google Highway

    Google Highway

    Performance-portable, length-agnostic SIMD with runtime dispatch

    Google Highway is a high-performance C++ library designed to provide portable SIMD (Single Instruction, Multiple Data) vectorization across multiple CPU architectures while maintaining predictable and efficient behavior. It abstracts low-level vector intrinsics into a consistent API that maps closely to hardware instructions, allowing developers to write high-performance code without relying heavily on compiler auto-vectorization.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    CUDA Python

    CUDA Python

    Performance meets Productivity

    ...The toolkit also includes utilities for profiling, memory management, distributed computing, and numerical operations, making it suitable for scientific computing, AI, and data processing workloads.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 21
    Infinispan

    Infinispan

    Infinispan is an open source data grid platform

    Infinispan is a distributed in-memory data grid and caching system designed for high-performance computing. It allows applications to scale dynamically by distributing data across multiple nodes, reducing latency and improving resilience.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    Zvec

    Zvec

    A lightweight, lightning-fast, in-process vector database

    ...Developed by Alibaba’s Tongyi Lab, it positions itself as the “SQLite of vector databases” by being easy to integrate, minimal in dependencies, and capable of handling high throughput with low latency on edge devices or small systems. Zvec excels at approximate nearest neighbor search and retrieval tasks that power features like semantic search, recommendation systems, and retrieval-augmented generation (RAG) setups. Its performance benchmarks show it achieving high queries-per-second and fast index build times compared to similar tools. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 23
    AsmJit

    AsmJit

    Low-latency machine code generation

    AsmJit is a low-level code generation library designed for dynamically creating machine code at runtime, enabling just-in-time (JIT) compilation for performance-critical applications. It provides a high-level API that abstracts away the complexity of writing raw assembly while still allowing fine-grained control over instruction generation. The library supports multiple architectures, including x86 and x64, making it versatile for cross-platform development. It is commonly used in applications such as emulators, compilers, and high-performance computing systems where runtime optimization is essential. asmjit emphasizes low latency and efficiency, ensuring that generated code executes quickly without significant overhead. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    TDengine

    TDengine

    Open-source time-series database with high-performance and scalability

    ...TDengine’s innovative design and purpose-built storage engine outperforms other time-series databases for data ingestion, querying and data compression while significantly reducing storage and computing costs. TDengine’s native distributed architecture powers out-of-the-box scalability and high availability. Nodes can be added through simple configuration to achieve greater data processing power. In addition, this feature is open source. TDengine uses SQL as the query language, thereby reducing learning and migration costs, while adding SQL extensions to handle time-series data better, and supporting convenient and flexible schemaless data ingestion.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Dagger.jl

    Dagger.jl

    A framework for out-of-core and parallel execution

    Dagger.jl is a framework for out-of-core and parallel computing in Julia that allows users to construct and execute dynamic task graphs. It is designed for large-scale, distributed, and memory-efficient computations. Dagger supports lazy evaluation and scheduling across multiple threads or machines, enabling high-performance workflows for data processing, scientific computing, and machine learning.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • 4
  • 5
  • Next
MongoDB Logo MongoDB