Showing 14 open source projects for "compute"

View related business solutions
  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build, govern, and optimize agents and models with Gemini Enterprise Agent Platform.
    Start Free
  • 1
    CubeCL

    CubeCL

    Multi-platform high-performance compute language extension for Rust

    CubeCL is a low-level compute language and compiler framework designed to simplify and optimize GPU programming for high-performance workloads, particularly in machine learning and numerical computing. It provides an abstraction layer that allows developers to write portable, hardware-efficient compute kernels without directly dealing with complex GPU APIs such as CUDA or OpenCL.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 2
    Databend

    Databend

    Cloud-native open source data warehouse for analytics and AI queries

    Databend is an open source cloud-native data warehouse designed for large-scale analytics and modern data workloads. Built in Rust, the system focuses on high performance, scalability, and efficient data processing for analytical queries. It is designed with a separation of compute and storage, allowing compute nodes to scale independently while storing data in object storage systems. This architecture enables cost-efficient storage and elastic scaling for workloads that involve large datasets and complex queries. Databend provides a unified engine capable of handling analytics, vector search, and full-text search within a single platform. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 3
    Vortex

    Vortex

    The LLVM of columnar file formats

    Vortex is a high-performance toolkit designed for working with compressed Apache Arrow arrays, providing functionality for in-memory, on-disk, and over-the-wire data handling. It aims to be an advanced successor to Apache Parquet, offering dramatically faster random access reads and scans, while maintaining similar compression ratios. Vortex's modular design allows for extensibility, enabling developers to implement custom encodings for efficient data management, particularly for large-scale...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 4
    Firecracker

    Firecracker

    Secure and fast microVMs for serverless computing

    Firecracker is an open-source virtualization technology developed by AWS for deploying secure micro-VMs (microVMs) that offer strong isolation with minimal overhead. Designed for serverless workloads (e.g., AWS Lambda, Fargate), it combines VM-level security with container-like performance and startup speed.
    Downloads: 2 This Week
    Last Update:
    See Project
  • Secure File Transfer for Windows with Cerberus by Redwood Icon
    Secure File Transfer for Windows with Cerberus by Redwood

    Protect and share files over FTP/S, SFTP, HTTPS and SCP with the #1 rated Windows file transfer server.

    Cerberus supports unlimited users and connections on a single IP, with built-in encryption, 2FA, and a browser-based web client — all deployable in under 15 minutes with a 25-day free trial.
    Try for Free
  • 5
    Neon Database

    Neon Database

    Neon: Serverless Postgres. We separated storage and compute

    The database you love, on a serverless platform designed to help you build reliable and scalable applications faster. Focus on building applications with time and money-saving features like instant provisioning, autoscaling according to load, and scale to zero. Instantly branch your data and schema to access isolated DB copies for development, CI/CD, and schema migrations with copy-on-write storage. Integrate it into your language or framework within minutes and unlock a simpler developer...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 6
    Burn

    Burn

    Burn is a new comprehensive dynamic Deep Learning Framework

    Burn is a new comprehensive dynamic Deep Learning Framework from Tracel AI built using Rust with extreme flexibility, compute efficiency and portability as its primary goals. Burn emphasizes performance, flexibility, and portability for both training and inference. Developed in Rust, it is designed to empower machine learning engineers and researchers across industry and academia.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 7
    wgpu

    wgpu

    Safe and portable GPU abstraction in Rust, implementing WebGPU API

    wgpu is a safe and portable graphics library for Rust based on the WebGPU API. It is suitable for general purpose graphics and compute on the GPU. Applications using wgpu run natively on Vulkan, Metal, DirectX 11/12, and OpenGL ES; and browsers via WebAssembly on WebGPU and WebGL2. Angle is a translation layer from GLES to other backends, developed by Google. We support running our GLES3 backend over it in order to reach platforms with GLES2 or DX11 support, which aren't accessible otherwise. ...
    Downloads: 8 This Week
    Last Update:
    See Project
  • 8
    Spice.ai OSS

    Spice.ai OSS

    A self-hostable CDN for databases

    Spice is a portable runtime offering developers a unified SQL interface to materialize, accelerate, and query data from any database, data warehouse, or data lake. Spice connects, fuses, and delivers data to applications, machine-learning models, and AI backends, functioning as an application-specific, tier-optimized Database CDN. The Spice runtime, written in Rust, is built-with industry-leading technologies such as Apache DataFusion, Apache Arrow, Apache Arrow Flight, SQLite, and DuckDB....
    Downloads: 4 This Week
    Last Update:
    See Project
  • 9
    Paddler

    Paddler

    Open-source LLM load balancer and serving platform for hosting LLMs

    ...The system acts as a specialized load balancer and serving layer for language models, enabling organizations to run inference workloads without relying on external API providers. It supports running models locally through engines such as llama.cpp while distributing requests across multiple compute nodes to improve performance and reliability. The architecture is designed with privacy and cost control in mind, making it suitable for organizations that handle sensitive data or require predictable operational costs. Paddler also includes tools for monitoring, request buffering, and autoscaling integration so that deployments can adapt dynamically to changing workloads. ...
    Downloads: 5 This Week
    Last Update:
    See Project
  • Build Securely on AWS with Proven Frameworks Icon
    Build Securely on AWS with Proven Frameworks

    Lay a foundation for success with Tested Reference Architectures developed by Fortinet’s experts. Learn more in this white paper.

    Moving to the cloud brings new challenges. How can you manage a larger attack surface while ensuring great network performance? Turn to Fortinet’s Tested Reference Architectures, blueprints for designing and securing cloud environments built by cybersecurity experts. Learn more and explore use cases in this white paper.
    Download Now
  • 10
    rspirv

    rspirv

    Rust implementation of SPIR-V module processing functionalities

    rspirv is a Rust-based parser, builder, and disassembler for SPIR-V, the intermediate binary format used in Vulkan and OpenCL for shaders and compute kernels. It’s part of the gfx-rs ecosystem, a suite of graphics tools aiming to provide cross-platform rendering capabilities in Rust. rspirv enables manipulation and inspection of SPIR-V modules, which is useful in shader compilers, graphics drivers, and development tools for low-level GPU programming. The library strictly follows the SPIR-V specification and is used in projects that need to generate, analyze, or modify shaders dynamically.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    Quickwit

    Quickwit

    Sub-second search & analytics engine on cloud storage

    Sub-second search & analytics engine on cloud storage. Quickwit is the fastest search engine on cloud storage. Quickwit has an Elasticsearch-compatible Ingest-API to make it easier to migrate your log shippers (Vector, Fluent Bit, Syslog, ...) to Quickwit. However, we only support ES aggregation DSL, the query DSL support is planned for Q2 2023. The core difference and advantage of Quickwit are its architecture built from the ground to search on cloud storage. We optimized IO paths, revamped...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    DCV Color Primitives

    DCV Color Primitives

    DCV Color Primitives Library

    ...You might want to propagate errors to the caller function or mix with some other error types. So far, buffers were sized taking into account the image pixel format and dimensions; However, you can use a function to compute how many bytes are needed to store an image of a given format and size. If your data is scattered in multiple buffers that are not necessarily contiguous, you can provide image planes. To take into account data which is not tightly packed, you can provide image strides.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 13
    Meteroid

    Meteroid

    Open-source pricing and billing infrastructure

    Meteroid is an open-source, cloud-native pricing and billing infrastructure designed to support usage-based billing, subscription management, invoicing, and cost-tracking — particularly for SaaS or product-led businesses that need flexible billing models beyond flat subscriptions. It provides an API-driven billing engine: you feed Meteroid usage data from your application (e.g. API calls, storage usage, seats, feature usage), define custom pricing rules or plans, and Meteroid calculates...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 14
    llm

    llm

    An ecosystem of Rust libraries for working with large language models

    ...Text generation can be done as a one-off based on a prompt, or interactively, through REPL or chat modes. The CLI can also be used to serialize (print) decoded models, quantize GGML files, or compute the perplexity of a model. It can be downloaded from the latest GitHub release or by installing it from crates.io.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB