Showing 537 open source projects for "compute"

View related business solutions
  • Application Monitoring That Won't Slow Your App Down Icon
    Application Monitoring That Won't Slow Your App Down

    AppSignal's Rust-based agent is lightweight and stable. Already running in thousands of production apps.

    Full APM with errors, performance, logs, and uptime monitoring. 99.999% uptime SLA on the platform itself.
    Start Free
  • Go From AI Idea to AI App Fast Icon
    Go From AI Idea to AI App Fast

    One platform to build, fine-tune, and deploy ML models. No MLOps team required.

    Access Gemini 3 and 200+ models. Build chatbots, agents, or custom models with built-in monitoring and scaling.
    Try Free
  • 1
    FlashMLA

    FlashMLA

    FlashMLA: Efficient Multi-head Latent Attention Kernels

    ...The library supports both BF16 and FP16 data types, and includes a paged KV cache implementation with a block size of 64 to efficiently manage memory during decoding. On very compute-bound settings, it can reach up to ~660 TFLOPS on H800 SXM5 hardware, while in memory-bound configurations it can push memory throughput to ~3000 GB/s. The team regularly updates it with performance improvements; for example, a 2025 update claims 5 % to 15 % gains on compute-bound workloads while maintaining API compatibility.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    Argo Workflows

    Argo Workflows

    Workflow engine for Kubernetes

    ...Define workflows where each step in the workflow is a container. Model multi-step workflows as a sequence of tasks or capture the dependencies between tasks using a directed acyclic graph (DAG). Easily run compute intensive jobs for machine learning or data processing in a fraction of the time using Argo Workflows on Kubernetes. Run CI/CD pipelines natively on Kubernetes without configuring complex software development products. Argo Workflows is the most popular workflow execution engine for Kubernetes. It can run 1000s of workflows a day, each with 1000s of concurrent tasks. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Feynman

    Feynman

    The open source AI research agent

    Feynman is a command-line AI research agent designed to automate complex research workflows by orchestrating multiple specialized agents that collaborate to gather, analyze, and synthesize information into structured outputs. It operates as a “Claude Code for research,” allowing users to input natural language queries and receive fully developed, source-grounded research briefs, literature reviews, or experimental analyses. The system is built around a multi-agent architecture that includes...
    Downloads: 11 This Week
    Last Update:
    See Project
  • 4
    Burn

    Burn

    Burn is a new comprehensive dynamic Deep Learning Framework

    Burn is a new comprehensive dynamic Deep Learning Framework from Tracel AI built using Rust with extreme flexibility, compute efficiency and portability as its primary goals. Burn emphasizes performance, flexibility, and portability for both training and inference. Developed in Rust, it is designed to empower machine learning engineers and researchers across industry and academia.
    Downloads: 2 This Week
    Last Update:
    See Project
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    More flexibility. More control.

    Generate interest, access liquidity without selling, and execute trades seamlessly. All in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 5
    Edit Banana

    Edit Banana

    Edit Banana: A framework for converting statistical figures

    ...The tool focuses on accessibility, giving hobbyists, content creators, and small teams a way to produce polished visuals without downloading heavyweight software or managing local compute resources. Through AI-driven features like content-aware fill and stylistic adjustments, users can modify or replace regions of an image with contextually relevant content that blends seamlessly with the rest of the composition.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 6
    Bittensor

    Bittensor

    Internet-scale Neural Networks

    Bittensor is a decentralized machine learning protocol that allows AI models to collaborate, learn, and earn tokens within a global network. It introduces a blockchain-based economy for neural networks, where participants are incentivized to contribute valuable knowledge and compute power. Bittensor combines peer-to-peer learning with on-chain rewards, creating a self-governing, scalable AI system that evolves without centralized control. It is a novel approach to aligning incentives in AI development, empowering open contributions while preserving model ownership and decentralization.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 7
    Retrofit For Dart

    Retrofit For Dart

    Retrofit.dart is an dio client generator using source_gen

    retrofit.dart is a type conversion dio client generator using source_gen and inspired by Chopper and Retrofit. If you want to parse models on a separate thread, you can take advantage of the compute function, just like Dio does when converting String data responses into json objects. Avoid using Map values, otherwise, multiple background isolates will be spawned to perform the computation, which is extremely intensive for Dart.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 8
    TurboQuant PyTorch

    TurboQuant PyTorch

    From-scratch PyTorch implementation of Google's TurboQuant

    TurboQuant PyTorch is a specialized deep learning optimization framework designed to accelerate neural network inference and training through advanced quantization techniques within the PyTorch ecosystem. The project focuses on reducing the computational and memory footprint of models by converting floating-point representations into lower-precision formats while preserving performance. It provides tools for experimenting with different quantization strategies, enabling developers to balance...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 9
    GLM-4.1V

    GLM-4.1V

    GLM-4.6V/4.5V/4.1V-Thinking, towards versatile multimodal reasoning

    GLM-4.1V — often referred to as a smaller / lighter version of the GLM-V family — offers a more resource-efficient option for users who want multimodal capabilities without requiring large compute resources. Though smaller in scale, GLM-4.1V maintains competitive performance, particularly impressive on many benchmarks for models of its size: in fact, on a number of multimodal reasoning and vision-language tasks it outperforms some much larger models from other families. It represents a trade-off: somewhat reduced capacity compared to 4.5V or 4.6V, but with benefits in terms of speed, deployability, and lower hardware requirements — making it especially useful for developers experimenting locally, building lightweight agents, or deploying on limited infrastructure. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Compliant and Reliable File Transfers Backed by Top Security Certifications Icon
    Compliant and Reliable File Transfers Backed by Top Security Certifications

    Cerberus FTP Server delivers SOC 2 Type II certified security and FIPS 140-2 validated encryption.

    Stop relying on non-certified, legacy file transfer tools that creak under the weight of modern security demands. Get full audit trails, advanced access controls and more supported by an award-winning team of experts. Start your free 25-day trial today.
    Start Free Trial
  • 10
    wgpu

    wgpu

    Safe and portable GPU abstraction in Rust, implementing WebGPU API

    wgpu is a safe and portable graphics library for Rust based on the WebGPU API. It is suitable for general purpose graphics and compute on the GPU. Applications using wgpu run natively on Vulkan, Metal, DirectX 11/12, and OpenGL ES; and browsers via WebAssembly on WebGPU and WebGL2. Angle is a translation layer from GLES to other backends, developed by Google. We support running our GLES3 backend over it in order to reach platforms with GLES2 or DX11 support, which aren't accessible otherwise. ...
    Downloads: 8 This Week
    Last Update:
    See Project
  • 11

    Microsoft SEAL

    Easy-to-use and powerful homomorphic encryption library

    ...Developed by the Cryptography and Privacy Research group at Microsoft, it enables software engineers to build end-to-end encrypted data storage and computation services that never have to procure the customer's key. Microsoft SEAL is very easy to use, compile and run in many different environments. Homomorphic encryption is an encryption scheme that allows the cloud to compute directly on the encrypted data, without requiring the data to be decrypted first. This results in encrypted computations remaining encrypted, decrypted only by the data owner using the secret key.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 12
    MiniMax-M1

    MiniMax-M1

    Open-weight, large-scale hybrid-attention reasoning model

    ...It is built on the MiniMax-Text-01 foundation and keeps the same massive parameter budget, but reworks the attention and training setup for better reasoning and test-time compute scaling. Architecturally, it combines Mixture-of-Experts layers with lightning attention, enabling the model to support a native context length of 1 million tokens while using far fewer FLOPs than comparable reasoning models for very long generations. The team emphasizes efficient scaling of test-time compute: at 100K-token generation lengths, M1 reportedly uses only about 25 percent of the FLOPs of some competing models, making extended “think step” traces more feasible. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Substra

    Substra

    Low-level Python library used to interact with a Substra network

    An open-source framework supporting privacy-preserving, traceable federated learning and machine learning orchestration. Offers a Python SDK, high-level FL library (SubstraFL), and web UI to define datasets, models, tasks, and orchestrate secure, auditable collaborations.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Librosa

    Librosa

    Python library for audio and music analysis

    Librosa is a powerful Python library for analyzing and processing audio and music signals. Built on top of NumPy, SciPy, and matplotlib, it provides a wide range of tools for feature extraction, time-series manipulation, audio display, and music information retrieval. Whether you're building machine learning models for audio classification or visualizing spectrograms, Librosa is a go-to library for researchers and developers working in audio signal processing.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Model Explorer

    Model Explorer

    A modern model graph visualizer and debugger

    Model Explorer is a visual tool for exploring, debugging, and optimizing ML models deployed on edge devices. Developed by Google AI Edge, it offers a browser-based interface to inspect layer-wise performance, memory usage, and inference timing of TensorFlow Lite and other supported models. It’s a powerful utility for developers optimizing models for constrained environments.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    oneAPI.jl

    oneAPI.jl

    Julia support for the oneAPI programming toolkit.

    Julia support for the oneAPI programming toolkit. oneAPI.jl provides support for working with the oneAPI unified programming model. The package is verified to work with the (currently) only implementation of this interface that is part of the Intel Compute Runtime, only available on Linux. This package is still under significant development, so expect bugs and missing features.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Multica

    Multica

    The open-source managed agents platform

    ...It introduces a paradigm where agents can be assigned tasks, participate in discussions, and autonomously execute work while reporting progress and blockers in real time. The system integrates with multiple AI coding tools and provides a unified interface for managing tasks, compute environments, and agent execution pipelines. It includes both a web interface and a CLI that connects local or cloud-based runtimes to the platform, enabling flexible deployment and scaling. Multica emphasizes collaboration between humans and AI by allowing agents to operate alongside developers in shared workspaces. It also supports reusable skill accumulation, meaning that solutions generated by agents can be reused across projects to improve efficiency over time.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 18
    DeepSeek Coder

    DeepSeek Coder

    DeepSeek Coder: Let the Code Write Itself

    DeepSeek-Coder is a series of code-specialized language models designed to generate, complete, and infill code (and mixed code + natural language) with high fluency in both English and Chinese. The models are trained from scratch on a massive corpus (~2 trillion tokens), of which about 87% is code and 13% is natural language. This dataset covers project-level code structure (not just line-by-line snippets), using a large context window (e.g. 16K) and a secondary fill-in-the-blank objective...
    Downloads: 11 This Week
    Last Update:
    See Project
  • 19
    CausalityTools.jl

    CausalityTools.jl

    Algorithms for detecting associations, dynamical influences

    CausalityTools.jl is a package for quantifying associations and dynamical coupling between datasets, independence testing, and causal inference. Association measures from conventional statistics, information theory, and dynamical systems theory, for example, distance correlation, mutual information, transfer entropy, convergent cross mapping and a lot more. A dedicated API for independence testing, which comes with automatic compatibility with every measure-estimator combination you can...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 20
    Spice.ai OSS

    Spice.ai OSS

    A self-hostable CDN for databases

    Spice is a portable runtime offering developers a unified SQL interface to materialize, accelerate, and query data from any database, data warehouse, or data lake. Spice connects, fuses, and delivers data to applications, machine-learning models, and AI backends, functioning as an application-specific, tier-optimized Database CDN. The Spice runtime, written in Rust, is built-with industry-leading technologies such as Apache DataFusion, Apache Arrow, Apache Arrow Flight, SQLite, and DuckDB....
    Downloads: 4 This Week
    Last Update:
    See Project
  • 21
    WebLLM

    WebLLM

    Bringing large-language models and chat to web browsers

    WebLLM is a modular, customizable javascript package that directly brings language model chats directly onto web browsers with hardware acceleration. Everything runs inside the browser with no server support and is accelerated with WebGPU. We can bring a lot of fun opportunities to build AI assistants for everyone and enable privacy while enjoying GPU acceleration. WebLLM offers a minimalist and modular interface to access the chatbot in the browser. The WebLLM package itself does not come...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 22
    NVTX (NVIDIA Tools Extension Library)

    NVTX (NVIDIA Tools Extension Library)

    C-based Application Programming Interface (API)

    ...It allows developers to insert markers, ranges, and events directly into their applications, providing contextual insight into how code executes on CPUs and GPUs. These annotations are visualized in tools such as NVIDIA Nsight Systems and Nsight Compute, enabling developers to identify performance bottlenecks, track execution flow, and correlate application behavior with hardware activity. The API is written in C and includes wrappers for C++ and Python, making it accessible across different programming environments and workloads. NVTX is particularly valuable in high-performance computing and AI workloads where understanding concurrency, memory usage, and kernel execution is critical for optimization.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 23
    OpenMeter

    OpenMeter

    Metering and Billing for AI, API and DevOps

    OpenMeter is an open-source metering and billing platform designed to collect, aggregate, and analyze usage events from APIs, cloud infrastructure, and software services in real time, enabling flexible usage-based billing for SaaS, AI, and DevOps offerings. It supports high-scale event ingestion and deduplication to accurately record how customers consume billable resources such as API calls, compute time, or storage, and then correlates that usage with payment systems and billing plans to automate invoicing and revenue recognition. The system includes metering, storage, cataloging of products and pricing rules, and tools to enforce limits or quotas, supporting both self-service customer portals and internal dashboards. OpenMeter integrates with payment processors, CRMs, and analytics systems, letting teams launch complex pricing models or subscription tiers without heavy custom engineering work.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 24
    GLM-4.5

    GLM-4.5

    GLM-4.5: Open-source LLM for intelligent agents by Z.ai

    GLM-4.5 is a cutting-edge open-source large language model designed by Z.ai for intelligent agent applications. The flagship GLM-4.5 model has 355 billion total parameters with 32 billion active parameters, while the compact GLM-4.5-Air version offers 106 billion total parameters and 12 billion active parameters. Both models unify reasoning, coding, and intelligent agent capabilities, providing two modes: a thinking mode for complex reasoning and tool usage, and a non-thinking mode for...
    Downloads: 54 This Week
    Last Update:
    See Project
  • 25
    Paddler

    Paddler

    Open-source LLM load balancer and serving platform for hosting LLMs

    ...The system acts as a specialized load balancer and serving layer for language models, enabling organizations to run inference workloads without relying on external API providers. It supports running models locally through engines such as llama.cpp while distributing requests across multiple compute nodes to improve performance and reliability. The architecture is designed with privacy and cost control in mind, making it suitable for organizations that handle sensitive data or require predictable operational costs. Paddler also includes tools for monitoring, request buffering, and autoscaling integration so that deployments can adapt dynamically to changing workloads. ...
    Downloads: 5 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB