208 projects for "compute" with 1 filter applied:

  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build, govern, and optimize agents and models with Gemini Enterprise Agent Platform.
    Start Free
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 1
    tt-metal

    tt-metal

    TT-NN operator library, and TT-Metalium low level kernel programming

    ...The project is designed for developers who need direct access to the company’s Tensix processor architecture, exposing a programming model that is closer to hardware control than high-level inference frameworks. Instead of following a traditional GPU model centered on massive thread parallelism, the platform is built around a grid of specialized compute nodes called Tensix cores, each with local SRAM, dedicated compute units, and multiple RISC-V control processors. The SDK provides the abstractions and APIs needed to manage data movement, compute kernels, memory coordination, and execution flow across this architecture.
    Downloads: 13 This Week
    Last Update:
    See Project
  • 2
    Bacalhau

    Bacalhau

    Community-driven, simple, yet powerful framework

    Bacalhau is a decentralized compute platform for running jobs on data stored across distributed networks, like IPFS or Filecoin, without moving the data to centralized cloud environments. It allows developers to run containerized workloads close to where the data lives, reducing latency, cost, and privacy risks. Bacalhau supports various runtime environments and is designed to make decentralized data processing as accessible as traditional cloud computing.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Hono

    Hono

    Web framework built on Web Standards

    Hono is a small, simple, and ultrafast web framework built on Web Standards. It works on any JavaScript runtime: Cloudflare Workers, Fastly Compute, Deno, Bun, Vercel, AWS Lambda, Lambda@Edge, and Node.js. The router RegExpRouter is really fast. The hono/tiny preset is under 14kB. Using only Web Standard APIs. Works on Cloudflare, Fastly, Deno, Bun, AWS, or Node.js. The same code runs on all platforms. Hono has built-in middleware, custom middleware, third-party middleware, and helpers. ...
    Downloads: 10 This Week
    Last Update:
    See Project
  • 4
    VulkanSceneGraph

    VulkanSceneGraph

    Vulkan & C++17 based Scene Graph Project

    VulkanSceneGraph (VSG), is a modern, cross-platform, high-performance scene graph library built upon Vulkan graphics/compute API. The software is written in C++17 and follows the CppCoreGuidelines and FOSS Best Practices. The source code is published under the MIT License, with the exception of vulkan.h, used for Vulkan extensions, which is under Apache License 2.0. This repository contains C++ headers and source and CMake build scripts to build the libvsg library.
    Downloads: 6 This Week
    Last Update:
    See Project
  • Full-stack observability with actually useful AI | Grafana Cloud Icon
    Full-stack observability with actually useful AI | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 5
    Google Cloud's buildpacks

    Google Cloud's buildpacks

    Builders and buildpacks designed to run on Google Cloud's container

    ...It is built around full compatibility with the Cloud Native Buildpacks specification, which means it can participate in standardized build workflows while still being tailored to Google Cloud environments. The project supports major runtime destinations such as Cloud Run, GKE, Anthos, and Compute Engine with Container-Optimized OS, and it also powers build flows for App Engine and Cloud Functions. One of its strengths is that it does not only provide buildpacks themselves, but also builder images that work with tools such as pack, kpack, Tekton, and Skaffold, giving developers flexibility in how they incorporate it into CI/CD systems.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 6
    clip-retrieval

    clip-retrieval

    Easily compute clip embeddings and build a clip retrieval system

    clip-retrieval is an open-source toolkit designed to build large-scale semantic search systems for images and text by leveraging CLIP embeddings to enable multimodal retrieval. It allows developers to compute embeddings for both images and text efficiently and then index them for fast similarity search across massive datasets. The system is optimized for performance and scalability, capable of processing tens or even hundreds of millions of embeddings using GPU acceleration. It includes components for inference, indexing, filtering, and serving results through APIs, making it a complete pipeline for building production-ready retrieval systems. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 7
    MiniRAG

    MiniRAG

    Making RAG Simpler with Small and Open-Sourced Language Models

    MiniRAG is a lightweight retrieval-augmented generation tool designed to bring the benefits of RAG workflows to smaller datasets, edge environments, and constrained compute settings by simplifying embedding, indexing, and retrieval. It extracts text from documents, codes, or other structured inputs and converts them into embeddings using efficient models, then stores these vectors for fast nearest-neighbor search without requiring huge databases or separate vector servers. When a query is issued, MiniRAG retrieves the most relevant contexts and feeds them into a generative model to produce an answer that is grounded in the source material rather than hallucinated. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 8
    DeepSeekMath-V2

    DeepSeekMath-V2

    Towards self-verifiable mathematical reasoning

    DeepSeekMath-V2 is a large-scale open-source AI model designed specifically for advanced mathematical reasoning, theorem proving, and rigorous proof verification. It’s built by DeepSeek as a successor to their earlier math-specialist models. Unlike general-purpose LLMs that might generate plausible-looking math but sometimes hallucinate or mishandle rigorous logic, Math-V2 is engineered to not only generate solutions but also self-verify them, meaning it examines the derivations, checks...
    Downloads: 9 This Week
    Last Update:
    See Project
  • 9
    Apache Polaris

    Apache Polaris

    Apache Polaris, the interoperable, open source catalog

    Apache Polaris is an open-source metadata catalog and data management service designed to manage Apache Iceberg tables in modern data lakehouse environments. It provides a centralized catalog that allows multiple compute engines and analytics systems to interact with the same datasets through a standardized interface. By implementing the Iceberg REST catalog API, Polaris enables distributed data platforms to access shared table metadata without tightly coupling storage systems and query engines. This design allows organizations to run queries on the same Iceberg tables using tools such as Apache Spark, Flink, Trino, and other analytics engines while maintaining consistency across platforms. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • Auth0 B2B Essentials: SSO, MFA, and RBAC Built In Icon
    Auth0 B2B Essentials: SSO, MFA, and RBAC Built In

    Unlimited organizations, 3 enterprise SSO connections, role-based access control, and pro MFA included. Dev and prod tenants out of the box.

    Auth0's B2B Essentials plan gives you everything you need to ship secure multi-tenant apps. Unlimited orgs, enterprise SSO, RBAC, audit log streaming, and higher auth and API limits included. Add on M2M tokens, enterprise MFA, or additional SSO connections as you scale.
    Sign Up Free
  • 10
    MiniMind-V

    MiniMind-V

    "Big Model" trains a visual multimodal VLM with 26M parameters

    MiniMind-V is an experimental open-source project that aims to train a very small multimodal vision–language model (VLM) from scratch with extremely low compute and cost, making research and experimentation accessible to more people. The repository showcases training workflows and code designed to produce a 26-million parameter model—including both image and text capabilities—using minimal resources in very little time, reflecting a trend toward democratizing AI research. MiniMind-V combines techniques from modern vision-language modeling but focuses on efficiency and simplicity so that individuals or small teams can explore multimodal learning without massive GPU clusters. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Argo Workflows

    Argo Workflows

    Workflow engine for Kubernetes

    ...Define workflows where each step in the workflow is a container. Model multi-step workflows as a sequence of tasks or capture the dependencies between tasks using a directed acyclic graph (DAG). Easily run compute intensive jobs for machine learning or data processing in a fraction of the time using Argo Workflows on Kubernetes. Run CI/CD pipelines natively on Kubernetes without configuring complex software development products. Argo Workflows is the most popular workflow execution engine for Kubernetes. It can run 1000s of workflows a day, each with 1000s of concurrent tasks. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 12
    Netflix Maestro

    Netflix Maestro

    Netflix’s Workflow Orchestrator

    ...It was designed to support the demanding internal infrastructure of Netflix, where thousands of workflows must process massive volumes of data reliably and efficiently every day. The platform enables engineers and data scientists to define workflows using structured configuration files and execute tasks across diverse compute environments, including scripts, containers, and notebook environments. Maestro provides built-in mechanisms for retry logic, task scheduling, dependency management, and error handling, which are essential when orchestrating production-scale pipelines.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    OpenMythos

    OpenMythos

    A theoretical reconstruction of the Claude Mythos architecture

    ...The architecture incorporates advanced techniques such as mixture-of-experts routing, adaptive computation time, and multiple attention mechanisms to dynamically allocate compute where needed. It is highly configurable through a centralized configuration system, allowing experimentation with different architectural parameters such as loop depth, attention type.
    Downloads: 20 This Week
    Last Update:
    See Project
  • 14
    FlashMLA

    FlashMLA

    FlashMLA: Efficient Multi-head Latent Attention Kernels

    ...The library supports both BF16 and FP16 data types, and includes a paged KV cache implementation with a block size of 64 to efficiently manage memory during decoding. On very compute-bound settings, it can reach up to ~660 TFLOPS on H800 SXM5 hardware, while in memory-bound configurations it can push memory throughput to ~3000 GB/s. The team regularly updates it with performance improvements; for example, a 2025 update claims 5 % to 15 % gains on compute-bound workloads while maintaining API compatibility.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Feynman

    Feynman

    The open source AI research agent

    Feynman is a command-line AI research agent designed to automate complex research workflows by orchestrating multiple specialized agents that collaborate to gather, analyze, and synthesize information into structured outputs. It operates as a “Claude Code for research,” allowing users to input natural language queries and receive fully developed, source-grounded research briefs, literature reviews, or experimental analyses. The system is built around a multi-agent architecture that includes...
    Downloads: 11 This Week
    Last Update:
    See Project
  • 16
    Edit Banana

    Edit Banana

    Edit Banana: A framework for converting statistical figures

    ...The tool focuses on accessibility, giving hobbyists, content creators, and small teams a way to produce polished visuals without downloading heavyweight software or managing local compute resources. Through AI-driven features like content-aware fill and stylistic adjustments, users can modify or replace regions of an image with contextually relevant content that blends seamlessly with the rest of the composition.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 17
    Multica

    Multica

    The open-source managed agents platform

    ...It introduces a paradigm where agents can be assigned tasks, participate in discussions, and autonomously execute work while reporting progress and blockers in real time. The system integrates with multiple AI coding tools and provides a unified interface for managing tasks, compute environments, and agent execution pipelines. It includes both a web interface and a CLI that connects local or cloud-based runtimes to the platform, enabling flexible deployment and scaling. Multica emphasizes collaboration between humans and AI by allowing agents to operate alongside developers in shared workspaces. It also supports reusable skill accumulation, meaning that solutions generated by agents can be reused across projects to improve efficiency over time.
    Downloads: 12 This Week
    Last Update:
    See Project
  • 18
    GLM-4.1V

    GLM-4.1V

    GLM-4.6V/4.5V/4.1V-Thinking, towards versatile multimodal reasoning

    GLM-4.1V — often referred to as a smaller / lighter version of the GLM-V family — offers a more resource-efficient option for users who want multimodal capabilities without requiring large compute resources. Though smaller in scale, GLM-4.1V maintains competitive performance, particularly impressive on many benchmarks for models of its size: in fact, on a number of multimodal reasoning and vision-language tasks it outperforms some much larger models from other families. It represents a trade-off: somewhat reduced capacity compared to 4.5V or 4.6V, but with benefits in terms of speed, deployability, and lower hardware requirements — making it especially useful for developers experimenting locally, building lightweight agents, or deploying on limited infrastructure. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    TurboQuant PyTorch

    TurboQuant PyTorch

    From-scratch PyTorch implementation of Google's TurboQuant

    TurboQuant PyTorch is a specialized deep learning optimization framework designed to accelerate neural network inference and training through advanced quantization techniques within the PyTorch ecosystem. The project focuses on reducing the computational and memory footprint of models by converting floating-point representations into lower-precision formats while preserving performance. It provides tools for experimenting with different quantization strategies, enabling developers to balance...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 20
    MiniMax-M1

    MiniMax-M1

    Open-weight, large-scale hybrid-attention reasoning model

    ...It is built on the MiniMax-Text-01 foundation and keeps the same massive parameter budget, but reworks the attention and training setup for better reasoning and test-time compute scaling. Architecturally, it combines Mixture-of-Experts layers with lightning attention, enabling the model to support a native context length of 1 million tokens while using far fewer FLOPs than comparable reasoning models for very long generations. The team emphasizes efficient scaling of test-time compute: at 100K-token generation lengths, M1 reportedly uses only about 25 percent of the FLOPs of some competing models, making extended “think step” traces more feasible. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    OpenMeter

    OpenMeter

    Metering and Billing for AI, API and DevOps

    OpenMeter is an open-source metering and billing platform designed to collect, aggregate, and analyze usage events from APIs, cloud infrastructure, and software services in real time, enabling flexible usage-based billing for SaaS, AI, and DevOps offerings. It supports high-scale event ingestion and deduplication to accurately record how customers consume billable resources such as API calls, compute time, or storage, and then correlates that usage with payment systems and billing plans to automate invoicing and revenue recognition. The system includes metering, storage, cataloging of products and pricing rules, and tools to enforce limits or quotas, supporting both self-service customer portals and internal dashboards. OpenMeter integrates with payment processors, CRMs, and analytics systems, letting teams launch complex pricing models or subscription tiers without heavy custom engineering work.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 22
    DeepSeek Coder

    DeepSeek Coder

    DeepSeek Coder: Let the Code Write Itself

    DeepSeek-Coder is a series of code-specialized language models designed to generate, complete, and infill code (and mixed code + natural language) with high fluency in both English and Chinese. The models are trained from scratch on a massive corpus (~2 trillion tokens), of which about 87% is code and 13% is natural language. This dataset covers project-level code structure (not just line-by-line snippets), using a large context window (e.g. 16K) and a secondary fill-in-the-blank objective...
    Downloads: 11 This Week
    Last Update:
    See Project
  • 23
    GLM-4.5

    GLM-4.5

    GLM-4.5: Open-source LLM for intelligent agents by Z.ai

    GLM-4.5 is a cutting-edge open-source large language model designed by Z.ai for intelligent agent applications. The flagship GLM-4.5 model has 355 billion total parameters with 32 billion active parameters, while the compact GLM-4.5-Air version offers 106 billion total parameters and 12 billion active parameters. Both models unify reasoning, coding, and intelligent agent capabilities, providing two modes: a thinking mode for complex reasoning and tool usage, and a non-thinking mode for...
    Downloads: 56 This Week
    Last Update:
    See Project
  • 24
    The Pope Bot

    The Pope Bot

    Autonomous AI agent that you can configure and build

    ...It’s designed so that every action taken by the agent is logged as a git commit, giving users complete visibility into what the agent did, why it did it, and when, which makes actions auditable and reversible. The framework treats the repository itself as the agent’s “brain,” and GitHub Actions serve as the compute layer, enabling tasks to run securely without exposing sensitive API keys to the underlying AI. The system integrates with messaging platforms like Telegram, where users can interact with the bot, trigger actions, or receive notifications, and supports scheduling and automation through patterns of request handling.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 25
    Attention Residuals (AttnRes)

    Attention Residuals (AttnRes)

    Drop-in replacement for standard residual connections in Transformers

    Attention Residuals is a research-driven architectural innovation for transformer-based models that replaces traditional residual connections with an attention-based mechanism to improve information flow across layers. In standard transformers, residual connections simply sum outputs from previous layers, which can lead to uncontrolled growth of hidden states and dilution of early-layer information in deep networks. Attention Residuals introduces a learnable softmax attention mechanism that...
    Downloads: 1 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • 4
  • 5
  • Next
MongoDB Logo MongoDB