Showing 9 open source projects for "cluster"

View related business solutions
  • Run Any Workload on Compute Engine VMs Icon
    Run Any Workload on Compute Engine VMs

    From dev environments to AI training, choose preset or custom VMs with 1–96 vCPUs and industry-leading 99.95% uptime SLA.

    Compute Engine delivers high-performance virtual machines for web apps, databases, containers, and AI workloads. Choose from general-purpose, compute-optimized, or GPU/TPU-accelerated machine types—or build custom VMs to match your exact specs. With live migration and automatic failover, your workloads stay online. New customers get $300 in free credits.
    Try Compute Engine
  • $300 in Free Credit for Your Google Cloud Projects Icon
    $300 in Free Credit for Your Google Cloud Projects

    Build, test, and explore on Google Cloud with $300 in free credit. No hidden charges. No surprise bills.

    Launch your next project with $300 in free Google Cloud credit—no hidden charges. Test, build, and deploy without risk. Use your credit across the Google Cloud platform to find what works best for your needs. After your credits are used, continue building with free monthly usage products. Only pay when you're ready to scale. Sign up in minutes and start exploring.
    Start Free Trial
  • 1
    AlphaZero.jl

    AlphaZero.jl

    A generic, simple and fast implementation of Deepmind's AlphaZero

    Beyond its much publicized success in attaining superhuman level at games such as Chess and Go, DeepMind's AlphaZero algorithm illustrates a more general methodology of combining learning and search to explore large combinatorial spaces effectively. We believe that this methodology can have exciting applications in many different research areas. Because AlphaZero is resource-hungry, successful open-source implementations (such as Leela Zero) are written in low-level languages (such as C++)...
    Downloads: 23 This Week
    Last Update:
    See Project
  • 2
    DualPipe

    DualPipe

    A bidirectional pipeline parallelism algorithm

    ...The main goal of DualPipe is to maximize overlap between computation and communication phases during distributed training, thus reducing idle GPU time (i.e. “pipeline bubbles”) and improving cluster efficiency. Traditional pipeline parallelism methods (e.g. 1F1B or staggered pipelining) leave gaps because forward and backward phases can’t fully overlap with communication. DualPipe addresses that by scheduling micro-batches from both ends of the pipeline in a bidirectional fashion—i.e. some micro-batches flow forward while others flow backward—so that computation on one partition can coincide with communication for another.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    tsfresh

    tsfresh

    Automatic extraction of relevant features from time series

    ...Further tsfresh is compatible with pythons pandas and scikit-learn APIs, two important packages for Data Science endeavours in python. The extracted features can be used to describe or cluster time series based on the extracted characteristics. Further, they can be used to build models that perform classification/regression tasks on the time series. Often the features give new insights into time series and their dynamics.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    supercluster

    supercluster

    A very fast geospatial point clustering library for browsers and Node

    A very fast JavaScript library for geospatial point clustering for browsers and Node. supercluster supports property aggregation with the following two options. Map, a function that returns cluster properties corresponding to a single point. Reduce, a reduce function that merges properties of two clusters into one.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Build AI Apps with Gemini 3 on Vertex AI Icon
    Build AI Apps with Gemini 3 on Vertex AI

    Access Google’s most capable multimodal models. Train, test, and deploy AI with 200+ foundation models on one platform.

    Vertex AI gives developers access to Gemini 3—Google’s most advanced reasoning and coding model—plus 200+ foundation models including Claude, Llama, and Gemma. Build generative AI apps with Vertex AI Studio, customize with fine-tuning, and deploy to production with enterprise-grade MLOps. New customers get $300 in free credits.
    Try Vertex AI Free
  • 5

    SWAPHI-LS: Alignment on Xeon Phi Cluster

    Smith-Waterman long DNA sequence alignment on Xeon Phi clusters

    The first parallel Smith-Waterman algorithm exploiting Intel Xeon Phi clusters to accelerate the alignment of long DNA sequences. This algorithm is written in C++ (with a set of SIMD intrinsic extensions), OpenMP and MPI. The performance evaluation revealed that our algorithm achieves very stable performance, and yields a performance of up to 30.1 GCUPS on a single Xeon Phi and up to 111.4 GCUPS on four Xeon Phis sharing a host.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    EASEA
    EASEA (EAsy Specification of Evolutionary Algorithms: pronounce "easy") is a high-level language dedicated to the specification of evolutionary algorithms. EASEA up to version 0.7 compiles .ez specification files into C++ or JAVA object files.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    IMPORTANT: The project moved over to github! You can find it at: https://github.com/exhuma/python-cluster
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    Celluster is an easy adaptable producer/consumer software for distributed computing on a cluster of Cell B.E. processors
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    FORce based Cluster Editing (FORCE) is a Java software heuristically solving the graph cluster editing problem on weighted edges using BLAST E-values. It further provides a training mode for heuristic parameter estimation.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Easily Host LLMs and Web Apps on Cloud Run Icon
    Easily Host LLMs and Web Apps on Cloud Run

    Run everything from popular models with on-demand NVIDIA L4 GPUs to web apps without infrastructure management.

    Run frontend and backend services, batch jobs, host LLMs, and queue processing workloads without the need to manage infrastructure. Cloud Run gives you on-demand GPU access for hosting LLMs and running real-time AI—with 5-second cold starts and automatic scale-to-zero so you only pay for actual usage. New customers get $300 in free credit to start.
    Try Cloud Run Free
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB