Showing 4 open source projects for "parallel computing datamaning"

View related business solutions
  • Run Any Workload on Compute Engine VMs Icon
    Run Any Workload on Compute Engine VMs

    From dev environments to AI training, choose preset or custom VMs with 1–96 vCPUs and industry-leading 99.95% uptime SLA.

    Compute Engine delivers high-performance virtual machines for web apps, databases, containers, and AI workloads. Choose from general-purpose, compute-optimized, or GPU/TPU-accelerated machine types—or build custom VMs to match your exact specs. With live migration and automatic failover, your workloads stay online. New customers get $300 in free credits.
    Try Compute Engine
  • Ship AI Apps Faster with Vertex AI Icon
    Ship AI Apps Faster with Vertex AI

    Go from idea to deployed AI app without managing infrastructure. Vertex AI offers one platform for the entire AI development lifecycle.

    Ship AI apps and features faster with Vertex AI—your end-to-end AI platform. Access Gemini 3 and 200+ foundation models, fine-tune for your needs, and deploy with enterprise-grade MLOps. Build chatbots, agents, or custom models. New customers get $300 in free credit.
    Try Vertex AI Free
  • 1
    LinDB

    LinDB

    LinDB is a scalable, high performance, high availability database

    LinDB is a scalable, high-performance, high-availability distributed time series database. A single server could easily support more than one million write TPS; With fundamental techniques like efficient compression storage and parallel computing, LinDB delivers highly optimized query performance. The multi-channel replication protocol supports any amount of nodes, and ensures the system's availability. Schema-free multi-dimensional data model with Metric, Tags, and Fields; The LinQL is flexible yet handy for real-time data analytics. Horizontal scalable is made simple by adding more new broker and storage nodes without too much thinking and manual operations. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2

    PDP-OmniSim

    PDP-OmniSim simulating parallel and distributed processing systems

    PDP-OmniSim 🧬 Scientific Overview PDP-OmniSim is an advanced computational framework for simulating parallel and distributed processing systems, with cutting-edge applications in computational neuroscience, distributed computing, and complex systems modeling. The framework provides researchers with robust tools for large-scale simulations of networked systems and their emergent behaviors. 🎯 Key Scientific Contributions 🔬 Interdisciplinary Research Domains Computational Neuroscience: Large-scale neural population dynamics, brain-inspired computing architectures, and neuro-symbolic AI systems Distributed Systems: Scalable parallel processing simulations, resource allocation optimization, and fault-tolerant computing Complex Systems: Emergent behavior in networked systems, self-organizing criticality, and adaptive network topologies
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    QuickRNASeq

    QuickRNASeq

    A pipeline for large scale RNA-seq data analysis

    We have implemented QuickRNASeq, an open-source based pipeline for large scale RNA-seq data analysis. QuickRNASeq takes advantage of parallel computing resources, a careful selection of previously published algorithms for RNA-seq read mapping, counting and quality control, and a three-stage strategy to build a fully automated workflow. We also implemented built-in functionalities to detect sample swapping or mislabeling in large-scale RNA-seq studies. Our pipeline significantly lifts large-scale RNA-seq data analysis to the next level of automation and visualization. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Parallel Colt is a multithreaded version of Colt - a library for high performance scientific computing in Java. It contains efficient algorithms for data analysis, linear algebra, multi-dimensional arrays, Fourier transforms, statistics and histogramming
    Downloads: 0 This Week
    Last Update:
    See Project
  • Easily Host LLMs and Web Apps on Cloud Run Icon
    Easily Host LLMs and Web Apps on Cloud Run

    Run everything from popular models with on-demand NVIDIA L4 GPUs to web apps without infrastructure management.

    Run frontend and backend services, batch jobs, host LLMs, and queue processing workloads without the need to manage infrastructure. Cloud Run gives you on-demand GPU access for hosting LLMs and running real-time AI—with 5-second cold starts and automatic scale-to-zero so you only pay for actual usage. New customers get $300 in free credit to start.
    Try Cloud Run Free
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB