Showing 1 open source project for "code server"

View related business solutions
  • Ship AI Apps Faster with Vertex AI Icon
    Ship AI Apps Faster with Vertex AI

    Go from idea to deployed AI app without managing infrastructure. Vertex AI offers one platform for the entire AI development lifecycle.

    Ship AI apps and features faster with Vertex AI—your end-to-end AI platform. Access Gemini 3 and 200+ foundation models, fine-tune for your needs, and deploy with enterprise-grade MLOps. Build chatbots, agents, or custom models. New customers get $300 in free credit.
    Try Vertex AI Free
  • Easily Host LLMs and Web Apps on Cloud Run Icon
    Easily Host LLMs and Web Apps on Cloud Run

    Run everything from popular models with on-demand NVIDIA L4 GPUs to web apps without infrastructure management.

    Run frontend and backend services, batch jobs, host LLMs, and queue processing workloads without the need to manage infrastructure. Cloud Run gives you on-demand GPU access for hosting LLMs and running real-time AI—with 5-second cold starts and automatic scale-to-zero so you only pay for actual usage. New customers get $300 in free credit to start.
    Try Cloud Run Free
  • 1
    progrep

    progrep

    Utility to show live progress, status & stats for running simulations

    ...progrep supports both single-threaded and parallel (multicore/multinode - e.g. OpenMP/MPI) jobs. progrep can also report for jobs running on remote hosts, e.g. jobs running on Linux Clusters. progrep works in client-server model. The server can be installed in your simulation source code with only 4 extra lines (API). When the progrep command is invoked, it queries this server as a client. progrep does not interfere with or slow down your simulation when not invoked. Even when invoked, the overhead is insignificant. Out of the box, the API works with Fortran/C/C++ code. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB