Showing 4 open source projects for "static code analysis"

View related business solutions
  • Go From Idea to Deployed AI App Fast Icon
    Go From Idea to Deployed AI App Fast

    One platform to build, fine-tune, and deploy. No MLOps team required.

    Access Gemini 3 and 200+ models. Build chatbots, agents, or custom models with built-in monitoring and scaling.
    Try Free
  • Cut Cloud Costs with Google Compute Engine Icon
    Cut Cloud Costs with Google Compute Engine

    Save up to 91% with Spot VMs and get automatic sustained-use discounts. One free VM per month, plus $300 in credits.

    Save on compute costs with Compute Engine. Reduce your batch jobs and workload bill 60-91% with Spot VMs. Compute Engine's committed use offers customers up to 70% savings through sustained use discounts. Plus, you get one free e2-micro VM monthly and $300 credit to start.
    Try Compute Engine
  • 1
    JPPF

    JPPF

    The open source grid computing solution

    JPPF makes it easy to parallelize computationally intensive tasks and execute them on a Grid.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 2
    NodeCom is an open-source component framework for extremely constrained WSN. This project sourceforge is under construction. The code will be added soon. See our laboratory web page: http://domus.usherbrooke.ca/
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Poor Man's HPC is a framework that allows distributing and running code on a server farm. pmHPC is a scaled down and simplified version of distributed computing projects such as SETI, so is a perfect fit for enthusiasts and universities.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Sight provides a friendly interface to create and connect agents for bioinformatics. The workflow supports multiple responses to a single request, structure transforms, filters and request history access.Can talk with ordinary http servers (get and post)
    Downloads: 0 This Week
    Last Update:
    See Project
  • Host LLMs in Production With On-Demand GPUs Icon
    Host LLMs in Production With On-Demand GPUs

    NVIDIA L4 GPUs. 5-second cold starts. Scale to zero when idle.

    Deploy your model, get an endpoint, pay only for compute time. No GPU provisioning or infrastructure management required.
    Try Free
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB