Showing 1 open source project for "inference"

View related business solutions
  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    More flexibility. More control.

    Generate interest, access liquidity without selling, and execute trades seamlessly. All in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 1
    PyTorch Lightning

    PyTorch Lightning

    The lightweight PyTorch wrapper for high-performance AI research

    ...It's designed to decouple the science from the engineering in your PyTorch code, simplifying complex network coding and giving you maximum flexibility. PyTorch Lightning can be used for just about any type of research, and was built for the fast inference needed in AI research and production. When you need to scale up things like BERT and self-supervised learning, Lightning responds accordingly by automatically exporting to ONNX or TorchScript. PyTorch Lightning can easily be applied for any use case. With just a quick refactor you can run your code on any hardware, run distributed training, perform logging, metrics, visualization and so much more!
    Downloads: 10 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB