Showing 6 open source projects for "inference"

View related business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • 1
    DALI

    DALI

    A GPU-accelerated library containing highly optimized building blocks

    ...Deep learning applications require complex, multi-stage data processing pipelines that include loading, decoding, cropping, resizing, and many other augmentations. These data processing pipelines, which are currently executed on the CPU, have become a bottleneck, limiting the performance and scalability of training and inference. DALI addresses the problem of the CPU bottleneck by offloading data preprocessing to the GPU. Additionally, DALI relies on its own execution engine, built to maximize the throughput of the input pipeline.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 2
    SageMaker MXNet Inference Toolkit

    SageMaker MXNet Inference Toolkit

    Toolkit for allowing inference and serving with MXNet in SageMaker

    SageMaker MXNet Inference Toolkit is an open-source library for serving MXNet models on Amazon SageMaker. This library provides default pre-processing, predict and postprocessing for certain MXNet model types and utilizes the SageMaker Inference Toolkit for starting up the model server, which is responsible for handling inference requests. AWS Deep Learning Containers (DLCs) are a set of Docker images for training and serving models in TensorFlow, TensorFlow 2, PyTorch, and MXNet. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    SageMaker TensorFlow Serving Container

    SageMaker TensorFlow Serving Container

    A TensorFlow Serving solution for use in SageMaker

    ...You can also run your container locally in Docker to test different models and input inference requests by hand.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Caramel

    Caramel

    Functional language for building type-safe applications

    ...Caramel leverages the OCaml compiler, to provide you with a pragmatic type system and industrial-strength type safety, and the Erlang VM, known for running low-latency, distributed, and fault-tolerant systems used in a wide range of industries. Excellent type inference, so you never need to annotate your code. Supports sources in OCaml (and soon Reason syntax too). Caramel aims to make building type-safe concurrent programs a productive and fun experience. Caramel should let anyone with existing OCaml or Reason experience be up and running without having to relearn the entire language. Caramel strives to integrate with the larger ecosystem of BEAM languages, like Erlang, Elixir, Gleam, Purerl, LFE, and Hamler.
    Downloads: 3 This Week
    Last Update:
    See Project
  • Forever Free Full-Stack Observability | Grafana Cloud Icon
    Forever Free Full-Stack Observability | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 5
    Amazon SageMaker Examples

    Amazon SageMaker Examples

    Jupyter notebooks that demonstrate how to build models using SageMaker

    ...It uses the familiar JupyterLab interface and has seamless integration with a variety of deep learning and data science environments and scalable compute resources for training, inference, and other ML operations. Studio offers teams and companies easy on-boarding for their team members, freeing them up from complex systems admin and security processes. Administrators control data access and resource provisioning for their users. Notebook Instances are another option. They have the familiar Jupyter and JuypterLab interfaces that work well for single users, or small teams where users are also administrators. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    The Prolog Expert System Shell (PESS) is a software that generates ES using basically two components: Knowledge Base, used by the ES to guide its decisions making, and Inference Machine, cable of collection the rules and generating new facts.
    Downloads: 15 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB