Showing 13 open source projects for "inference"

View related business solutions
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • AI-generated apps that pass security review Icon
    AI-generated apps that pass security review

    Stop waiting on engineering. Build production-ready internal tools with AI—on your company data, in your cloud.

    Retool lets you generate dashboards, admin panels, and workflows directly on your data. Type something like “Build me a revenue dashboard on my Stripe data” and get a working app with security, permissions, and compliance built in from day one. Whether on our cloud or self-hosted, create the internal software your team needs without compromising enterprise standards or control.
    Try Retool free
  • 1
    whisper.cpp

    whisper.cpp

    Port of OpenAI's Whisper model in C/C++

    ...The entire high-level implementation of the model is contained in whisper.h and whisper.cpp. The rest of the code is part of the ggml machine learning library. The command downloads the base.en model converted to custom ggml format and runs the inference on all .wav samples in the folder samples. whisper.cpp supports integer quantization of the Whisper ggml models. Quantized models require less memory and disk space and depending on the hardware can be processed more efficiently.
    Downloads: 372 This Week
    Last Update:
    See Project
  • 2
    Genv

    Genv

    GPU environment management and cluster orchestration

    Genv is an open-source environment and cluster management system for GPUs. Genv lets you easily control, configure, monitor and enforce the GPU resources that you are using in a GPU machine or cluster. It is intended to ease up the process of GPU allocation for data scientists without code changes.
    Downloads: 19 This Week
    Last Update:
    See Project
  • 3
    WiFi DensePose

    WiFi DensePose

    Turn WiFi signals into real-time human pose estimation and detection

    ...It is designed to showcase the emerging field of RF-based sensing, where machine learning models interpret wireless channel data to reconstruct human movement and posture. The repository includes components for data processing, model inference, and real-time visualization, making it suitable for research and experimental deployments. Its architecture emphasizes performance and reproducibility, allowing developers to explore non-visual motion capture systems using accessible hardware. Overall, WiFi DensePose functions as an advanced research-grade toolkit for WiFi-based human sensing and pose estimation.
    Downloads: 53 This Week
    Last Update:
    See Project
  • 4
    Arize Phoenix

    Arize Phoenix

    Uncover insights, surface problems, monitor, and fine tune your LLM

    Phoenix provides ML insights at lightning speed with zero-config observability for model drift, performance, and data quality. Phoenix is an Open Source ML Observability library designed for the Notebook. The toolset is designed to ingest model inference data for LLMs, CV, NLP and tabular datasets. It allows Data Scientists to quickly visualize their model data, monitor performance, track down issues & insights, and easily export to improve. Deep Learning Models (CV, LLM, and Generative) are an amazing technology that will power many of future ML use cases. A large set of these technologies are being deployed into businesses (the real world) in what we consider a production setting.
    Downloads: 2 This Week
    Last Update:
    See Project
  • Go from Code to Production URL in Seconds Icon
    Go from Code to Production URL in Seconds

    Cloud Run deploys apps in any language instantly. Scales to zero. Pay only when code runs.

    Skip the Kubernetes configs. Cloud Run handles HTTPS, scaling, and infrastructure automatically. Two million requests free per month.
    Try it free
  • 5
    MiniSearch

    MiniSearch

    Minimalist web-searching platform with an AI assistant

    MiniSearch is a minimalist web search application with a built-in AI assistant that runs largely inside the browser for privacy-focused information retrieval. The project combines metasearch capabilities with local or remote language model inference to provide conversational answers alongside traditional search results. It is designed to be lightweight, easy to deploy with Docker, and configurable for both personal and hosted use cases. The platform supports browser-level integration so users can set it as their default search engine for quick access. Its architecture emphasizes privacy by avoiding tracking and minimizing data collection while still enabling advanced AI features. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 6
    OnnxStream

    OnnxStream

    Lightweight inference library for ONNX files, written in C++

    The challenge is to run Stable Diffusion 1.5, which includes a large transformer model with almost 1 billion parameters, on a Raspberry Pi Zero 2, which is a microcomputer with 512MB of RAM, without adding more swap space and without offloading intermediate results on disk. The recommended minimum RAM/VRAM for Stable Diffusion 1.5 is typically 8GB. Generally, major machine learning frameworks and libraries are focused on minimizing inference latency and/or maximizing throughput, all of which at the cost of RAM usage. So I decided to write a super small and hackable inference library specifically focused on minimizing memory consumption: OnnxStream. OnnxStream is based on the idea of decoupling the inference engine from the component responsible for providing the model weights, which is a class derived from WeightsProvider. ...
    Downloads: 8 This Week
    Last Update:
    See Project
  • 7
    DeepSpeech

    DeepSpeech

    Open source embedded speech-to-text engine

    ...A pre-trained English model is available for use and can be downloaded following the instructions in the usage docs. If you want to use the pre-trained English model for performing speech-to-text, you can download it (along with other important inference material) from the DeepSpeech releases page.
    Downloads: 11 This Week
    Last Update:
    See Project
  • 8
    Albedo

    Albedo

    A recommender system for discovering GitHub repos

    Albedo is an open-source recommender system aimed at helping developers discover GitHub repositories by learning from activity signals. It treats repositories and developers as a graph of interactions and applies large-scale matrix factorization to model affinities, with Apache Spark providing the distributed data processing. The project focuses on implicit feedback—stars, watches, and other engagement metrics—so it can build useful recommendations without explicit ratings. A reproducible...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    core.typed

    core.typed

    An optional type system for Clojure

    core.typed is an optional static type system for Clojure, providing gradual typing features like type declarations, checking, and inference. It supports modules for development-time checking and runtime typing, though note that it's deprecated as of Clojure 1.11 in favor of Typed Clojure. core.typed is separated into modules. You'll want the full type checker at development time, and the runtime dependencies in production. Then, lein repl will automatically activate the :dev profile. ...
    Downloads: 8 This Week
    Last Update:
    See Project
  • AI-powered service management for IT and enterprise teams Icon
    AI-powered service management for IT and enterprise teams

    Enterprise-grade ITSM, for every business

    Give your IT, operations, and business teams the ability to deliver exceptional services—without the complexity. Maximize operational efficiency with refreshingly simple, AI-powered Freshservice.
    Try it Free
  • 10
    Mandarax is a pure Java implementation of a rule engine. It supports mutiples types of facts and rules based on reflection, databases, EJB etc, supporting XML standards (RuleML 0.8). It provides a J2EE compliant inference engine using backward chaining.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    Self-aware dynamic, reflective, and declarative semantic space representation, transformation, and exploration system.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    SIES is a forward-chaining rule engine server. Clients connect to update a global knowledge-base and/or can register to be notified when specified elements are inferred. Local access for a client's own inferences is also possible.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    The Prolog Expert System Shell (PESS) is a software that generates ES using basically two components: Knowledge Base, used by the ES to guide its decisions making, and Inference Machine, cable of collection the rules and generating new facts.
    Downloads: 1 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB