5 projects for "loading" with 2 filters applied:

  • Fully Managed MySQL, PostgreSQL, and SQL Server Icon
    Fully Managed MySQL, PostgreSQL, and SQL Server

    Automatic backups, patching, replication, and failover. Focus on your app, not your database.

    Cloud SQL handles your database ops end to end, so you can focus on your app.
    Try Free
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 1
    Google AI Edge Gallery

    Google AI Edge Gallery

    A gallery that showcases on-device ML/GenAI use cases

    ...The project bundles runnable samples that show how to run TensorFlow Lite/Edge TPU models (and similar lightweight runtimes) on mobile and embedded platforms, demonstrating common tasks like image classification, object detection, audio recognition, and pose estimation. Each sample is intended to be both a learning aid and a practical starting point: code is organized to show model loading, pre/post-processing, performance measurement, and common optimization knobs (quantization, NNAPI/Delegate usage, and hardware accelerators). The repo also collects small, well-documented models and conversion scripts so developers can reproduce a pipeline from a full-size model down to a device-friendly artifact.
    Downloads: 42 This Week
    Last Update:
    See Project
  • 2
    qvac-fabric-llm.cpp

    qvac-fabric-llm.cpp

    QVAC Fabric: cross-platform LLM inference and fine-tuning

    qvac-fabric-llm.cpp is a cross-platform large language model inference and fine-tuning engine built as an advanced fork of llama.cpp, designed to run efficiently across desktops, mobile devices, and heterogeneous GPU environments. The project focuses on removing hardware limitations traditionally associated with LLM deployment by enabling support for a wide range of backends, including Vulkan, Metal, CUDA, and CPU, making it accessible on devices ranging from smartphones to enterprise...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 3
    LLMFarm

    LLMFarm

    llama and other large language models on iOS and MacOS offline

    ...It enables users to run LLMs on personal hardware or private infrastructure, reducing dependency on external APIs and improving data privacy. The system typically provides a user-friendly interface for loading models, configuring inference parameters, and interacting with them through chat or task-based workflows. It emphasizes modularity, allowing users to integrate different models, backends, or tools depending on their needs and hardware capabilities. LLMFarm is particularly useful for developers and researchers experimenting with local AI systems, as it lowers the barrier to entry for running and testing models without extensive setup. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    KoboldCpp

    KoboldCpp

    Run GGUF models easily with a UI or API. One File. Zero Install.

    KoboldCpp is an easy-to-use AI text-generation software for GGML and GGUF models, inspired by the original KoboldAI. It's a single self-contained distributable that builds off llama.cpp and adds many additional powerful features.
    Leader badge
    Downloads: 270 This Week
    Last Update:
    See Project
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 5
    BoostMultiDex

    BoostMultiDex

    Load multiple dex files on low Android version devices

    BoostMultiDex is a specialized Android-library created by ByteDance to address a long-standing performance and user-experience problem on older Android devices (specifically Android 4.x and below, i.e. SDK < 21), where loading multiple dex files after installation or upgrade can lead to extremely long “black-screen” freeze times during first startup. Instead of relying on the standard Android MultiDex mechanism, BoostMultiDex reworks the loading strategy by tapping into hidden Dalvik VM internals so that original DEX bytecode can be loaded directly, bypassing costly ODEX generation. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB