32 projects for "loading" with 2 filters applied:

  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Fully Managed MySQL, PostgreSQL, and SQL Server Icon
    Fully Managed MySQL, PostgreSQL, and SQL Server

    Automatic backups, patching, replication, and failover. Focus on your app, not your database.

    Cloud SQL handles your database ops end to end, so you can focus on your app.
    Try Free
  • 1
    Scene Framework

    Scene Framework

    Android Single Activity Framework compatible with Fragment

    ...The repository may aim to help developers manage complex state, transitions, or UI/navigation flows in applications, or provide a structured framework for rendering or content pipelines. As such, Scene could serve as a backbone for media apps, games, or modular user interface projects, offering abstractions for scene management, resource loading, event dispatching, and lifecycle handling. Because of its open-source status, developers may adapt it to their own use-cases — customizing rendering, UI transitions, or content management — though the maturity and documentation may vary.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    Gemma in PyTorch

    Gemma in PyTorch

    The official PyTorch implementation of Google's Gemma models

    gemma_pytorch provides the official PyTorch reference for running and fine-tuning Google’s Gemma family of open models. It includes model definitions, configuration files, and loading utilities for multiple parameter scales, enabling quick evaluation and downstream adaptation. The repository demonstrates text generation pipelines, tokenizer setup, quantization paths, and adapters for low-rank or parameter-efficient fine-tuning. Example notebooks walk through instruction tuning and evaluation so teams can benchmark and iterate rapidly. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    PyTorch3D

    PyTorch3D

    PyTorch3D is FAIR's library of reusable components for deep learning

    ...The library provides fast GPU-accelerated implementations of rendering pipelines, transformations, rasterization, and lighting—making it possible to compute gradients through full 3D rendering processes. Researchers use it for tasks like shape generation, reconstruction, view synthesis, and visual reasoning. PyTorch3D also includes utilities for loading, transforming, and sampling 3D assets, so models can be trained end-to-end from 2D supervision or partial data. Its modular design allows easy extension—components like differentiable rasterizers, mesh blending, or signed distance field (SDF) modules can be swapped or combined to test new architectures quickly.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 4
    Ruler AI

    Ruler AI

    Centralize and sync AI coding rules across tools and projects

    ...Instead of maintaining separate configuration files for tools like GitHub Copilot, Claude, or Cursor, it stores all rules in a single .ruler/ directory and distributes them automatically. This reduces duplication, avoids inconsistent outputs, and keeps guidance aligned as projects evolve. Ruler supports nested rule loading, allowing teams to define context-specific instructions for different parts of a codebase. It also manages MCP server settings, automates .gitignore updates, and provides simple commands to initialize, apply, and revert configurations. Designed for teams using multiple AI agents, it improves workflow consistency, simplifies onboarding, and ensures all assistants follow the same standards across projects.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • 5
    FLUX.1

    FLUX.1

    Official inference repo for FLUX.1 models

    ...The project is part of a larger family of FLUX models developed by Black Forest Labs, designed to produce high-quality, detailed visuals from text descriptions with competitive prompt adherence and artistic fidelity. This repo focuses on running the open-source model variants efficiently, providing scripts, model loading logic, and examples for local installations, and supports integration with Python toolchains like PyTorch and popular generative pipelines. Users can launch CLI tools to generate images, experiment with different FLUX variants, and extend the base code for research-oriented applications.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 6
    HunyuanImage-3.0

    HunyuanImage-3.0

    A Powerful Native Multimodal Model for Image Generation

    ...The model is intended to be competitive with closed-source image generation systems, aiming for high fidelity, prompt adherence, fine detail, and even “world knowledge” reasoning (i.e. leveraging context, semantics, or common sense in generation). The GitHub repo includes code, scripts, model loading instructions, inference utilities, prompt handling, and integration with standard ML tooling (e.g. Hugging Face / Transformers).
    Downloads: 11 This Week
    Last Update:
    See Project
  • 7
    ChatGLM-6B

    ChatGLM-6B

    ChatGLM-6B: An Open Bilingual Dialogue Language Model

    ChatGLM-6B is an open bilingual (Chinese + English) conversational language model based on the GLM architecture, with approximately 6.2 billion parameters. The project provides inference code, demos (command line, web, API), quantization support for lower memory deployment, and tools for finetuning (e.g., via P-Tuning v2). It is optimized for dialogue and question answering with a balance between performance and deployability in consumer hardware settings. Support for quantized inference...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 8
    LLM Scraper

    LLM Scraper

    Extract structured data from webpages using LLM-powered scraping

    LLM Scraper is a TypeScript library designed to extract structured data from webpages using large language models. Instead of relying on fragile HTML selectors or manual parsing rules, the tool interprets webpage content with language models and converts it into structured data according to a defined schema. Developers can specify the data structure using tools such as Zod or JSON Schema, enabling the model to extract relevant information directly into typed objects. LLM Scraper integrates...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 9
    wllama

    wllama

    WebAssembly binding for llama.cpp - Enabling on-browser LLM inference

    wllama is a WebAssembly-based library that enables large language model inference directly inside a web browser. Built as a binding for the llama.cpp inference engine, the project allows developers to run LLM models locally without requiring a server backend or dedicated GPU hardware. The library leverages WebAssembly SIMD capabilities to achieve efficient execution within modern browsers while maintaining compatibility across platforms. By running models locally on the user’s device, wllama...
    Downloads: 1 This Week
    Last Update:
    See Project
  • Full-stack observability with actually useful AI | Grafana Cloud Icon
    Full-stack observability with actually useful AI | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 10
    VLMEvalKit

    VLMEvalKit

    Open-source evaluation toolkit of large multi-modality models (LMMs)

    ...Instead of requiring complex data preparation pipelines or multiple repositories for each benchmark, the system enables evaluation through simple commands that automatically handle dataset loading, model inference, and metric computation. VLMEvalKit supports generation-based evaluation methods, allowing models to produce textual responses to visual inputs while measuring performance through techniques such as exact matching or language-model-assisted answer extraction.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    mosaicml composer

    mosaicml composer

    Supercharge Your Model Training

    ...At the center of the project is a highly optimized Trainer abstraction that simplifies the management of training loops, parallelization, metrics, logging, and data loading. The framework is intended for modern workloads that may span anything from a single GPU to very large distributed training environments, which makes it suitable for both experimentation and production-scale development. It includes built-in support for distributed training strategies such as Fully Sharded Data Parallelism and standard Distributed Data Parallel execution, helping teams scale models without having to assemble as much infrastructure by hand.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    AirLLM

    AirLLM

    AirLLM 70B inference with single 4GB GPU

    AirLLM is an open source Python library that enables extremely large language models to run on consumer hardware with very limited GPU memory. The project addresses one of the main barriers to local LLM experimentation by introducing a memory-efficient inference technique that loads model layers sequentially rather than storing the entire model in GPU memory. This layer-wise inference approach allows models with tens of billions of parameters to run on devices with only a few gigabytes of...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Web Quality Skills

    Web Quality Skills

    Agent Skills for optimizing web quality based on Lighthouse

    ...These skills are framework-agnostic, meaning they apply to React, Vue, Svelte, Angular, Astro, or even plain HTML projects. For example, an agent can use these skills to audit a page’s performance, identify bottlenecks in loading speed, fix layout shift issues, suggest accessibility enhancements, or recommend SEO improvements. The guidance is designed to be triggered by typical developer requests related to web quality and to produce actionable suggestions rooted in well-established standards.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Granite TSFM

    Granite TSFM

    Foundation Models for Time Series

    granite-tsfm collects public notebooks, utilities, and serving components for IBM’s Time Series Foundation Models (TSFM), giving practitioners a practical path from data prep to inference for forecasting and anomaly-detection use cases. The repository focuses on end-to-end workflows: loading data, building datasets, fine-tuning forecasters, running evaluations, and serving models. It documents the currently supported Python versions and points users to where the core TSFM models are hosted and how to wire up service components. Issues and examples in the tracker illustrate common tasks such as slicing inference windows or using pipeline helpers that return pandas DataFrames, grounding the library in day-to-day time-series operations. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    MARS5

    MARS5

    MARS5 speech model (TTS) from CAMB.AI

    MARS5-TTS is CAMB.AI’s open-source English speech model designed for high-quality text-to-speech and voice emulation. It uses a two-stage architecture that combines an autoregressive (AR) model with a non-autoregressive (NAR) model, giving it both expressiveness and speed. The model is built to handle prosodically challenging content such as sports commentary, anime dialogue, and other high-energy or highly varied speech patterns with realistic rhythm and intonation. To control speaker...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Deep Learning Models

    Deep Learning Models

    A collection of various deep learning architectures, models, and tips

    This repository collects clear, well-documented implementations of deep learning models and training utilities written by Sebastian Raschka. The code favors readability and pedagogy: components are organized so you can trace data flow through layers, losses, optimizers, and evaluation. Examples span fundamental architectures—MLPs, CNNs, RNN/Transformers—and practical tasks like image classification or text modeling. Reproducible training scripts and configuration files make it...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    TensorFlow Hub

    TensorFlow Hub

    A library for transfer learning by reusing parts of TensorFlow models

    TensorFlow Hub is a repository that provides a library and platform for publishing, discovering, and reusing pre-trained machine learning models built with TensorFlow. The project enables developers to integrate high-quality models into their applications without needing to train them from scratch. Through TensorFlow Hub, researchers and practitioners can share reusable model components such as image classifiers, text embedding models, and object detection networks. These models can be...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    RAGxplorer

    RAGxplorer

    Open-source tool to visualise your RAG

    RAGxplorer is an open-source visualization tool designed to help developers analyze and understand Retrieval-Augmented Generation (RAG) pipelines. Retrieval-augmented generation combines language models with external document retrieval systems in order to produce more accurate and grounded responses. However, RAG systems can be complex because they involve multiple components such as embedding models, vector databases, and retrieval algorithms. RAGxplorer provides visual tools that allow...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    Grok-1

    Grok-1

    Open-source, high-performance Mixture-of-Experts large language model

    ...In March 2024, xAI released Grok-1's model weights and architecture under the Apache 2.0 license, making them openly accessible to developers. The accompanying GitHub repository provides JAX example code for loading and running the model. Due to its substantial size, utilizing Grok-1 requires a machine with significant GPU memory. The repository's MoE layer implementation prioritizes correctness over efficiency, avoiding the need for custom kernels. This is a full repo snapshot ZIP file of the Grok-1 code.
    Leader badge
    Downloads: 34 This Week
    Last Update:
    See Project
  • 20
    Mixtral offloading

    Mixtral offloading

    Run Mixtral-8x7B models in Colab or consumer desktops

    ...This approach takes advantage of the sparse activation properties of mixture-of-experts architectures, where only a subset of expert networks are used for each token during generation. By selectively loading and caching the required experts, the system avoids keeping the entire model in GPU memory at once. The repository includes notebooks and code examples that demonstrate how to run large language models on consumer hardware such as personal GPUs or cloud notebook environments.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 21
    LLaMA-MoE

    LLaMA-MoE

    Building Mixture-of-Experts from LLaMA with Continual Pre-training

    ...The project is not just a model release, but also a research framework that includes multiple expert construction methods, several gating strategies, and tooling for continual pre-training on filtered SlimPajama-based datasets. It also emphasizes training efficiency through features such as FlashAttention-v2 integration and fast streaming dataset loading, which are important for large-scale experimentation.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    LLM Applications

    LLM Applications

    A comprehensive guide to building RAG-based LLM applications

    LLM Applications is a practical reference repository that demonstrates how to build production-grade applications powered by large language models. The project focuses particularly on Retrieval-Augmented Generation architectures, which combine language models with external knowledge sources to improve accuracy and reliability. It provides step-by-step guidance for constructing systems that ingest documents, split them into chunks, generate embeddings, index them in vector databases, and...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    Video Pre-Training

    Video Pre-Training

    Learning to Act by Watching Unlabeled Online Videos

    ...The repository contains demonstration models of different widths, fine-tuned variants (e.g. for building houses or early-game tasks), and inference scripts that instantiate agents from pretrained weights. Key modules include the behavioral cloning logic, the agent wrapper, and data loading pipelines (with an accessible skeleton for loading Minecraft demonstration data). The repo also includes a run_agent.py script for testing an agent interactively, and an agent.py module encapsulating the control logic.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    tf2_course

    tf2_course

    Notebooks for my "Deep Learning with TensorFlow 2 and Keras" course

    tf2_course provides the notebooks for the “Deep Learning with TensorFlow 2 and Keras” course authored by the same author, Aurélien Géron. It is structured as a teaching toolkit: you’ll find notebooks covering neural networks with Keras, lower-level TensorFlow APIs, data loading & preprocessing, convolutional and recurrent networks, and deployment/distribution of models. The material is intended for learners who already have foundational knowledge of ML and wish to deepen their understanding of deep learning frameworks and practices. The repo supports experimentation: you can run code, tweak hyperparameters, and follow guided exercises that strengthen practical mastery. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25

    Agentopia

    Java5 mobile agents in peer2peer containers without stubs/skeletons.

    Agentopia is a programming framework (API) for Java 5 mobile agents in peer-to-peer networks. Main features: Routing around firewalls, anonymity, and it is extremely easy to write new agents. No RMI, no CORBA, just plain Java bytecode loading.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • Next
MongoDB Logo MongoDB