Showing 574 open source projects for "input-leap"

View related business solutions
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build, govern, and optimize agents and models with Gemini Enterprise Agent Platform.
    Start Free
  • Forever Free Full-Stack Observability | Grafana Cloud Icon
    Forever Free Full-Stack Observability | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 1
    Uncertainty Baselines

    Uncertainty Baselines

    High-quality implementations of standard and SOTA methods

    Uncertainty Baselines is a collection of strong, well-documented training pipelines that make it straightforward to evaluate predictive uncertainty in modern machine learning models. Rather than offering toy scripts, it provides end-to-end recipes—data input, model architectures, training loops, evaluation metrics, and logging—so results are comparable across runs and research groups. The library spans canonical modalities and tasks, from image classification and NLP to tabular problems, with baselines that cover both deterministic and probabilistic approaches. Techniques include deep ensembles, Monte Carlo dropout, temperature scaling, stochastic variational inference, heteroscedastic heads, and out-of-distribution detection workflows. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    HunyuanCustom

    HunyuanCustom

    Multimodal-Driven Architecture for Customized Video Generation

    HunyuanCustom is a multimodal video customization framework by Tencent Hunyuan, aimed at generating customized videos featuring particular subjects (people, characters) under flexible conditions, while maintaining subject/identity consistency. It supports conditioning via image, audio, video, and text, and can perform subject replacement in videos, generate avatars speaking given audio, or combine multiple subject images. The architecture builds on HunyuanVideo, with added modules for...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Changelog CI

    Changelog CI

    Changelog CI is a GitHub Action that enables a project

    Changelog CI is a GitHub Action that enables a project to automatically generate changelogs. Changelog CI can be triggered on pull_request, workflow_dispatch, and any other events that can provide the required inputs. Changelog CI uses python and GitHub API to generate a changelog for a repository. First, it tries to get the latest release from the repository (If available). Then, it checks all the pull requests/commits merged after the last release using the GitHub API. After that, it...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Kaleidoscope-SDK

    Kaleidoscope-SDK

    User toolkit for analyzing and interfacing with Large Language Models

    kaleidoscope-sdk is a Python module used to interact with large language models hosted via the Kaleidoscope service available at: https://github.com/VectorInstitute/kaleidoscope. It provides a simple interface to launch LLMs on an HPC cluster, asking them to perform basic features like text generation, but also retrieve intermediate information from inside the model, such as log probabilities and activations. Users must authenticate using their Vector Institute cluster credentials. This can...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Our Free Plans just got better! | Auth0 Icon
    Our Free Plans just got better! | Auth0

    With up to 25k MAUs and unlimited Okta connections, our Free Plan lets you focus on what you do best—building great apps.

    You asked, we delivered! Auth0 is excited to expand our Free and Paid plans to include more options so you can focus on building, deploying, and scaling applications without having to worry about your security. Auth0 now, thank yourself later.
    Try free now
  • 5
    GLM-4.6V

    GLM-4.6V

    GLM-4.6V/4.5V/4.1V-Thinking, towards versatile multimodal reasoning

    GLM-4.6V represents the latest generation of the GLM-V family and marks a major step forward in multimodal AI by combining advanced vision-language understanding with native “tool-call” capabilities, long-context reasoning, and strong generalization across domains. Unlike many vision-language models that treat images and text separately or require intermediate conversions, GLM-4.6V allows inputs such as images, screenshots or document pages directly as part of its reasoning pipeline — and...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 6
    UNO

    UNO

    A Universal Customization Method for Single and Multi Conditioning

    UNO is a project by ByteDance introduced in 2025, titled “A Universal Customization Method for Both Single and Multi-Subject Conditioning.” It suggests a framework for image (or more general generative) modeling where the model can be conditioned either on a single subject or multiple subjects — which may correspond to generating or customizing images featuring specific people, styles, or objects, possibly with fine-grained control over subject identity or composition. Because the project is...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 7
    Qwen2-Audio

    Qwen2-Audio

    Repo of Qwen2-Audio chat & pretrained large audio language model

    ...It is trained to accept various audio signal inputs (including speech, sounds, etc.) and perform both voice chat and audio analysis, producing textual responses. It supports two major modes: Voice Chat (interactive voice only input) and Audio Analysis (audio + text instructions), with both base and instruction-tuned models. It is evaluated on many benchmarks (speech recognition, translation, sound classification, emotion, etc.), and offers pretrained models (e.g. 7B) released via ModelScope and Hugging Face. Code & examples provided with Hugging Face transformers, and usage via AutoProcessor, model classes etc. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 8
    website-to-gif

    website-to-gif

    Turn your website into a GIF

    ...In your GitHub repo, create a workflow file or extend an existing one. You have to also include a step to checkout and commit to the repo. You can use the following example gif.yml. Make sure to modify the url value and add any other input you want to use. WebP rendering will take a lot of time to benefit from lossless quality and file size optimization.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 9
    AG-UI

    AG-UI

    The Agent-User Interaction Protocol

    ...Instead of treating an AI agent as a black-box chat endpoint, AG-UI defines structured events for messages, tool calls, state changes, lifecycle updates, and user interactions. This makes it easier for developers to build agent-powered apps that stream progress, request human input, update UI state, and coordinate complex workflows. The project is especially useful for teams building copilots, workflow assistants, multi-agent products, or custom AI interfaces. Overall, AG-UI provides a shared communication layer between autonomous systems and the interfaces where people actually use them.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 10
    One-Lin3r

    One-Lin3r

    Gives you one-liners that aids in penetration testing operations

    ...It is designed to be cross-platform, supporting environments like Windows, Linux, macOS, and BSD systems, making it highly versatile for different penetration testing scenarios. One of its standout aspects is its intelligent autocomplete system, which not only predicts commands but also corrects user input errors and suggests the closest matching commands.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Conversational Health Agents (CHA)

    Conversational Health Agents (CHA)

    A Personalized LLM-powered Agent Frameworks

    ...It integrates orchestration capabilities that allow the agent to gather information from APIs, knowledge bases, and external services in order to generate more accurate and context-aware responses. The framework supports modular components such as planning, tool execution, and multimodal input processing, which makes it suitable for complex healthcare applications. It also includes a web-based interface for interacting with the agent, making it accessible for testing and deployment in real-world scenarios.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    deepjazz

    deepjazz

    Deep learning driven jazz generation using Keras & Theano

    ...It uses the Keras and Theano libraries to build a two-layer Long Short-Term Memory network capable of learning temporal patterns in music. The system analyzes musical sequences from an input MIDI file and then generates new musical notes that follow similar stylistic patterns. The project was originally created during a hackathon and was designed to show how neural networks can emulate creative tasks traditionally associated with human musicians. The repository includes preprocessing scripts for preparing MIDI data, training scripts for building the neural network model, and code for generating new compositions.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    verl-agent

    verl-agent

    Designed for training LLM/VLM agents via RL

    verl-agent is an open-source reinforcement learning framework designed to train large language model agents and vision-language model agents for complex interactive environments. Built as an extension of the veRL reinforcement learning infrastructure, the project focuses on enabling scalable training for agents that perform multi-step reasoning and decision-making tasks. The framework supports multi-turn interactions between agents and their environments, allowing the system to receive...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Ling-V2

    Ling-V2

    Ling-V2 is a MoE LLM provided and open-sourced by InclusionAI

    Ling-V2 is an open-source family of Mixture-of-Experts (MoE) large language models developed by the InclusionAI research organization with the goal of combining state-of-the-art performance, efficiency, and openness for next-generation AI applications. It introduces highly sparse architectures where only a fraction of the model’s parameters are activated per input token, enabling models like Ling-mini-2.0 to achieve reasoning and instruction-following capabilities on par with much larger dense models while remaining significantly more computationally efficient. Trained on more than 20 trillion tokens of high-quality data and enhanced through multi-stage supervised fine-tuning and reinforcement learning, Ling-V2’s models demonstrate strong general reasoning, mathematical problem-solving, coding understanding, and knowledge-intensive task performance.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    LLM TLDR

    LLM TLDR

    95% token savings. 155x faster queries. 16 languages

    LLM TLDR is a tool that leverages large language models (LLMs) to generate concise, coherent summaries (TL;DRs) of long documents, articles, or text files, helping users quickly understand large amounts of content without reading every word. It integrates with LLM APIs to handle input texts of varying lengths and complexity, applying techniques like chunking, context management, and multi-pass summarization to preserve accuracy even when the source is very large. The system supports both extractive and abstractive summarization styles so that users can choose whether they want condensed highlights or a more narrative paraphrase of key ideas. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    EKS Best Practices

    EKS Best Practices

    A best practices guide for day 2 operations

    The Amazon EKS Best Practices Guide is a public repository containing comprehensive documentation and guidance for operating production-grade Kubernetes clusters on AWS’s managed service, Amazon EKS. Rather than a code library, it serves as a reference catalogue of patterns, anti-patterns, checklists and architectures across domains such as security, reliability, scalability, networking, cost optimization and hybrid cloud deployments. The repository is maintained by AWS but open to...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Flax

    Flax

    Flax is a neural network library for JAX

    ...The library is widely used in vision, language, and reinforcement learning, often serving as a thin layer atop NumPy-like JAX primitives. Tutorials and examples show patterns for multi-host training, mixed precision, and advanced input pipelines that scale from laptops to TPUs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    4M

    4M

    4M: Massively Multimodal Masked Modeling

    ...The repository releases code and models for multiple variants (e.g., 4M-7 and 4M-21), emphasizing transfer to unseen tasks and modalities. Training/inference configs and issues discuss things like depth tokenizers, input masks for generation, and CUDA build questions, signaling active research iteration. The design leans into flexibility and steerability, so prompts and masks can shape behavior without bespoke heads per task. In short, 4M provides a unified recipe to pretrain large multimodal models that generalize broadly while remaining practical to fine-tune.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    CycleGAN and pix2pix in PyTorch

    CycleGAN and pix2pix in PyTorch

    Image-to-Image Translation in PyTorch

    CycleGAN and pix2pix in PyTorch repository is a PyTorch implementation of two influential image-to-image translation frameworks: CycleGAN (for unpaired translation) and pix2pix (for paired translation). This repo gives developers and researchers a convenient, modern (PyTorch-based) platform to train and test these methods — supporting both paired datasets (input to output) and unpaired datasets (domain-to-domain) with minimal changes. The code supports standard training and inference pipelines, and as of recent updates, compatibility with the latest Python and PyTorch versions (e.g. Python 3.11, PyTorch 2.4) as well as support for distributed/multi-GPU training for scalable workflows. Because of its flexibility, users can apply it to many tasks: e.g. style transfer between domains (e.g. season changes, art-to-photo, etc.), mapping sketches/edges to real images, image colorization, day-to-night, photo enhancement, and more.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    Segmentation Models

    Segmentation Models

    Segmentation models with pretrained backbones. PyTorch

    ...Pytorch Image Models (a.k.a. timm) has a lot of pretrained models and interface which allows using these models as encoders in smp, however, not all models are supported. Input channels parameter allows you to create models, which process tensors with an arbitrary number of channels.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    Pytorch-toolbelt

    Pytorch-toolbelt

    PyTorch extensions for fast R&D prototyping and Kaggle farming

    A pytorch-toolbelt is a Python library with a set of bells and whistles for PyTorch for fast R&D prototyping and Kaggle farming. Easy model building using flexible encoder-decoder architecture. Modules: CoordConv, SCSE, Hypercolumn, Depthwise separable convolution and more. GPU-friendly test-time augmentation TTA for segmentation and classification. GPU-friendly inference on huge (5000x5000) images. Every-day common routines (fix/restore random seed, filesystem utils, metrics). Losses:...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    Matrix

    Matrix

    Multi-Agent daTa geneRation Infra and eXperimentation framework

    Matrix is a distributed, large-scale engine for multi-agent synthetic data generation and experiments: it provides the infrastructure to run thousands of “agentic” workflows concurrently (e.g. multiple LLMs interacting, reasoning, generating content, data-processing pipelines) by leveraging distributed computing (like Ray + cluster management). The idea is to treat data generation as a “data-to-data” transformation: each input item defines a task, and the runtime orchestrates asynchronous, peer-to-peer agent workflows, avoiding global synchronization bottlenecks. That design makes Matrix particularly well-suited for large-batch inference, model benchmarking, data curation, augmentation, or generation — whether for language, code, dialogue, or multimodal tasks. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    shuyuan

    shuyuan

    Reading book source

    ...For learners, researchers, or avid readers, Shuyuan offers a way to bridge from plain text files or eBooks into a manageable, interactive resource — one where notes, references, and reading progress can be tracked. It likely supports different input formats (text, HTML, PDF), and may integrate optional translation or text normalization tools.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    Raster Vision

    Raster Vision

    Open source framework for deep learning satellite and aerial imagery

    ...Raster Vision allows engineers to quickly and repeatably configure pipelines that go through core components of a machine learning workflow: analyzing training data, creating training chips, training models, creating predictions, evaluating models, and bundling the model files and configuration for easy deployment. The input to a Raster Vision pipeline is a set of images and training data, optionally with Areas of Interest (AOIs) that describe where the images are labeled. The output of a Raster Vision pipeline is a model bundle that allows you to easily utilize models in various deployment scenarios.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Step1X-3D

    Step1X-3D

    High-Fidelity and Controllable Generation of Textured 3D Assets

    ...It combines a hybrid architecture: a geometry generation stage using a VAE-DiT model to output a watertight 3D representation (e.g. TSDF surface), and a texture synthesis stage that conditions on geometry and optionally reference input (or prompts) to produce view-consistent textures using a diffusion-based texture module. The result is fully 3D assets — meshes + textures — which can be rendered from any viewpoint, textured consistently, and used in 3D applications. To achieve this, the project includes a massive curated dataset: among more than 5 million candidate 3D assets, it filters and standardizes to produce a high-quality 2 million–asset subset suitable for training.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB