Showing 54 open source projects for "low vision"

View related business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 8 Monitoring Tools in One APM. Install in 5 Minutes. Icon
    8 Monitoring Tools in One APM. Install in 5 Minutes.

    Errors, performance, logs, uptime, hosts, anomalies, dashboards, and check-ins. One interface.

    AppSignal works out of the box for Ruby, Elixir, Node.js, Python, and more. 30-day free trial, no credit card required.
    Start Free
  • 1
    Open Vision Agents by Stream

    Open Vision Agents by Stream

    Build Vision Agents quickly with any model or video provider

    ...The framework uses Stream’s ultra low latency edge network so agents can join sessions quickly and maintain very low audio and video latency while processing frames and generating responses. Developers work with an agent abstraction that connects video edge providers, LLMs, and processors into pipelines, making it easier to orchestrate tasks like object detection, pose estimation, and conversational guidance.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 2
    Exclusively Dark Image Dataset

    Exclusively Dark Image Dataset

    ExDARK dataset is the largest collection of low-light images

    ...The dataset was created to address the lack of large-scale low-light datasets available for research in object detection, recognition, and enhancement. It has been widely used in studies of low-light image enhancement, deep learning approaches, and domain adaptation for vision models. Researchers can also explore its associated source code for low-light image enhancement tasks, making it an essential resource for advancing work in night-time and low-light visual recognition.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 3
    Phi-3-MLX

    Phi-3-MLX

    Phi-3.5 for Mac: Locally-run Vision and Language Models

    Phi-3-Vision-MLX is an Apple MLX (machine learning on Apple silicon) implementation of Phi-3 Vision, a lightweight multi-modal model designed for vision and language tasks. It focuses on running vision-language AI efficiently on Apple hardware like M1 and M2 chips.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    LRSLibrary

    LRSLibrary

    Low-Rank and Sparse Tools for Background Modeling and Subtraction

    ...The algorithms can also be adapted to other computer vision or machine learning problems beyond video. Large algorithm collection: > 100 matrix- and tensor-based low-rank + sparse methods. Open-source license, documentation and references included.
    Downloads: 1 This Week
    Last Update:
    See Project
  • Our Free Plans just got better! | Auth0 Icon
    Our Free Plans just got better! | Auth0

    With up to 25k MAUs and unlimited Okta connections, our Free Plan lets you focus on what you do best—building great apps.

    You asked, we delivered! Auth0 is excited to expand our Free and Paid plans to include more options so you can focus on building, deploying, and scaling applications without having to worry about your security. Auth0 now, thank yourself later.
    Try free now
  • 5
    Kornia

    Kornia

    Open Source Differentiable Computer Vision Library

    Kornia is a differentiable computer vision library for PyTorch. It consists of a set of routines and differentiable modules to solve generic computer vision problems. At its core, the package uses PyTorch as its main backend both for efficiency and to take advantage of the reverse-mode auto-differentiation to define and compute the gradient of complex functions. Inspired by existing packages, this library is composed by a subset of packages containing operators that can be inserted within neural networks to train models to perform image transformations, epipolar geometry, depth estimation, and low-level image processing such as filtering and edge detection that operate directly on tensors. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    Kimi K2.5

    Kimi K2.5

    Moonshot's most powerful AI model

    Kimi K2.5 is Moonshot AI’s open-source, native multimodal agentic model built through continual pretraining on approximately 15 trillion mixed vision and text tokens. Based on a 1T-parameter Mixture-of-Experts (MoE) architecture with 32B activated parameters, it integrates advanced language reasoning with strong visual understanding. K2.5 supports both “Thinking” and “Instant” modes, enabling either deep step-by-step reasoning or low-latency responses depending on the task. Designed for agentic workflows, it features an Agent Swarm mechanism that decomposes complex problems into coordinated sub-agents executing in parallel. ...
    Downloads: 33 This Week
    Last Update:
    See Project
  • 7
    MiniMind-V

    MiniMind-V

    "Big Model" trains a visual multimodal VLM with 26M parameters

    MiniMind-V is an experimental open-source project that aims to train a very small multimodal vision–language model (VLM) from scratch with extremely low compute and cost, making research and experimentation accessible to more people. The repository showcases training workflows and code designed to produce a 26-million parameter model—including both image and text capabilities—using minimal resources in very little time, reflecting a trend toward democratizing AI research. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    LingBot-VLA

    LingBot-VLA

    A Pragmatic VLA Foundation Model

    ...The model aims to bridge vision, language understanding, and motor control within one unified architecture, making it capable of understanding high-level instructions and generating coherent low-level actions in physical environments. Because LingBot-VLA includes not just the model weights but also a full production-ready codebase with tools for data handling, training, and evaluation, developers can adapt it to custom robots or simulation environments efficiently.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 9
    CleanVision

    CleanVision

    Automatically find issues in image datasets

    CleanVision automatically detects potential issues in image datasets like images that are: blurry, under/over-exposed, (near) duplicates, etc. This data-centric AI package is a quick first step for any computer vision project to find problems in the dataset, which you want to address before applying machine learning. CleanVision is super simple -- run the same couple lines of Python code to audit any image dataset! The quality of machine learning models hinges on the quality of the data used to train them, but it is hard to manually identify all of the low-quality data in a big dataset. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Go From AI Idea to AI App Fast Icon
    Go From AI Idea to AI App Fast

    One platform to build, fine-tune, and deploy ML models. No MLOps team required.

    Access Gemini 3 and 200+ models. Build chatbots, agents, or custom models with built-in monitoring and scaling.
    Try Free
  • 10
    dots.ocr

    dots.ocr

    Multilingual Document Layout Parsing in a Single Vision-Language Model

    dots.ocr is a cutting-edge multilingual document parsing system built on a unified vision-language model that combines layout detection, text recognition, and structural understanding into a single architecture. Unlike traditional OCR pipelines that rely on multiple specialized components, dots.ocr integrates these processes end-to-end, reducing error propagation and improving consistency across tasks. The model is designed to recognize virtually any human script, making it highly effective for global and low-resource language scenarios. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    RuView

    RuView

    Turn WiFi signals into real-time human sensing and spatial awareness.

    ...Built on the concept of WiFi DensePose, it analyzes disturbances in WiFi Channel State Information (CSI) caused by human movement to reconstruct body position, breathing patterns, heart rate, and presence. Unlike traditional vision systems, RuView operates without cameras, wearables, or cloud connectivity, making it a privacy-first sensing solution. The system runs on low-cost hardware such as ESP32 sensor meshes and performs signal processing and machine learning directly at the edge. By learning the RF signature of each environment over time, RuView adapts automatically to different spaces and improves its sensing accuracy. ...
    Downloads: 61 This Week
    Last Update:
    See Project
  • 12
    Magnitude

    Magnitude

    Vision AI browser agent for automation, testing, and extraction

    ...Developers can use it to automate repetitive web tasks, integrate services without APIs, or build advanced browser-based agents. It also provides flexible abstraction levels, allowing both high-level task execution and precise low-level control of actions like mouse movements and keyboard input.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Portkey AI Gateway

    Portkey AI Gateway

    A blazing fast AI Gateway with integrated guardrails

    ...It’s lightweight and optimized for low latency with a small footprint.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 14
    fastai

    fastai

    Deep learning library

    fastai is a deep learning library which provides practitioners with high-level components that can quickly and easily provide state-of-the-art results in standard deep learning domains, and provides researchers with low-level components that can be mixed and matched to build new approaches. It aims to do both things without substantial compromises in ease of use, flexibility, or performance. This is possible thanks to a carefully layered architecture, which expresses common underlying...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    The Optimism Monorepo

    The Optimism Monorepo

    Optimism is Ethereum, scaled

    Optimism is a low-cost and lightning-fast Ethereum L2 blockchain, but it's also so much more than that. Optimism is the technical foundation for the Optimism Collective, a band of communities, companies, and citizens united by a mutually beneficial pact to adhere to the axiom of impact=profit — the principle that positive impact to the collective should be rewarded with profit to the individual.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 16
    SageAttention

    SageAttention

    NeurIPS2025 Spotlight] Quantized Attention

    SageAttention is an open-source optimization library designed to accelerate the attention mechanism used in transformer-based neural networks. Since attention operations are often the most computationally expensive component of modern AI models, SageAttention introduces quantization techniques that significantly reduce computational overhead while preserving model accuracy. The system achieves this by using low-precision numerical formats such as INT4, FP8, or INT8 to represent key matrices...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 17
    FastDeploy

    FastDeploy

    High-performance Inference and Deployment Toolkit for LLMs and VLMs

    FastDeploy is an open-source inference and deployment toolkit designed to simplify the process of running and serving deep learning models across a wide range of hardware platforms. Developed within the PaddlePaddle ecosystem, the toolkit focuses on providing high-performance deployment capabilities for modern AI models including large language models and vision-language systems. The platform enables developers to deploy trained models quickly using optimized inference pipelines that support...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    Cosmos-RL

    Cosmos-RL

    Cosmos-RL is a flexible and scalable Reinforcement Learning framework

    Cosmos-RL is a scalable reinforcement learning framework designed specifically for physical AI systems such as robotics, autonomous agents, and multimodal models. It provides a distributed training architecture that separates policy learning and environment rollout processes, enabling efficient and asynchronous reinforcement learning at scale. The framework supports multiple parallelism strategies, including tensor, pipeline, and data parallelism, allowing it to leverage large GPU clusters...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    CoreNet

    CoreNet

    CoreNet: A library for training deep neural networks

    CoreNet is Apple’s internal deep learning framework for distributed neural network training, designed for high scalability, low-latency communication, and strong hardware efficiency. It focuses on enabling large-scale model training across clusters of GPUs and accelerators by optimizing data flow and parallelism strategies. CoreNet provides abstractions for data, tensor, and pipeline parallelism, allowing models to scale without code duplication or heavy manual configuration.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    Qwen2.5-Omni

    Qwen2.5-Omni

    Capable of understanding text, audio, vision, video

    Qwen2.5-Omni is an end-to-end multimodal flagship model in the Qwen series by Alibaba Cloud, designed to process multiple modalities (text, images, audio, video) and generate responses both as text and natural speech in streaming real-time. It supports “Thinker-Talker” architecture, and introduces innovations for aligning modalities over time (for example synchronizing video/audio), robust speech generation, and low-VRAM/quantized versions to make usage more accessible. It holds...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    FLUX.2

    FLUX.2

    Official inference repo for FLUX.2 models

    ...It supports high-resolution output (up to ~4 megapixels), which allows for photography-quality images, detailed product shots, infographics or UI mockups rather than just low-resolution drafts. FLUX.2 is built with a modern architecture (a flow-matching transformer + a revamped VAE + a strong vision-language encoder), enabling strong prompt adherence, correct rendering of text/typography in images, reliable lighting, layout, and physical realism, and consistent style/character/product identity across multiple generations or edits.
    Downloads: 51 This Week
    Last Update:
    See Project
  • 22
    HomeRobot

    HomeRobot

    Mobile manipulation research tools for roboticists

    HomeRobot is an open source robotics framework developed by Facebook Research (Meta AI) that provides a complete software stack for mobile manipulation in real-world and simulated environments. Designed for low-cost robots such as the Hello Robot Stretch, HomeRobot enables agents to perceive, navigate, and interact with their surroundings through vision and manipulation. The system focuses on Open Vocabulary Mobile Manipulation (OVMM) — a challenging task in which a robot must explore an unknown environment, identify an object by name, locate a suitable receptacle, and place the object appropriately. ...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 23
    Qwen3-VL

    Qwen3-VL

    Qwen3-VL, the multimodal large language model series by Alibaba Cloud

    ...It also brings advanced perception capabilities, including spatial grounding, object recognition, OCR across 32 languages, and robust handling of challenging inputs like low-light or distorted text.
    Downloads: 7 This Week
    Last Update:
    See Project
  • 24
    GLM-4.6V

    GLM-4.6V

    GLM-4.6V/4.5V/4.1V-Thinking, towards versatile multimodal reasoning

    GLM-4.6V represents the latest generation of the GLM-V family and marks a major step forward in multimodal AI by combining advanced vision-language understanding with native “tool-call” capabilities, long-context reasoning, and strong generalization across domains. Unlike many vision-language models that treat images and text separately or require intermediate conversions, GLM-4.6V allows inputs such as images, screenshots or document pages directly as part of its reasoning pipeline — and...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    SteadyDancer

    SteadyDancer

    Harmonized and Coherent Human Image Animation

    SteadyDancer is a research-oriented motion stabilization and dancer tracking system designed to analyze and correct motion in videos, making captured performances appear smoother and more stable while preserving expressiveness. It employs computer vision and motion modeling to estimate and reduce unwanted jitters, shakes, or camera wobbles — particularly in dance or movement sequences where traditional smoothing would distort intentional motion. By differentiating between intentional...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • Next
MongoDB Logo MongoDB