Computer Vision Libraries for BSD

Browse free open source Computer Vision Libraries and projects for BSD below. Use the toggles on the left to filter open source Computer Vision Libraries by OS, license, language, programming language, and project status.

  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • 1
    OpenCV

    OpenCV

    Open Source Computer Vision Library

    The Open Source Computer Vision Library has >2500 algorithms, extensive documentation and sample code for real-time computer vision. It works on Windows, Linux, Mac OS X, Android, iOS in your browser through JavaScript. Languages: C++, Python, Julia, Javascript Homepage: https://opencv.org Q&A forum: https://forum.opencv.org/ Documentation: https://docs.opencv.org Source code: https://github.com/opencv Please pay special attention to our tutorials! https://docs.opencv.org/master Books about the OpenCV are described here: https://opencv.org/books.html
    Leader badge
    Downloads: 2,952 This Week
    Last Update:
    See Project
  • 2
    Armadillo

    Armadillo

    fast C++ library for linear algebra & scientific computing

    * Fast C++ library for linear algebra (matrix maths) and scientific computing * Easy to use functions and syntax, deliberately similar to Matlab / Octave * Uses template meta-programming techniques to increase efficiency * Provides user-friendly wrappers for OpenBLAS, Intel MKL, LAPACK, ATLAS, ARPACK, SuperLU and FFTW libraries * Useful for machine learning, pattern recognition, signal processing, bioinformatics, statistics, finance, etc. * Downloads: http://arma.sourceforge.net/download.html * Documentation: http://arma.sourceforge.net/docs.html * Bug reports: http://arma.sourceforge.net/faq.html * Git repo: https://gitlab.com/conradsnicta/armadillo-code
    Leader badge
    Downloads: 3,085 This Week
    Last Update:
    See Project
  • 3
    COLMAP

    COLMAP

    Structure-from-Motion and Multi-View Stereo

    COLMAP is a general-purpose Structure-from-Motion (SfM) and Multi-View Stereo (MVS) pipeline with a graphical and command-line interface. It offers a wide range of features for the reconstruction of ordered and unordered image collections. The software is licensed under the new BSD license.
    Downloads: 55 This Week
    Last Update:
    See Project
  • 4
    DensePose

    DensePose

    A real-time approach for mapping all human pixels of 2D RGB images

    DensePose is a computer vision system that maps all human pixels in an RGB image to the 3D surface of a human body model. It extends human pose estimation from predicting joint keypoints to providing dense correspondences between 2D images and a canonical 3D mesh (such as the SMPL model). This enables detailed understanding of human shape, motion, and surface appearance directly from images or videos. The repository includes the DensePose network architecture, training code, pretrained models, and dataset tools for annotation and visualization. DensePose is widely used in augmented reality, motion capture, virtual try-on, and visual effects applications because it enables real-time 3D human mapping from 2D inputs. The model architecture builds on Mask R-CNN, using additional regression heads to predict UV coordinates that map image pixels to 3D surfaces.
    Downloads: 6 This Week
    Last Update:
    See Project
  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • 5
    SAM 2

    SAM 2

    The repository provides code for running inference with SAM 2

    SAM2 is a next-generation version of the Segment Anything Model (SAM), designed to improve performance, generalization, and efficiency in promptable image segmentation tasks. It retains the core promptable interface—accepting points, boxes, or masks—but incorporates architectural and training enhancements to produce higher-fidelity masks, better boundary adherence, and robustness to complex scenes. The updated model is optimized for faster inference and lower memory use, enabling real-time interactivity even on larger images or constrained hardware. SAM2 comes with pretrained weights and easy-to-use APIs, enabling developers and researchers to integrate promptable segmentation into annotation tools, vision pipelines, or downstream tasks. The project also includes scripts and notebooks to compare SAM2 against SAM on edge cases, benchmarks showing improvements, and evaluation suites to measure mask quality metrics like IoU and boundary error.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 6
    PIFuHD

    PIFuHD

    High-Resolution 3D Human Digitization from A Single Image

    PIFuHD (Pixel-Aligned Implicit Function for 3D human reconstruction at high resolution) is a method and codebase to reconstruct high-fidelity 3D human meshes from a single image. It extends prior PIFu work by increasing resolution and detail, enabling fine geometry in cloth folds, hair, and subtle surface features. The method operates by learning an implicit occupancy / surface function conditioned on the image and camera projection; at inference time it queries dense points to reconstruct a mesh via marching cubes. It also uses a two-stage architecture: a coarse global model followed by local refinement patches to capture fine detail, balancing global consistency and local detail. The repo includes training pipelines, dataset loaders (for Multi-POP, etc.), and inference scripts for mesh output including depth maps for postprocessing. To help practical use, there are utilities for normal estimation, texture back-projection, mesh cleanup, and integration with rendering pipelines.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 7
    Mobile Robot Programming Toolkit (MRPT)

    Mobile Robot Programming Toolkit (MRPT)

    **MOVED TO GITHUB** ==> https://github.com/MRPT/mrpt

    **MOVED TO GITHUB** ==> https://github.com/MRPT/mrpt The Mobile Robot Programming Toolkit (MRPT) is an extensive, cross-platform, and open source C++ library aimed for robotics researchers to design and implement algorithms about Localization, SLAM, Navigation, computer vision. http://www.mrpt.org/
    Leader badge
    Downloads: 44 This Week
    Last Update:
    See Project
  • 8
    Segment Anything

    Segment Anything

    Provides code for running inference with the SegmentAnything Model

    Segment Anything (SAM) is a foundation model for image segmentation that’s designed to work “out of the box” on a wide variety of images without task-specific fine-tuning. It’s a promptable segmenter: you guide it with points, boxes, or rough masks, and it predicts high-quality object masks consistent with the prompt. The architecture separates a powerful image encoder from a lightweight mask decoder, so the heavy vision work can be computed once and the interactive part stays fast. A bundled automatic mask generator can sweep an image and propose many object masks, which is useful for dataset bootstrapping or bulk annotation. The repository includes ready-to-use weights, Python APIs, and example notebooks demonstrating both interactive and automatic modes. Because SAM was trained with an extremely large and diverse mask dataset, it tends to generalize well to new domains, making it a practical starting point for research and production annotation tools.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 9
    OpenNN - Open Neural Networks Library

    OpenNN - Open Neural Networks Library

    Machine learning algorithms for advanced analytics

    OpenNN is a software library written in C++ for advanced analytics. It implements neural networks, the most successful machine learning method. Some typical applications of OpenNN are business intelligence (customer segmentation, churn prevention…), health care (early diagnosis, microarray analysis…) and engineering (performance optimization, predictive maitenance…). OpenNN does not deal with computer vision or natural language processing. The main advantage of OpenNN is its high performance. This library outstands in terms of execution speed and memory allocation. It is constantly optimized and parallelized in order to maximize its efficiency. The documentation is composed by tutorials and examples to offer a complete overview about the library. OpenNN is developed by Artelnics, a company specialized in artificial intelligence.
    Downloads: 11 This Week
    Last Update:
    See Project
  • AI-powered service management for IT and enterprise teams Icon
    AI-powered service management for IT and enterprise teams

    Enterprise-grade ITSM, for every business

    Give your IT, operations, and business teams the ability to deliver exceptional services—without the complexity. Maximize operational efficiency with refreshingly simple, AI-powered Freshservice.
    Try it Free
  • 10
    Phi-3-MLX

    Phi-3-MLX

    Phi-3.5 for Mac: Locally-run Vision and Language Models

    Phi-3-Vision-MLX is an Apple MLX (machine learning on Apple silicon) implementation of Phi-3 Vision, a lightweight multi-modal model designed for vision and language tasks. It focuses on running vision-language AI efficiently on Apple hardware like M1 and M2 chips.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    The Integrating Vision Toolkit (IVT) is a powerful and fast C++ computer vision library with an easy-to-use object-oriented architecture. It offers its own multi-platform GUI toolkit. OpenCV is integrated optionally. Website: http://ivt.sourceforge.net
    Downloads: 5 This Week
    Last Update:
    See Project
  • 12
    A multi-platform collection of C++ software libraries for Computer Vision and Image Understanding.
    Leader badge
    Downloads: 2 This Week
    Last Update:
    See Project
  • 13
    Panzer Combat II

    Panzer Combat II

    Computer-assisted miniature tank game.

    Panzer Combat II is a multi-player voice and webcam enabled computer-assisted distributed miniature wargame of World War II tank combat. Firing is done by placing a webcam behind the aiming unit. Distance to target is computed using computer vision. Action inside the tanks is performed on the computer screen while battlefield strategy is played on the miniature terrain. Both camps can use a different laptop or tablet, the game will interconnect. You can try it online : http://server.panzercombat.com/PCII_Web/move.htm Look at battle reports : http://www.flickr.com/photos/panzercombatii Or watch a demo : http://www.youtube.com/watch?v=WcjfV8Odtss 100% CLEAN : http://games.softpedia.com/progClean/Panzer-Combat-II-Clean-95530.html
    Downloads: 3 This Week
    Last Update:
    See Project
  • 14
    Blazeface

    Blazeface

    Blazeface is a lightweight model that detects faces in images

    Blazeface is a lightweight, high-performance face detection model designed for mobile and embedded devices, developed by TensorFlow. It is optimized for real-time face detection tasks and runs efficiently on mobile CPUs, ensuring minimal latency and power consumption. Blazeface is based on a fast architecture and uses deep learning techniques to detect faces with high accuracy, even in challenging conditions. It supports multiple face detection in varying lighting and poses, and is designed to work in real-world applications like mobile apps, robotics, and other resource-constrained environments.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 15
    OpenPR
    OpenPR stands for Open Pattern Recognition project and is intended to be an open source library for algorithms of image processing, computer vision, natural language processing, pattern recognition, machine learning and the related fields.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 16
    CVSharp (aka Computer Vision in C#) is a Computer Vision project. Until the present day just one part of the whole project was actually developed. It's called CVSharp Lab, an Image Processing Tool.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 17
    i3D-converter creates a 3D representation from a couple of images (or a pair of stereo images). This program also performs other Computer Vision operations such as, edge and corner detection, image filtering, getting geometric shapes,...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 18
    Computer Vision Software designed for AttyTheWalker, a made from scratch hexapod robot. Entirely developed in C language for GNU/Linux platforms it's the best way to move the robot using a personal computer with RS-232 serial communication.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    BADGr

    BADGr

    Toolbox for Box Approximation, Decomposition, and Grasping

    BADGr, the BoxGrasping toolbox, is a package for Box Approximation, Decomposition, and Grasping. The toolbox was developed in the Computer Vision & Active Perception Lab, at the Royal Institute of Technology, as a participant of the EU research project PACO-PLUS, and published at the project's end in Summer 2010. BADGr provides modules to approximate the shape of a point cloud (possibly from sensor data) by box primitives. These box primitives then serve as a base for the generation of box-based pre-grasp hypotheses for robot grippers.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    The CVR-Lib (Computer Vision and Robotics Library) is a C++ object oriented library for computer vision. It provides lots of functionality to solve mathematical problems, many image processing and analysis algorithms, classification tools, and much more.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    CAM

    CAM

    Class Activation Mapping

    This repository implements Class Activation Mapping (CAM), a technique to expose the implicit attention of convolutional neural networks by generating heatmaps that highlight the most discriminative image regions influencing a network’s class prediction. The method involves modifying a CNN model slightly (e.g., using global average pooling before the final layer) to produce a weighted combination of feature maps as the class activation map. Integration with existing CNNs (with light modifications). Sample scripts/examples using standard architectures. The repo provides example code and instructions for applying CAM to existing CNN architectures. Visualization of discriminative regions per class.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    Cambio is a computer simulator of a robot with stereo, 3D vision. It is intended mainly as a tool for studying computer vision algorithms, but I might expand it to cover other topics in robotics of interest (sensorymotor cognition, reliability, etc).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    Camera Kombat is an opensource fighting game based on computer vision that enables free, unencumbered interaction. In order to enable this level of interaction, images of the users are captured by a webcam and their gestures are recognized in real-time.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    CoTracker

    CoTracker

    CoTracker is a model for tracking any point (pixel) on a video

    CoTracker is a learning-based point tracking system that jointly follows many user-specified points across a video, rather than tracking each point independently. By reasoning about all tracks together, it can maintain temporal consistency, handle mutual occlusions, and reduce identity swaps when trajectories cross. The model takes sparse point queries on one frame and predicts their sub-pixel locations and a visibility score for every subsequent frame, producing long, coherent trajectories. Its transformer-style architecture aggregates information both along time and across points, allowing it to recover tracks even after brief disappearances. The repository ships with inference scripts, pretrained weights, and simple interfaces to seed points, run tracking, and export trajectories for downstream tasks. Typical uses include correspondence building, motion analysis, dynamic SLAM priors, video editing masks, and evaluation of geometric consistency in real scenes.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    The goal of cvtk is to provide an easy to use computer vision framework that allows real-time tracking of color-marked objects in 2 dimensions.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • Next
MongoDB Logo MongoDB