Computer Vision Libraries for BSD

Browse free open source Computer Vision Libraries and projects for BSD below. Use the toggles on the left to filter open source Computer Vision Libraries by OS, license, language, programming language, and project status.

  • Cloud-based help desk software with ServoDesk Icon
    Cloud-based help desk software with ServoDesk

    Full access to Enterprise features. No credit card required.

    What if You Could Automate 90% of Your Repetitive Tasks in Under 30 Days? At ServoDesk, we help businesses like yours automate operations with AI, allowing you to cut service times in half and increase productivity by 25% - without hiring more staff.
    Try ServoDesk for free
  • HSI Donesafe | #1 Rated Cloud EHS Software | ESG Reporting Icon
    HSI Donesafe | #1 Rated Cloud EHS Software | ESG Reporting

    Trusted by industries worldwide, Donesafe allows organizations to track, manage, and report on essential EHS functions from one central platform.

    HSI Donesafe is a no-code, cloud-based EHS software that simplifies safety, compliance, and risk management, transforming complex processes into user-friendly workflows.
    Learn More
  • 1
    OpenCV

    OpenCV

    Open Source Computer Vision Library

    The Open Source Computer Vision Library has >2500 algorithms, extensive documentation and sample code for real-time computer vision. It works on Windows, Linux, Mac OS X, Android, iOS in your browser through JavaScript. Languages: C++, Python, Julia, Javascript Homepage: https://opencv.org Q&A forum: https://forum.opencv.org/ Documentation: https://docs.opencv.org Source code: https://github.com/opencv Please pay special attention to our tutorials! https://docs.opencv.org/master Books about the OpenCV are described here: https://opencv.org/books.html
    Leader badge
    Downloads: 2,606 This Week
    Last Update:
    See Project
  • 2
    Armadillo

    Armadillo

    fast C++ library for linear algebra & scientific computing

    * Fast C++ library for linear algebra (matrix maths) and scientific computing * Easy to use functions and syntax, deliberately similar to Matlab / Octave * Uses template meta-programming techniques to increase efficiency * Provides user-friendly wrappers for OpenBLAS, Intel MKL, LAPACK, ATLAS, ARPACK, SuperLU and FFTW libraries * Useful for machine learning, pattern recognition, signal processing, bioinformatics, statistics, finance, etc. * Downloads: http://arma.sourceforge.net/download.html * Documentation: http://arma.sourceforge.net/docs.html * Bug reports: http://arma.sourceforge.net/faq.html * Git repo: https://gitlab.com/conradsnicta/armadillo-code
    Leader badge
    Downloads: 2,041 This Week
    Last Update:
    See Project
  • 3
    COLMAP

    COLMAP

    Structure-from-Motion and Multi-View Stereo

    COLMAP is a general-purpose Structure-from-Motion (SfM) and Multi-View Stereo (MVS) pipeline with a graphical and command-line interface. It offers a wide range of features for the reconstruction of ordered and unordered image collections. The software is licensed under the new BSD license.
    Downloads: 32 This Week
    Last Update:
    See Project
  • 4
    PIFuHD

    PIFuHD

    High-Resolution 3D Human Digitization from A Single Image

    PIFuHD (Pixel-Aligned Implicit Function for 3D human reconstruction at high resolution) is a method and codebase to reconstruct high-fidelity 3D human meshes from a single image. It extends prior PIFu work by increasing resolution and detail, enabling fine geometry in cloth folds, hair, and subtle surface features. The method operates by learning an implicit occupancy / surface function conditioned on the image and camera projection; at inference time it queries dense points to reconstruct a mesh via marching cubes. It also uses a two-stage architecture: a coarse global model followed by local refinement patches to capture fine detail, balancing global consistency and local detail. The repo includes training pipelines, dataset loaders (for Multi-POP, etc.), and inference scripts for mesh output including depth maps for postprocessing. To help practical use, there are utilities for normal estimation, texture back-projection, mesh cleanup, and integration with rendering pipelines.
    Downloads: 9 This Week
    Last Update:
    See Project
  • Kryptos is a versatile and user-friendly platform designed to simplify cryptocurrency tax management and portfolio tracking. Icon
    Kryptos is a versatile and user-friendly platform designed to simplify cryptocurrency tax management and portfolio tracking.

    Kryptos is ideal for cryptocurrency investors, businesses, and accountants seeking efficient portfolio management and tailored tax solutions.

    Generate quick tax reports with Kryptos: Ultimate tax reporting platform, backed by Binance, for seamless tax reporting across 30+ countries, 3000+ DeFi protocols, 100+ wallets & exchanges, 50+ blockchains, NFTs and much more.
    Learn More
  • 5
    DensePose

    DensePose

    A real-time approach for mapping all human pixels of 2D RGB images

    DensePose is a computer vision system that maps all human pixels in an RGB image to the 3D surface of a human body model. It extends human pose estimation from predicting joint keypoints to providing dense correspondences between 2D images and a canonical 3D mesh (such as the SMPL model). This enables detailed understanding of human shape, motion, and surface appearance directly from images or videos. The repository includes the DensePose network architecture, training code, pretrained models, and dataset tools for annotation and visualization. DensePose is widely used in augmented reality, motion capture, virtual try-on, and visual effects applications because it enables real-time 3D human mapping from 2D inputs. The model architecture builds on Mask R-CNN, using additional regression heads to predict UV coordinates that map image pixels to 3D surfaces.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 6
    Vision Transformer Pytorch

    Vision Transformer Pytorch

    Implementation of Vision Transformer, a simple way to achieve SOTA

    This repository provides a from-scratch, minimalist implementation of the Vision Transformer (ViT) in PyTorch, focusing on the core architectural pieces needed for image classification. It breaks down the model into patch embedding, positional encoding, multi-head self-attention, feed-forward blocks, and a classification head so you can understand each component in isolation. The code is intentionally compact and modular, which makes it easy to tinker with hyperparameters, depth, width, and attention dimensions. Because it stays close to vanilla PyTorch, you can integrate custom datasets and training loops without framework lock-in. It’s widely used as an educational reference for people learning transformers in vision and as a lightweight baseline for research prototypes. The project encourages experimentation—swap optimizers, change augmentations, or plug the transformer backbone into downstream tasks.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 7
    Mobile Robot Programming Toolkit (MRPT)

    Mobile Robot Programming Toolkit (MRPT)

    **MOVED TO GITHUB** ==> https://github.com/MRPT/mrpt

    **MOVED TO GITHUB** ==> https://github.com/MRPT/mrpt The Mobile Robot Programming Toolkit (MRPT) is an extensive, cross-platform, and open source C++ library aimed for robotics researchers to design and implement algorithms about Localization, SLAM, Navigation, computer vision. http://www.mrpt.org/
    Leader badge
    Downloads: 10 This Week
    Last Update:
    See Project
  • 8
    Screenshot to Code

    Screenshot to Code

    A neural network that transforms a design mock-up into static websites

    Screenshot-to-code is a tool or prototype that attempts to convert UI screenshots (e.g., of mobile or web UIs) into code representations, likely generating layouts, HTML, CSS, or markup from image inputs. It is part of a research/proof-of-concept domain in UI automation and image-to-UI code generation. Mapping visual design to code constructs. Code/UI layout (HTML, CSS, or markup). Examples/demo scripts showing “image UI code”.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 9
    Segment Anything

    Segment Anything

    Provides code for running inference with the SegmentAnything Model

    Segment Anything (SAM) is a foundation model for image segmentation that’s designed to work “out of the box” on a wide variety of images without task-specific fine-tuning. It’s a promptable segmenter: you guide it with points, boxes, or rough masks, and it predicts high-quality object masks consistent with the prompt. The architecture separates a powerful image encoder from a lightweight mask decoder, so the heavy vision work can be computed once and the interactive part stays fast. A bundled automatic mask generator can sweep an image and propose many object masks, which is useful for dataset bootstrapping or bulk annotation. The repository includes ready-to-use weights, Python APIs, and example notebooks demonstrating both interactive and automatic modes. Because SAM was trained with an extremely large and diverse mask dataset, it tends to generalize well to new domains, making it a practical starting point for research and production annotation tools.
    Downloads: 1 This Week
    Last Update:
    See Project
  • Digital Adoption and User Training - in one solution Icon
    Digital Adoption and User Training - in one solution

    ClickLearn produces the right learning format for any training situation

    ClickLearn captures work processes in enterprise software - and auto-produces learning content in 7 formats and 45 languages, creates a customisable e-learning portal and keeps documentation current with automatic updates.
    Learn More
  • 10
    VGGT

    VGGT

    [CVPR 2025 Best Paper Award] VGGT

    VGGT is a transformer-based framework aimed at unifying classic visual geometry tasks—such as depth estimation, camera pose recovery, point tracking, and correspondence—under a single model. Rather than training separate networks per task, it shares an encoder and leverages geometric heads/decoders to infer structure and motion from images or short clips. The design emphasizes consistent geometric reasoning: outputs from one head (e.g., correspondences or tracks) reinforce others (e.g., pose or depth), making the system more robust to challenging viewpoints and textures. The repo provides inference pipelines to estimate geometry from monocular inputs, stereo pairs, or brief sequences, together with evaluation harnesses for common geometry benchmarks. Training utilities highlight data curation and augmentations that preserve geometric cues while improving generalization across scenes and cameras.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    OpenNN - Open Neural Networks Library

    OpenNN - Open Neural Networks Library

    Machine learning algorithms for advanced analytics

    OpenNN is a software library written in C++ for advanced analytics. It implements neural networks, the most successful machine learning method. Some typical applications of OpenNN are business intelligence (customer segmentation, churn prevention…), health care (early diagnosis, microarray analysis…) and engineering (performance optimization, predictive maitenance…). OpenNN does not deal with computer vision or natural language processing. The main advantage of OpenNN is its high performance. This library outstands in terms of execution speed and memory allocation. It is constantly optimized and parallelized in order to maximize its efficiency. The documentation is composed by tutorials and examples to offer a complete overview about the library. OpenNN is developed by Artelnics, a company specialized in artificial intelligence.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 12
    A multi-platform collection of C++ software libraries for Computer Vision and Image Understanding.
    Leader badge
    Downloads: 1 This Week
    Last Update:
    See Project
  • 13
    Panzer Combat II

    Panzer Combat II

    Computer-assisted miniature tank game.

    Panzer Combat II is a multi-player voice and webcam enabled computer-assisted distributed miniature wargame of World War II tank combat. Firing is done by placing a webcam behind the aiming unit. Distance to target is computed using computer vision. Action inside the tanks is performed on the computer screen while battlefield strategy is played on the miniature terrain. Both camps can use a different laptop or tablet, the game will interconnect. You can try it online : http://server.panzercombat.com/PCII_Web/move.htm Look at battle reports : http://www.flickr.com/photos/panzercombatii Or watch a demo : http://www.youtube.com/watch?v=WcjfV8Odtss 100% CLEAN : http://games.softpedia.com/progClean/Panzer-Combat-II-Clean-95530.html
    Downloads: 1 This Week
    Last Update:
    See Project
  • 14
    Scilab Image Processing Toolbox

    Scilab Image Processing Toolbox

    Advanced image processing toolbox for Scilab on Unix/Linux/Mac OS

    SIP is the image processing and computer vision package for SciLab, a free Matlab-like programming environment. SIP reads/writes images in formats like JPEG, PNG, and BMP. It does filtering, segmentation, edge detection, morphology, and shape analysis. Download from Git http://siptoolbox.sourceforge.net/devel
    Downloads: 2 This Week
    Last Update:
    See Project
  • 15
    Blazeface

    Blazeface

    Blazeface is a lightweight model that detects faces in images

    Blazeface is a lightweight, high-performance face detection model designed for mobile and embedded devices, developed by TensorFlow. It is optimized for real-time face detection tasks and runs efficiently on mobile CPUs, ensuring minimal latency and power consumption. Blazeface is based on a fast architecture and uses deep learning techniques to detect faces with high accuracy, even in challenging conditions. It supports multiple face detection in varying lighting and poses, and is designed to work in real-world applications like mobile apps, robotics, and other resource-constrained environments.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 16
    OpenPR
    OpenPR stands for Open Pattern Recognition project and is intended to be an open source library for algorithms of image processing, computer vision, natural language processing, pattern recognition, machine learning and the related fields.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 17
    Computer Vision Software designed for AttyTheWalker, a made from scratch hexapod robot. Entirely developed in C language for GNU/Linux platforms it's the best way to move the robot using a personal computer with RS-232 serial communication.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    BADGr

    BADGr

    Toolbox for Box Approximation, Decomposition, and Grasping

    BADGr, the BoxGrasping toolbox, is a package for Box Approximation, Decomposition, and Grasping. The toolbox was developed in the Computer Vision & Active Perception Lab, at the Royal Institute of Technology, as a participant of the EU research project PACO-PLUS, and published at the project's end in Summer 2010. BADGr provides modules to approximate the shape of a point cloud (possibly from sensor data) by box primitives. These box primitives then serve as a base for the generation of box-based pre-grasp hypotheses for robot grippers.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    The CVR-Lib (Computer Vision and Robotics Library) is a C++ object oriented library for computer vision. It provides lots of functionality to solve mathematical problems, many image processing and analysis algorithms, classification tools, and much more.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    CAM

    CAM

    Class Activation Mapping

    This repository implements Class Activation Mapping (CAM), a technique to expose the implicit attention of convolutional neural networks by generating heatmaps that highlight the most discriminative image regions influencing a network’s class prediction. The method involves modifying a CNN model slightly (e.g., using global average pooling before the final layer) to produce a weighted combination of feature maps as the class activation map. Integration with existing CNNs (with light modifications). Sample scripts/examples using standard architectures. The repo provides example code and instructions for applying CAM to existing CNN architectures. Visualization of discriminative regions per class.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    Cambio is a computer simulator of a robot with stereo, 3D vision. It is intended mainly as a tool for studying computer vision algorithms, but I might expand it to cover other topics in robotics of interest (sensorymotor cognition, reliability, etc).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    Camera Kombat is an opensource fighting game based on computer vision that enables free, unencumbered interaction. In order to enable this level of interaction, images of the users are captured by a webcam and their gestures are recognized in real-time.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    CoTracker

    CoTracker

    CoTracker is a model for tracking any point (pixel) on a video

    CoTracker is a learning-based point tracking system that jointly follows many user-specified points across a video, rather than tracking each point independently. By reasoning about all tracks together, it can maintain temporal consistency, handle mutual occlusions, and reduce identity swaps when trajectories cross. The model takes sparse point queries on one frame and predicts their sub-pixel locations and a visibility score for every subsequent frame, producing long, coherent trajectories. Its transformer-style architecture aggregates information both along time and across points, allowing it to recover tracks even after brief disappearances. The repository ships with inference scripts, pretrained weights, and simple interfaces to seed points, run tracking, and export trajectories for downstream tasks. Typical uses include correspondence building, motion analysis, dynamic SLAM priors, video editing masks, and evaluation of geometric consistency in real scenes.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    The goal of cvtk is to provide an easy to use computer vision framework that allows real-time tracking of color-marked objects in 2 dimensions.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    CVSharp (aka Computer Vision in C#) is a Computer Vision project. Until the present day just one part of the whole project was actually developed. It's called CVSharp Lab, an Image Processing Tool.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • Next