Showing 178 open source projects for "video image extractor"

View related business solutions
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build, govern, and optimize agents and models with Gemini Enterprise Agent Platform.
    Start Free
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • 1
    Video-subtitle-extractor

    Video-subtitle-extractor

    A GUI tool for extracting hard-coded subtitle (hardsub) from videos

    Video hard subtitle extraction, generate srt file. There is no need to apply for a third-party API, and text recognition can be implemented locally. A deep learning-based video subtitle extraction framework, including subtitle region detection and subtitle content extraction. A GUI tool for extracting hard-coded subtitles (hardsub) from videos and generating srt files. Use local OCR recognition, no need to set up and call any API, and do not need to access online OCR services such as Baidu...
    Downloads: 39 This Week
    Last Update:
    See Project
  • 2
    OpenShot Video Editor

    OpenShot Video Editor

    Award-Winning Open Source Video Editing Software

    OpenShot Video Editor is a powerful yet very simple and easy-to-use video editor that delivers high quality video editing and animation solutions. OpenShot offers a myriad of features and capabilities, including powerful curve-based Key frame animations, 3D animated titles and effects, slow motion and time effects, audio mixing and editing, and so much more. It’s available for Linux, Mac and Windows, with a very simple and friendly interface. Start creating stunning videos quickly and easily...
    Downloads: 125 This Week
    Last Update:
    See Project
  • 3
    QualityScaler

    QualityScaler

    Image/video AI upscaler app (BSRGAN)

    Qualityscaler is a Windows app that uses BSRGAN Artificial Intelligence to enhance, enlarge and reduce noise in photographs and videos. QualityScaler is completely written in Python, from the backend to the front end. Image/list of images upscale. Video upscale. Drag&drop files [image / multiple images/video] Automatic image tiling and merging to avoid gpu VRAM limitation. Resize image/video before upscaling. Multiple Gpu support. Compatible images - png, jpeg, bmp, webp, tif. Compatible video - mp4, wemb, gif, mkv, flv, avi, mov, qt.
    Downloads: 137 This Week
    Last Update:
    See Project
  • 4
    Video Frame Extractor

    Video Frame Extractor

    Extracts semi-random frames from all MP4 videos

    This simple tool extracts frames from all MP4 videos in the same folder as this program. ## How to use: - Place this program in the folder containing your MP4 videos. - Double-click on VideoFrameExtractor.exe to run it. - When prompted, enter the number of frames you want to extract from each video. - Wait for the program to finish processing all videos. - Find your extracted frames in the 'extracted_frames' folder. The frames are extracted at evenly distributed points...
    Leader badge
    Downloads: 21 This Week
    Last Update:
    See Project
  • Forever Free Full-Stack Observability | Grafana Cloud Icon
    Forever Free Full-Stack Observability | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 5
    Deep-Live-Cam

    Deep-Live-Cam

    Real time face swap and one-click video deepfake

    Real time face swap and one-click video deepfake with only a single image. Choose a face (image with the desired face) and the target image/video (image/video in which you want to replace the face) and click on Start. Open File Explorer and navigate to the directory you select your output to be in. You will find a directory named <video_title> where you can see the frames being swapped in real time.
    Downloads: 586 This Week
    Last Update:
    See Project
  • 6
    Image-Editor

    Image-Editor

    AI based photo editing website for changing image background

    ...Then to, create a new Django project using django-admin startproject Website1, replacing 'Website1' with the name of your choice. Image-Editor uses Python's cv2 library, which provides an easy and efficient way to work with images and videos, including a wide range of image processing and computer vision algorithms. With cv2, you can easily read, write, filter, and display images, and much more. Image-Editor uses Mediapipe's selfie_segmentation model for background removal in real-time video streams. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Video2X

    Video2X

    A lossless video/GIF/image upscaler achieved with waifu2x, Anime4K

    A lossless video/GIF/image upscale achieved with waifu2x, Anime4K, SRMD and RealSR. Started in Hack the Valley 2, 2018. The latest Windows update is built based on version 4.8.1. GUI is not available for 5.0.0 yet, but is already under development. Go to the GUI page to see the basic usage of the GUI. Try the mirror if you can't download releases directly from GitHub.
    Downloads: 509 This Week
    Last Update:
    See Project
  • 8
    Real-ESRGAN GUI

    Real-ESRGAN GUI

    Cross-platform GUI for image upscaler Real-ESRGAN

    The graphic interface of Real-ESRGAN, a practical and beautiful image magnification tool, refers to the design of waifu2x-caffe. This program is a graphical interface of Real-ESRGAN's command line program Real-ESRGAN-ncnn-vulkan , written in Python and tkinter, and supports Windows, Ubuntu and macOS platforms. According to actual measurements, arm64the single-architecture performance is better than universal2the dual- architecture Mac on the Apple chip, so Apple chip users are advised to...
    Downloads: 81 This Week
    Last Update:
    See Project
  • 9
    Anime Player

    Anime Player

    Video player for improving quality of hand-drawn images

    A video player that enhances the quality of a hand-drawn image using Anime4K's high-performance scaling algorithm. This program is a video player written in the Python programming language using the PySimpleGUI graphical user interface library, an mpv media player, and the Anime4K scaling algorithm . Anime Player is designed to play video and audio files and includes functions such as opening files, URLs and folders, setting image scaling parameters using the Anime4K algorithm, creating an mpv config for watching videos using the Anime4K algorithm on Android, viewing help and information about tuning the algorithm. ...
    Downloads: 27 This Week
    Last Update:
    See Project
  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • 10
    labelme Image Polygonal Annotation

    labelme Image Polygonal Annotation

    Image polygonal annotation with Python

    Labelme is a graphical image annotation tool. It is written in Python and uses Qt for its graphical interface. Image annotation for polygon, rectangle, circle, line and point. Image flag annotation for classification and cleaning. Video annotation. (video annotation). GUI customization (predefined labels / flags, auto-saving, label validation, etc). Exporting VOC-format dataset for semantic/instance segmentation.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 11
    Wan2.1

    Wan2.1

    Wan2.1: Open and Advanced Large-Scale Video Generative Model

    ...The model supports text-to-video and image-to-video generation tasks with flexible resolution options suitable for various GPU hardware configurations. Wan2.1’s architecture balances generation quality and inference cost, paving the way for later improvements seen in Wan2.2 such as Mixture-of-Experts and enhanced aesthetics. It was trained on large-scale video and image datasets, providing generalization across diverse scenes and motion patterns.
    Downloads: 67 This Week
    Last Update:
    See Project
  • 12
    Wan2.2

    Wan2.2

    Wan2.2: Open and Advanced Large-Scale Video Generative Model

    ...The model is trained on significantly larger datasets than its predecessor, greatly enhancing motion complexity, semantic understanding, and aesthetic diversity. Wan2.2 also open-sources a 5-billion parameter high-compression VAE-based hybrid text-image-to-video (TI2V) model that supports 720P video generation at 24fps on consumer-grade GPUs like the RTX 4090. It supports multiple video generation tasks including text-to-video.
    Downloads: 99 This Week
    Last Update:
    See Project
  • 13
    CogVideo

    CogVideo

    Text and image to video generation: CogVideoX and CogVideo

    CogVideo is an open-source family of advanced video generation models that can create videos from text, images, or existing video inputs. Built on large-scale Transformer and diffusion architectures, it enables multimodal generation across text-to-video, image-to-video, and video continuation tasks. The latest CogVideoX models offer higher resolution outputs, longer video durations, and improved controllability through prompt engineering. ...
    Downloads: 20 This Week
    Last Update:
    See Project
  • 14
    Make-A-Video - Pytorch (wip)

    Make-A-Video - Pytorch (wip)

    Implementation of Make-A-Video, new SOTA text to video generator

    ...The gist of the paper comes down to, take a SOTA text-to-image model (here they use DALL-E2, but the same learning points would easily apply to Imagen), make a few minor modifications for attention across time and other ways to skimp on the compute cost, do frame interpolation correctly, get a great video model out. Passing in images (if one were to pretrain on images first), both temporal convolution and attention will be automatically skipped.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    BackgroundRemover

    BackgroundRemover

    Background Remover lets you Remove Background from images and video

    BackgroundRemover is a command line tool to remove background from image and video, made by nadermx to power BackgroundRemoverAI. If you wonder why it was made read this short blog post.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 16
    nunif

    nunif

    Misc; latest version of waifu2x; 2D video to stereo 3D video

    nunif is a deep learning–based image processing framework focused on image upscaling, restoration, denoising, and enhancement tasks using neural network models. The project provides a collection of AI-powered utilities designed primarily for anime-style artwork, illustrations, and high-quality image restoration workflows. It includes command-line tools and graphical interfaces for applying trained neural models to improve image resolution and visual clarity while minimizing artifacts. nunif...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 17
    HunyuanCustom

    HunyuanCustom

    Multimodal-Driven Architecture for Customized Video Generation

    HunyuanCustom is a multimodal video customization framework by Tencent Hunyuan, aimed at generating customized videos featuring particular subjects (people, characters) under flexible conditions, while maintaining subject/identity consistency. It supports conditioning via image, audio, video, and text, and can perform subject replacement in videos, generate avatars speaking given audio, or combine multiple subject images.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    WanGP

    WanGP

    AI video generator optimized for low VRAM and older GPUs use

    Wan2GP is an open source AI video generation toolkit designed to make modern generative models accessible on consumer-grade hardware with limited GPU memory. It acts as a unified interface for running multiple video, image, and audio generation models, including Wan-based models as well as other systems like Hunyuan Video, Flux, and Qwen. A key focus of the project is reducing VRAM requirements, enabling some workflows to run on as little as 6 GB while still supporting older Nvidia and certain AMD GPUs. ...
    Downloads: 48 This Week
    Last Update:
    See Project
  • 19
    HunyuanVideo-I2V

    HunyuanVideo-I2V

    A Customizable Image-to-Video Model based on HunyuanVideo

    HunyuanVideo-I2V is a customizable image-to-video generation framework from Tencent Hunyuan, built on their HunyuanVideo foundation. It extends video generation so that given a static reference image plus an optional prompt, it generates a video sequence that preserves the reference image’s identity (especially in the first frame) and allows stylized effects via LoRA adapters.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    LTX-2.3

    LTX-2.3

    Official Python inference and LoRA trainer package

    LTX-2.3 is an open-source multimodal artificial intelligence foundation model developed by Lightricks for generating synchronized video and audio from prompts or other inputs. Unlike most earlier video generation systems that only produced silent clips, LTX-2 combines video and audio generation in a unified architecture capable of producing coherent audiovisual scenes. The model uses a diffusion-transformer-based architecture designed to generate high-fidelity visual frames while...
    Downloads: 153 This Week
    Last Update:
    See Project
  • 21
    SwarmUI

    SwarmUI

    Modular AI image and video generation web UI with extensible tools

    SwarmUI is a modular web-based user interface designed for AI-driven image generation, with a strong focus on usability, performance, and extensibility. It serves as a unified environment for working with multiple AI models, including Stable Diffusion and newer image and video generation systems, allowing users to create and manage outputs through a browser interface. SwarmUI is built to accommodate both beginners and advanced users by offering a simple “Generate” interface alongside more advanced workflow tools that expose deeper configuration options. ...
    Downloads: 10 This Week
    Last Update:
    See Project
  • 22
    FaceFusion

    FaceFusion

    Industry leading face manipulation platform

    FaceFusion is an open-source face swapping and facial enhancement toolkit designed for high-quality video and image manipulation workflows. The project enables users to replace faces in images or videos while maintaining temporal consistency and visual realism. It integrates modern deep learning models for face detection, alignment, and blending to produce smoother results than traditional approaches. FaceFusion is built with a modular pipeline that allows users to customize processing steps and optimize performance for different hardware environments. ...
    Downloads: 311 This Week
    Last Update:
    See Project
  • 23
    AV1 AVIF

    AV1 AVIF

    AV1 Image File Format Specification - ISO-BMFF/HEIF derivative

    AV1 AVIF is the official specification and reference design for the AV1 Image File Format (AVIF), defining how AV1-encoded bitstreams are packaged into the HEIF container format (based on ISOBMFF) to produce AVIF files. The project outlines the syntax and semantics required for AVIF compliance, including support for multiple image profiles, color depths, chroma subsampling modes, HDR/WCG, alpha channels, animation/image sequences, and various color-space/bit-depth combinations — making AVIF...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 24
    SD.Next

    SD.Next

    All-in-one WebUI for AI generative image and video creation

    SD.Next is an all-in-one web user interface for generative image creation that expands beyond basic Stable Diffusion workflows to cover broader image and video generation, captioning, and processing tasks. It is designed as a power-user environment where model management, generation features, and workflow controls are centralized in a single UI rather than spread across separate scripts and utilities.
    Downloads: 22 This Week
    Last Update:
    See Project
  • 25
    MiniMax-MCP

    MiniMax-MCP

    Official MiniMax Model Context Protocol (MCP) server

    ...It acts as a bridge between tools like Claude Desktop, Cursor, Windsurf, OpenAI Agents, and the MiniMax platform, exposing capabilities such as text-to-speech, voice cloning, image generation, text-to-image, video generation, image-to-video, text-to-video, and music generation. The server is written in Python and distributed under the MIT license, with a pyproject.toml and uv-based workflow that makes installation and execution reproducible. Configuration is handled through JSON files that tell MCP clients how to launch the server (typically via uvx minimax-mcp) and which environment variables to use for the API key, host, and output directory. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • 4
  • 5
  • Next
MongoDB Logo MongoDB