Open Source BSD Artificial Intelligence Software - Page 5

Artificial Intelligence Software for BSD

  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • Go from Code to Production URL in Seconds Icon
    Go from Code to Production URL in Seconds

    Cloud Run deploys apps in any language instantly. Scales to zero. Pay only when code runs.

    Skip the Kubernetes configs. Cloud Run handles HTTPS, scaling, and infrastructure automatically. Two million requests free per month.
    Try it free
  • 1
    Virastyar

    Virastyar

    Virastyar is an spell checker for low-resource languages

    Virastyar is a free and open-source (FOSS) spell checker. It stands upon the shoulders of many free/libre/open-source (FLOSS) libraries developed for processing low-resource languages, especially Persian and RTL languages Publications: Kashefi, O., Nasri, M., & Kanani, K. (2010). Towards Automatic Persian Spell Checking. SCICT. Kashefi, O., Sharifi, M., & Minaie, B. (2013). A novel string distance metric for ranking Persian respelling suggestions. Natural Language Engineering, 19(2), 259-284. Rasooli, M. S., Kahefi, O., & Minaei-Bidgoli, B. (2011). Effect of adaptive spell checking in Persian. In NLP-KE Contributors: Omid Kashefi Azadeh Zamanifar Masoumeh Mashaiekhi Meisam Pourafzal Reza Refaei Mohammad Hedayati Kamiar Kanani Mehrdad Senobari Sina Iravanin Mohammad Sadegh Rasooli Mohsen Hoseinalizadeh Mitra Nasri Alireza Dehlaghi Fatemeh Ahmadi Neda PourMorteza
    Downloads: 61 This Week
    Last Update:
    See Project
  • 2
    MOA - Massive Online Analysis

    MOA - Massive Online Analysis

    Big Data Stream Analytics Framework.

    A framework for learning from a continuous supply of examples, a data stream. Includes classification, regression, clustering, outlier detection and recommender systems. Related to the WEKA project, also written in Java, while scaling to adaptive large scale machine learning.
    Leader badge
    Downloads: 85 This Week
    Last Update:
    See Project
  • 3
    A Java JNA wrapper for Tesseract OCR API
    Leader badge
    Downloads: 79 This Week
    Last Update:
    See Project
  • 4
    AI YouTube Shorts Generator

    AI YouTube Shorts Generator

    A python tool that uses GPT-4, FFmpeg, and OpenCV

    AI-YouTube-Shorts-Generator is a Python-based tool that automates the creation of short-form vertical video clips (“shorts”) from longer source videos — ideal for adapting content for platforms like YouTube Shorts, Instagram Reels, or TikTok. It analyzes input video (whether a local file or a YouTube URL), transcribes audio (with optional GPU-accelerated speech-to-text), uses an AI model to identify the most compelling or engaging segments, and then crops/resizes the video and applies subtitle overlays, producing a polished short video without manual editing. The tool streamlines multiple steps of the tedious short-form video workflow: highlight detection, clipping, subtitle generation, cropping to vertical 9:16 format, and final rendering — reducing hours of editing to a mostly automated pipeline. Because it supports both local and online video sources, it's flexible whether you're working with your own recorded content or repurposing existing longer-form videos.
    Downloads: 12 This Week
    Last Update:
    See Project
  • Go From AI Idea to AI App Fast Icon
    Go From AI Idea to AI App Fast

    One platform to build, fine-tune, and deploy ML models. No MLOps team required.

    Access Gemini 3 and 200+ models. Build chatbots, agents, or custom models with built-in monitoring and scaling.
    Try Free
  • 5
    GLM-Image

    GLM-Image

    GLM-Image: Auto-regressive for Dense-knowledge and High-fidelity Image

    GLM-Image is an open-source generative AI model designed to create high-fidelity images from text prompts using a hybrid architecture that combines autoregressive semantic understanding with diffusion-based detail refinement. It excels at generating images that include complex layouts and detailed text content, making it especially useful for posters, diagrams, info-graphics, social media graphics, and visual content that requires precise text placement and semantic alignment. Because it blends linguistic reasoning with image synthesis, GLM-Image produces visual outputs where semantic relationships and textual accuracy are prioritized alongside artistic style and realism, and its model structure enables it to handle dense visual knowledge tasks that challenge many pure diffusion models. The model’s design and weights are available under an open-source license that encourages experimentation, integration, and deployment across a range of creative workflows.
    Downloads: 12 This Week
    Last Update:
    See Project
  • 6
    SwarmUI

    SwarmUI

    Modular AI image and video generation web UI with extensible tools

    SwarmUI is a modular web-based user interface designed for AI-driven image generation, with a strong focus on usability, performance, and extensibility. It serves as a unified environment for working with multiple AI models, including Stable Diffusion and newer image and video generation systems, allowing users to create and manage outputs through a browser interface. SwarmUI is built to accommodate both beginners and advanced users by offering a simple “Generate” interface alongside more advanced workflow tools that expose deeper configuration options. It integrates with underlying systems like node-based workflows, enabling flexible and customizable pipelines for complex generation tasks. SwarmUI also emphasizes scalability, originally inspired by the idea of coordinating multiple GPUs to work together for large batch or grid-based image generation. SwarmUI includes a variety of built-in tools such as image editing, prompt handling, and automation features.
    Downloads: 12 This Week
    Last Update:
    See Project
  • 7
    typeui.sh

    typeui.sh

    Design system skills for agentic tools

    typeui.sh is a type-safe UI framework designed to streamline the development of user interfaces by leveraging strong typing and structured definitions. It focuses on improving developer productivity by reducing runtime errors and ensuring consistency across UI components. The framework integrates closely with TypeScript, allowing developers to define interfaces and components with strict type validation. It supports component-based architecture, making it easier to build scalable and maintainable applications. TypeUI also emphasizes reusability, enabling developers to create standardized components that can be shared across projects. Its design encourages clean and predictable code, which is especially valuable in large-scale applications. Overall, TypeUI provides a structured and reliable approach to building modern user interfaces with strong typing guarantees.
    Downloads: 12 This Week
    Last Update:
    See Project
  • 8
    AutoAgent AI

    AutoAgent AI

    Autonomous harness engineering

    AutoAgent is an experimental AI framework focused on autonomous agent engineering, where a meta-agent iteratively improves another agent’s architecture without direct human intervention. Instead of manually tuning prompts or workflows, developers define high-level goals in a configuration file, and the system continuously modifies its own tools, orchestration, and logic based on benchmark performance. It operates through a loop of testing, analyzing failures, and refining the agent’s configuration to maximize a scoring metric. The framework uses a single-file agent harness combined with structured tasks and evaluation suites to guide optimization. It runs inside Docker for safe execution and reproducibility. This approach shifts agent development from manual design to automated optimization. The system is particularly useful for building domain-specific agents that need continuous performance improvement.
    Downloads: 11 This Week
    Last Update:
    See Project
  • 9
    Claude Autoresearch

    Claude Autoresearch

    Claude Autoresearch Skill, autonomous goal-directed iteration

    Claude Autoresearch is an autonomous research assistant system that automates the process of exploring, collecting, and synthesizing information across multiple iterations. It is designed to mimic human research behavior by generating queries, evaluating results, and refining its approach based on previous findings. The system likely integrates with external data sources, allowing it to gather information from diverse inputs and organize it into structured outputs. Its iterative loop enables deeper exploration of topics over time, making it particularly useful for complex or open-ended research questions. The architecture emphasizes autonomy, reducing the need for constant user input while still producing meaningful insights. It may also include summarization and reporting capabilities to present findings in a digestible format. Overall, autoresearch represents a step toward self-directed knowledge discovery systems that continuously improve their outputs through iteration.
    Downloads: 11 This Week
    Last Update:
    See Project
  • Full-stack observability with actually useful AI | Grafana Cloud Icon
    Full-stack observability with actually useful AI | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 10
    Eidos

    Eidos

    An extensible framework for Personal Data Management

    Eidos is an extensible personal data management platform designed to help users organize and interact with their information using a local-first architecture. The system transforms SQLite into a flexible personal database that can store structured and unstructured information such as notes, documents, datasets, and knowledge resources. Its interface is inspired by tools like Notion, allowing users to create documents, databases, and custom views to organize personal information. Unlike cloud-based knowledge tools, Eidos runs entirely on the user’s machine, ensuring privacy and high performance through local storage. The platform integrates large language models to enable AI-assisted features such as summarizing documents, translating content, and interacting with stored data conversationally. It also includes an extension system that allows developers to create custom tools, scripts, and workflows using programming languages such as TypeScript or Python.
    Downloads: 11 This Week
    Last Update:
    See Project
  • 11
    FinGPT

    FinGPT

    Open-Source Financial Large Language Models

    FinGPT is an open-source, finance-specialized large language model framework that blends the capabilities of general LLMs with real-time financial data feeds, domain-specific knowledge bases, and task-oriented agents to support market analysis, research automation, and decision support. It extends traditional GPT-style models by connecting them to live or historical financial datasets, news APIs, and economic indicators so that outputs are grounded in relevant and recent market conditions rather than generic knowledge alone. The platform typically includes tools for fine-tuning, context engineering, and prompt templating, enabling users to build specialized assistants for tasks like sentiment analysis, earnings summary generation, risk profiling, trading signal interpretation, and document extraction from financial reports.
    Downloads: 11 This Week
    Last Update:
    See Project
  • 12
    HunyuanImage-3.0

    HunyuanImage-3.0

    A Powerful Native Multimodal Model for Image Generation

    HunyuanImage-3.0 is a powerful, native multimodal text-to-image generation model released by Tencent’s Hunyuan team. It unifies multimodal understanding and generation in a single autoregressive framework, combining text and image modalities seamlessly rather than relying on separate image-only diffusion components. It uses a Mixture-of-Experts (MoE) architecture with many expert subnetworks to scale efficiently, deploying only a subset of experts per token, which allows large parameter counts without linear inference cost explosion. The model is intended to be competitive with closed-source image generation systems, aiming for high fidelity, prompt adherence, fine detail, and even “world knowledge” reasoning (i.e. leveraging context, semantics, or common sense in generation). The GitHub repo includes code, scripts, model loading instructions, inference utilities, prompt handling, and integration with standard ML tooling (e.g. Hugging Face / Transformers).
    Downloads: 11 This Week
    Last Update:
    See Project
  • 13
    LTX-Video

    LTX-Video

    Official repository for LTX-Video

    LTX-Video is a sophisticated multimedia processing framework from Lightricks designed to handle high-quality video editing, compositing, and transformation tasks with performance and scalability. It provides runtime components that efficiently decode, encode, and manipulate video streams, frame buffers, and audio tracks while exposing a rich API for building customized editing features like transitions, effects, color grading, and keyframe automation. The toolkit is built with both real-time and offline workflows in mind, enabling applications from consumer editing to professional content creation and batch processing. Internally optimized for multi-core processors and hardware acceleration where available, LTX-Video makes it feasible to work with high-resolution content and complex timelines without sacrificing responsiveness.
    Downloads: 11 This Week
    Last Update:
    See Project
  • 14
    OpenShorts

    OpenShorts

    Free & open source AI video platform

    OpenShorts is an open-source, self-hosted AI video automation platform designed to generate, edit, and distribute short-form vertical content across social media platforms. It combines multiple tools into a single pipeline, including clip generation, AI-driven video creation, and YouTube optimization features. The system can transform long videos or uploaded files into short clips by detecting engaging moments, reframing content, and adding subtitles and visual effects. It also supports generating marketing videos using AI actors, voiceovers, and scripted narratives without requiring cameras or production resources. The platform integrates publishing capabilities, allowing users to distribute content directly to TikTok, Instagram, and YouTube. Its architecture uses modern technologies such as FastAPI, FFmpeg, and AI models for transcription, analysis, and rendering. Designed for creators and businesses, it automates the entire lifecycle of short-form video production.
    Downloads: 11 This Week
    Last Update:
    See Project
  • 15
    OpenWork

    OpenWork

    An open-source alternative to Claude Cowork, powered by opencode

    OpenWork is a framework for building decentralized collaborative work environments powered by AI and human contributions. At its core, the project enables contributors to define tasks, workflows, and goals that can be split, shared, and recombined across distributed nodes while agents and humans cooperate to advance progress. It offers structured templates for work items, decision logic for task allocation, and consensus mechanisms that let groups verify and validate results toward shared objectives. This project also includes moderation and reputation layers so that contributor trust and quality can be assessed and integrated into future task assignments. Rather than a single monolithic workflow engine, it emphasizes openness — providing APIs and interfaces so communities can build custom dashboards, integrate specialized agents, or add bespoke evaluation criteria.
    Downloads: 11 This Week
    Last Update:
    See Project
  • 16
    AutoSubs

    AutoSubs

    Instantly generate AI-powered subtitles on your device

    AutoSubs is an open-source, AI-powered subtitle generation tool that enables users to automatically transcribe audio and video content into accurate, editable subtitles directly on their device. It supports both standalone usage and integration with professional video editing software such as DaVinci Resolve, allowing creators to generate and edit subtitles within their existing workflows. The tool leverages speech-to-text models, including OpenAI Whisper, to produce high-quality transcriptions and can differentiate between speakers using diarization techniques. Users can customize subtitle styling, adjust timing, and export results in multiple formats, making it suitable for content creators, filmmakers, and editors. AutoSubs is designed with performance in mind, offering efficient processing through a Rust-based backend and supporting multiple operating systems including Windows, macOS, and Linux.
    Downloads: 10 This Week
    Last Update:
    See Project
  • 17
    Cloudflare Agents

    Cloudflare Agents

    Build and deploy AI Agents on Cloudflare

    Cloudflare Agents is an open-source framework designed to help developers build, deploy, and manage AI agents that run at the network edge. It provides infrastructure for creating stateful, event-driven agents capable of real-time interaction while maintaining low latency through Cloudflare’s distributed platform. The project includes SDKs, templates, and deployment tooling that simplify the process of connecting agents to external APIs, storage systems, and workflows. Its architecture emphasizes persistent memory, enabling agents to maintain context across sessions and interactions. Developers can orchestrate complex behaviors using workflows and durable objects, making it suitable for production-grade autonomous systems. Overall, Cloudflare Agents aims to streamline the development of scalable AI automation that operates close to users for improved performance.
    Downloads: 10 This Week
    Last Update:
    See Project
  • 18
    DeepSeek Coder

    DeepSeek Coder

    DeepSeek Coder: Let the Code Write Itself

    DeepSeek-Coder is a series of code-specialized language models designed to generate, complete, and infill code (and mixed code + natural language) with high fluency in both English and Chinese. The models are trained from scratch on a massive corpus (~2 trillion tokens), of which about 87% is code and 13% is natural language. This dataset covers project-level code structure (not just line-by-line snippets), using a large context window (e.g. 16K) and a secondary fill-in-the-blank objective to encourage better contextual completions and infilling. Multiple sizes of the model are offered (e.g. 1B, 5.7B, 6.7B, 33B) so users can trade off inference cost vs capability. The repo provides model weights, documentation on training setup, evaluation results on common benchmarks (HumanEval, MultiPL-E, APPS, etc.), and inference tools.
    Downloads: 10 This Week
    Last Update:
    See Project
  • 19
    DeepSeek LLM

    DeepSeek LLM

    DeepSeek LLM: Let there be answers

    The DeepSeek-LLM repository hosts the code, model files, evaluations, and documentation for DeepSeek’s LLM series (notably the 67B Chat variant). Its tagline is “Let there be answers.” The repo includes an “evaluation” folder (with results like math benchmark scores) and code artifacts (e.g. pre-commit config) that support model development and deployment. According to the evaluation files, DeepSeek LLM 67B Chat achieves strong performance on math benchmarks under both chain-of-thought (CoT) and tool-assisted reasoning modes. The model is trained from scratch, reportedly on a vast multilingual + code + reasoning dataset, and competes with other open or open-weight models. The architecture mirrors established decoder-only transformer families: pre-norm structure, rotational embeddings (RoPE), grouped query attention (GQA), and mixing in languages and tasks. It supports both “Base” (foundation model) and “Chat” (instruction / conversation tuned) variants.
    Downloads: 10 This Week
    Last Update:
    See Project
  • 20
    Every Code

    Every Code

    Local AI coding agent CLI with multi-agent orchestration tools

    Every Code (often referred to simply as Code) is a fast, local AI-powered coding agent designed to run directly in the terminal environment. It is a community-driven fork of the Codex CLI, with a strong emphasis on improving real-world developer ergonomics and workflows. Every Code enhances the traditional coding assistant model by introducing multi-agent orchestration, allowing multiple AI agents to collaborate, compare solutions, and refine outputs in parallel. It supports integration with various AI providers, enabling users to route tasks across different models depending on their needs. Every Code also includes browser integration and automation capabilities, extending its usefulness beyond simple code generation into more complex development tasks. Customization is a key focus, with support for theming, configurable settings, and reasoning controls that allow developers to fine-tune how the agent behaves.
    Downloads: 10 This Week
    Last Update:
    See Project
  • 21
    LLM Council

    LLM Council

    LLM Council works together to answer your hardest questions

    LLM Council is a creative open-source web application by Andrej Karpathy that lets you consult multiple large language models together to answer questions more reliably than querying a single model. Instead of relying on one provider, this application sends your query simultaneously to several LLMs supported via OpenRouter, collects each model’s independent response, and then orchestrates a multi-stage evaluation where the models critique and rank each other’s outputs anonymously. After this peer-review process, a designated “Chairman” model synthesizes a final consolidated answer drawing on the strengths and insights of all participants. The interface looks like a familiar chat app but under the hood it implements this ensemble and consensus workflow to reduce bias and leverage diverse reasoning styles.
    Downloads: 10 This Week
    Last Update:
    See Project
  • 22
    ML Intern

    ML Intern

    ML engineer that reads papers, trains models, and ships ML models

    ML Intern is a repository by Hugging Face that provides educational content and projects aimed at helping learners gain practical experience in machine learning and AI development. It is designed to simulate the experience of working as a machine learning intern, offering tasks and exercises that mirror real-world workflows. The project includes tutorials, datasets, and example implementations that guide users through different aspects of ML development. It emphasizes hands-on learning, encouraging users to build and experiment rather than passively consume information. The repository also introduces tools and libraries commonly used in the Hugging Face ecosystem. It is structured to help users progressively build skills and confidence in AI development. Overall, ML Intern is a practical learning platform for aspiring machine learning engineers.
    Downloads: 10 This Week
    Last Update:
    See Project
  • 23
    Qwen-Image-Layered

    Qwen-Image-Layered

    Qwen-Image-Layered: Layered Decomposition for Inherent Editablity

    Qwen-Image-Layered is an extension of the Qwen series of multimodal models that introduces layered image understanding, enabling the model to reason about hierarchical visual structures — such as separating foreground, background, objects, and contextual layers within an image. This architecture allows richer semantic interpretation, enabling use cases such as scene decomposition, object-level editing, layered captioning, and more fine-grained multimodal reasoning than with flat image encodings alone. By combining text and structured image representations, it aims to facilitate tasks where both descriptive and structural understanding are important, such as detailed image QA, interactive image editing via prompt layers, and image-conditioned generation with structural control. The layered approach supports training signals that help the model learn how visual elements relate to each other and to textual context, rather than simply learning global image embeddings.
    Downloads: 10 This Week
    Last Update:
    See Project
  • 24
    VideoCaptioner

    VideoCaptioner

    AI-powered tool for generating, optimizing, and translating subtitles

    VideoCaptioner is an open source AI-powered subtitle processing tool designed to simplify the workflow of creating subtitles for videos. It integrates speech recognition, language processing, and translation technologies to automatically generate and refine subtitles from video or audio sources. VideoCaptioner uses speech-to-text engines such as Whisper variants to transcribe spoken content and convert it into subtitle text with accurate timestamps. After transcription, large language models are used to intelligently restructure subtitles into natural sentences, correct wording, and improve readability for viewers. It can also translate subtitles into other languages while preserving the original timing, making it suitable for multilingual video publishing and accessibility. In addition to generating subtitles, it supports editing, formatting, and embedding subtitles into videos as either hard or soft subtitles.
    Downloads: 10 This Week
    Last Update:
    See Project
  • 25
    super-agent-party

    super-agent-party

    All-in-one AI companion! Desktop girlfriend + virtual streamer

    Super Agent Party is an open-source experimental framework designed to demonstrate collaborative multi-agent AI systems interacting within a shared environment. The project explores how multiple specialized AI agents can coordinate to solve complex tasks by communicating with each other and sharing information. Instead of relying on a single monolithic model, the framework organizes agents with different roles or capabilities that cooperate to achieve goals. Each agent may handle different responsibilities such as planning, execution, reasoning, or knowledge retrieval, allowing the system to tackle more complex problems than a single agent might handle alone. The platform is primarily intended as a research and demonstration environment for experimenting with agent collaboration strategies. Developers can use it to study coordination patterns, communication protocols, and task decomposition in multi-agent systems.
    Downloads: 10 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB