Open Source Artificial Intelligence Software - Page 16

Artificial Intelligence Software

View 13598 business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    Let your crypto work for you

    Put idle assets to work with competitive interest rates, borrow without selling, and trade with precision. All in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 1
    Moltis

    Moltis

    A Rust-native claw you can trust

    Moltis is an open-source personal AI assistant platform written in Rust that is designed to run as a fully self-hosted, local-first agent environment. It compiles the entire assistant stack, including the web interface, model routing, memory, and tools, into a single self-contained binary with no external runtime dependencies. The system supports multiple large language model providers alongside local models, enabling users to maintain privacy while still accessing cloud capabilities when needed. Moltis emphasizes security through sandboxed execution environments, where commands and browsing tasks run in isolated containers and require explicit approval. The platform also includes long-term memory powered by hybrid vector and full-text search, allowing the assistant to retain context across sessions. With multi-channel access such as web UI, Telegram, and API endpoints, Moltis functions as a unified automation hub intended for developers and advanced users who want full control.
    Downloads: 19 This Week
    Last Update:
    See Project
  • 2
    Onlook

    Onlook

    The Cursor for Designers • An Open-Source AI-First Design tool

    Seamlessly integrate with any website or web app running on React + TailwindCSS, and make live edits directly in the browser DOM. Customize your design, control your codebase, and push changes your changes without compromise. Link Onlook to your React project with just one command. Run this command on your project's root folder to get set up in seconds. Onlook writes reliable code you can trust, exactly where it needs to go. Adjust layouts, change colors, modify text, and more.
    Downloads: 19 This Week
    Last Update:
    See Project
  • 3
    Qwen3.5

    Qwen3.5

    Qwen3.5 is the large language model series developed by Qwen team

    Qwen3.5 is part of Alibaba’s Qwen family of large language and multimodal foundation models, designed to power advanced AI applications such as chatbots, coding assistants, and autonomous agents. The project represents a significant step toward “agentic AI,” meaning models that can reason through multi-step tasks and interact with external tools or environments rather than only generating text. Qwen3.5 builds on earlier Qwen generations by improving multilingual understanding, reasoning ability, and efficiency, while also introducing native multimodal capabilities that allow the model to work with both language and visual inputs. Architecturally, the system leverages modern large-scale training techniques and mixture-of-experts style efficiency so that very large parameter counts can be used while keeping inference practical.
    Downloads: 19 This Week
    Last Update:
    See Project
  • 4
    Short Video Factory

    Short Video Factory

    AI tool for automatic batch short video creation and editing

    Short Video Factory is an open source desktop application designed to simplify the creation of short-form videos using AI-driven automation. It enables users to generate product marketing clips and general content videos by combining simple prompt-based input with pre-prepared media assets. Short Video Factory integrates multiple stages of video production, including script generation, voice synthesis, video editing, and subtitle effects, into a single streamlined workflow. By leveraging AI technologies, it significantly reduces the manual effort required to produce high-quality short videos at scale. Short Video Factory supports batch processing, allowing users to automatically generate multiple videos based on predefined templates and configurations. It is built as a cross-platform desktop solution with a focus on usability, making it accessible to both beginners and content creators who need fast turnaround times.
    Downloads: 19 This Week
    Last Update:
    See Project
  • Full-stack observability with actually useful AI | Grafana Cloud Icon
    Full-stack observability with actually useful AI | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 5
    Superset LLM

    Superset LLM

    Run an army of Claude Code, Codex, etc. on your machine

    Superset is a development environment and terminal-based platform designed to orchestrate multiple AI coding agents simultaneously within a single workspace. The tool enables developers to run many autonomous coding agents in parallel without the typical overhead of manually managing multiple terminals, repositories, or branches. Each agent task is isolated in its own Git worktree, ensuring that code changes from different agents do not interfere with each other while allowing developers to track their progress independently. The platform includes built-in monitoring capabilities so users can observe the activity of each agent, receive notifications when tasks are completed, and quickly review changes produced by automated coding workflows. Superset also integrates tools for reviewing code differences, editing generated outputs, and managing the development environment directly from the interface.
    Downloads: 19 This Week
    Last Update:
    See Project
  • 6
    The Julia Programming Language

    The Julia Programming Language

    High-level, high-performance dynamic language for technical computing

    Julia is a fast, open source high-performance dynamic language for technical computing. It can be used for data visualization and plotting, deep learning, machine learning, scientific computing, parallel computing and so much more. Having a high level syntax, Julia is easy to use for programmers of every level and background. Julia has more than 2,800 community-registered packages including various mathematical libraries, data manipulation tools, and packages for general purpose computing. Libraries from Python, R, C/Fortran, C++, and Java can also be used.
    Downloads: 19 This Week
    Last Update:
    See Project
  • 7
    Torch-TensorRT

    Torch-TensorRT

    PyTorch/TorchScript/FX compiler for NVIDIA GPUs using TensorRT

    Torch-TensorRT is a compiler for PyTorch/TorchScript, targeting NVIDIA GPUs via NVIDIA’s TensorRT Deep Learning Optimizer and Runtime. Unlike PyTorch’s Just-In-Time (JIT) compiler, Torch-TensorRT is an Ahead-of-Time (AOT) compiler, meaning that before you deploy your TorchScript code, you go through an explicit compile step to convert a standard TorchScript program into a module targeting a TensorRT engine. Torch-TensorRT operates as a PyTorch extension and compiles modules that integrate into the JIT runtime seamlessly. After compilation using the optimized graph should feel no different than running a TorchScript module. You also have access to TensorRT’s suite of configurations at compile time, so you are able to specify operating precision (FP32/FP16/INT8) and other settings for your module.
    Downloads: 19 This Week
    Last Update:
    See Project
  • 8
    node-llama-cpp

    node-llama-cpp

    Run AI models locally on your machine with node.js bindings for llama

    node-llama-cpp is a JavaScript and Node.js binding that allows developers to run large language models locally using the high-performance inference engine provided by llama.cpp. The library enables applications built with Node.js to interact directly with local LLM models without requiring a remote API or external service. By using native bindings and optimized model execution, the framework allows developers to integrate advanced language model capabilities into desktop applications, server software, and command-line tools. The system automatically detects the available hardware on a machine and selects the most appropriate compute backend, including CPU or GPU acceleration. Developers can use the library to perform tasks such as text generation, conversational chat, embedding generation, and structured output generation. Because it runs models locally, the platform is particularly useful for privacy-sensitive environments or offline AI deployments.
    Downloads: 19 This Week
    Last Update:
    See Project
  • 9
    yek

    yek

    Serialize repositories into LLM-ready context w/ smart prioritization

    Yek is a Rust-based CLI tool designed to serialize text-based files from a repository or directory into a single structured output for large language model use. It scans projects using .gitignore rules to exclude irrelevant files and automatically filters out binary or oversized content. Yek prioritizes files based on Git history, placing more important content later in the output to align with how language models process context. Yek supports multiple directories, individual files, and glob patterns, making it flexible for different workflows. It can stream output when piped or save results to a temporary file, depending on usage. Configuration is handled through a yek.yaml file, allowing users to define ignore rules and priority settings. By consolidating code and documents into a single, ordered format, Yek simplifies preparing repositories for AI-driven analysis, debugging, or automation tasks.
    Downloads: 19 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 10
    Basic Memory

    Basic Memory

    Persistent AI memory using local Markdown knowledge graphs

    Basic Memory is an open source knowledge system that turns AI conversations into persistent, structured knowledge you control. Instead of losing context after each chat, it stores information as simple Markdown files on your device, allowing both you and AI to read and write to the same knowledge base. It uses the Model Context Protocol (MCP) so compatible AI tools can access, update, and build on your notes across sessions. Basic Memory creates a semantic knowledge graph by linking related ideas, making it easier to retrieve, expand, and connect information over time. With a local-first design, your data stays private and portable, while optional cloud sync enables cross-device access. It combines simplicity with powerful indexing and search, giving you a flexible way to build long-term memory for projects, research, and workflows.
    Downloads: 18 This Week
    Last Update:
    See Project
  • 11
    Claude-Mem

    Claude-Mem

    Claude Code plugin that automatically captures everything Claude does

    Claude-Mem is a persistent memory compression system built specifically for Claude Code to preserve context across coding sessions. It automatically captures Claude’s tool usage, observations, and decisions, then compresses them into semantic memories that carry forward into future sessions. By enabling long-term continuity, Claude-Mem helps Claude “remember” project history, past fixes, and prior reasoning even after restarts or reconnects. Its progressive disclosure approach intelligently injects only the most relevant context, balancing usefulness with token efficiency. Claude-Mem runs automatically in the background with no manual workflow changes required. Designed for serious developers, it transforms Claude Code into a continuously learning, project-aware coding assistant.
    Downloads: 18 This Week
    Last Update:
    See Project
  • 12
    Coursera Machine Learning

    Coursera Machine Learning

    Coursera Machine Learning By Prof. Andrew Ng

    CourseraMachineLearning is a personal collection of resources, notes, and programming exercises from Andrew Ng’s popular Machine Learning course on Coursera. It consolidates lecture references, programming tutorials, test cases, and supporting materials into one repository for easier review and practice. The project highlights fundamental machine learning concepts such as hypothesis functions, cost functions, gradient descent, bias-variance tradeoffs, and regression models. It also organizes week-by-week course schedules with links to exercises, lecture notes, and additional resources. Alongside the official coursework, the repository includes supplemental explanations, code snippets, and references to recommended textbooks and external materials. By gathering course-related resources into a single space, this project acts as a practical study companion for learners revisiting or supplementing the original course.
    Downloads: 18 This Week
    Last Update:
    See Project
  • 13
    Dev Janitor

    Dev Janitor

    Your Vibe Coding Toolkit A cross-platform desktop application

    Dev Janitor is an open-source developer productivity tool designed to automatically clean up stale, unused, or poorly maintained code patterns in a codebase, helping teams keep their repositories tidy without consuming engineering time manually pruning technical debt. The tool analyzes project files and identifies opportunities to perform cleanup tasks such as removing dead imports, fixing outdated syntax, simplifying redundant expressions, and consolidating duplicated logic, all while observing established conventions for the languages it supports. Through a pluggable rule system, it allows teams to enforce their own style guides or cleanup policies, enabling tailored automation that fits each codebase’s unique needs. It is built for integration into continuous integration and deployment (CI/CD) pipelines or pre-commit workflows, so cleanup suggestions can be surfaced and applied incrementally rather than as a massive refactor.
    Downloads: 18 This Week
    Last Update:
    See Project
  • 14
    Faster Whisper

    Faster Whisper

    Faster Whisper transcription with CTranslate2

    Faster Whisper is an optimized implementation of the Whisper speech recognition model designed to deliver significantly faster inference while maintaining comparable accuracy. It leverages efficient inference engines and optimized computation strategies to reduce latency and resource consumption. The system is particularly useful for real-time or large-scale transcription tasks where performance is critical. It supports multiple model sizes, allowing users to balance speed and accuracy based on their needs. The architecture is designed to run efficiently on both CPUs and GPUs, making it accessible across different environments. It also includes support for streaming and batch processing, enabling flexible deployment scenarios. Overall, faster-whisper makes state-of-the-art speech recognition more practical for production use cases by improving speed and efficiency without sacrificing quality.
    Downloads: 18 This Week
    Last Update:
    See Project
  • 15
    LLaMA 3

    LLaMA 3

    The official Meta Llama 3 GitHub site

    This repository is the former home for Llama 3 model artifacts and getting-started code, covering pre-trained and instruction-tuned variants across multiple parameter sizes. It introduced the public packaging of weights, licenses, and quickstart examples that helped developers fine-tune or run the models locally and on common serving stacks. As the Llama stack evolved, Meta consolidated repositories and marked this one deprecated, pointing users to newer, centralized hubs for models, utilities, and docs. Even as a deprecated repo, it documents the transition path and preserves references that clarify how Llama 3 releases map into the current ecosystem. Practically, it functioned as a bridge between Llama 2 and later Llama releases by standardizing distribution and starter code for inference and fine-tuning. Teams still treat it as historical reference material for version lineage and migration notes.
    Downloads: 18 This Week
    Last Update:
    See Project
  • 16
    LiteLLM

    LiteLLM

    lightweight package to simplify LLM API calls

    Call all LLM APIs using the OpenAI format [Anthropic, Huggingface, Cohere, Azure OpenAI etc.] liteLLM supports streaming the model response back, pass stream=True to get a streaming iterator in response. Streaming is supported for OpenAI, Azure, Anthropic, and Huggingface models.
    Downloads: 18 This Week
    Last Update:
    See Project
  • 17
    Luna AI

    Luna AI

    Virtual AI anchor that combines state-of-the-art technology

    Luna AI is a virtual AI streamer framework designed to power an interactive VTuber that can go live on major platforms and chat with viewers in real time. It is built around a core assistant persona called “Luna AI,” which can be driven by a wide range of large language models and platforms, including GPT-style APIs, Claude, LangChain-based backends, ChatGLM, Kimi, Ollama, and many others. The project supports multiple rendering backends for the avatar, such as Live2D, Unreal Engine (UE), and “xuniren,” and can output to streaming platforms like Bilibili, Douyin, Kuaishou, WeChat Channels, Pinduoduo, Douyu, YouTube, Twitch, and TikTok. For voice, it integrates with numerous TTS engines (Edge-TTS, VITS-Fast, ElevenLabs, VALL-E-X, OpenVoice, GPT-SoVITS, Azure TTS, fish-speech, ChatTTS, CosyVoice, F5-TTS, MultiTTS, MeloTTS, and others), and can optionally pass the output through voice conversion systems like so-vits-svc or DDSP-SVC to change timbre.
    Downloads: 18 This Week
    Last Update:
    See Project
  • 18
    Nimbalyst

    Nimbalyst

    Run multiple Codex and Claude Code AI sessions

    Crystal is an open-source project focused on building a lightweight and flexible system for managing structured data, workflows, or automation pipelines, typically oriented toward developer productivity and extensible backend tooling. It is designed with modularity in mind, allowing developers to define reusable components and compose them into larger workflows that can adapt to different use cases. The project emphasizes simplicity and clarity, making it easier to understand and extend compared to heavier enterprise frameworks. Crystal often leverages modern programming practices and clean architecture principles to ensure maintainability and scalability as projects grow. It can be used as a foundation for building internal tools, automation systems, or data processing pipelines, depending on how developers configure its components. The system is particularly useful for teams that want control over their infrastructure without relying on overly complex or opinionated platforms.
    Downloads: 18 This Week
    Last Update:
    See Project
  • 19
    SentencePiece

    SentencePiece

    Unsupervised text tokenizer for Neural Network-based text generation

    SentencePiece is an unsupervised text tokenizer and detokenizer mainly for Neural Network-based text generation systems where the vocabulary size is predetermined prior to the neural model training. SentencePiece implements subword units (e.g., byte-pair-encoding (BPE) [Sennrich et al.]) and unigram language model [Kudo.]) with the extension of direct training from raw sentences. SentencePiece allows us to make a purely end-to-end system that does not depend on language-specific pre/postprocessing. Purely data driven, sentencePiece trains tokenization and detokenization models from sentences. Pre-tokenization (Moses tokenizer/MeCab/KyTea) is not always required. SentencePiece treats the sentences just as sequences of Unicode characters. There is no language-dependent logic.
    Downloads: 18 This Week
    Last Update:
    See Project
  • 20
    Unsloth Studio

    Unsloth Studio

    Unified web UI for training and running open models locally

    Unsloth Studio is a web-based interface for running and training AI models locally with a unified and user-friendly experience. It allows users to work with a wide range of models for text, audio, vision, embeddings, and more without relying heavily on cloud infrastructure. Built on top of the Unsloth framework, it focuses on high-performance training with reduced VRAM usage and faster speeds compared to traditional methods. The platform supports fine-tuning, pretraining, and reinforcement learning workflows, making it suitable for both experimentation and production use. Users can interact with models through chat, upload files like PDFs or images, and execute code within the environment to improve outputs. By combining powerful optimization techniques with an intuitive UI, Unsloth Studio simplifies the process of building and customizing AI models locally.
    Downloads: 18 This Week
    Last Update:
    See Project
  • 21
    VideoCaptioner

    VideoCaptioner

    AI-powered tool for generating, optimizing, and translating subtitles

    VideoCaptioner is an open source AI-powered subtitle processing tool designed to simplify the workflow of creating subtitles for videos. It integrates speech recognition, language processing, and translation technologies to automatically generate and refine subtitles from video or audio sources. VideoCaptioner uses speech-to-text engines such as Whisper variants to transcribe spoken content and convert it into subtitle text with accurate timestamps. After transcription, large language models are used to intelligently restructure subtitles into natural sentences, correct wording, and improve readability for viewers. It can also translate subtitles into other languages while preserving the original timing, making it suitable for multilingual video publishing and accessibility. In addition to generating subtitles, it supports editing, formatting, and embedding subtitles into videos as either hard or soft subtitles.
    Downloads: 18 This Week
    Last Update:
    See Project
  • 22
    graphify

    graphify

    AI coding assistant skill (Claude Code, Codex, OpenCode, OpenClaw)

    graphify is a data visualization and transformation tool designed to convert structured or semi-structured data into graph-based representations, enabling better understanding of relationships and dependencies. It focuses on building visual models such as nodes and edges that represent entities and their connections, making complex datasets easier to interpret. The system likely supports dynamic updates, allowing graphs to evolve as data changes or new inputs are introduced. It is particularly useful in domains such as network analysis, knowledge graphs, and system architecture visualization. The architecture emphasizes flexibility, enabling users to customize how data is mapped and displayed. It may also include analytical features to explore patterns, clusters, or anomalies within the graph. Overall, graphify serves as a bridge between raw data and visual insight.
    Downloads: 18 This Week
    Last Update:
    See Project
  • 23
    kubectl-ai

    kubectl-ai

    AI assistant for managing Kubernetes clusters from the terminal

    kubectl-ai is an AI-powered command-line assistant designed to help users manage and interact with Kubernetes clusters through natural language queries. It acts as an intelligent interface that interprets user intent and translates it into appropriate Kubernetes operations and commands. By integrating large language models, it enables users to ask questions or request actions in plain language instead of manually crafting complex Kubernetes commands. kubectl-ai runs directly in the terminal and integrates with the existing kubectl workflow, making it familiar for Kubernetes administrators and developers. It can help perform tasks such as inspecting resources, retrieving logs, troubleshooting issues, and modifying cluster configurations. kubectl-ai supports both cloud-based and locally hosted language models, allowing it to adapt to different infrastructure and privacy requirements.
    Downloads: 18 This Week
    Last Update:
    See Project
  • 24
    tlm

    tlm

    Local CLI Copilot, powered by Ollama

    tlm is an open-source command-line AI assistant designed to provide intelligent terminal support using locally running large language models. The project functions as a CLI copilot that helps developers generate commands, explain shell instructions, and answer technical questions directly from the terminal. Instead of relying on cloud APIs or paid AI services, TLM runs entirely on the user’s workstation and integrates with local models managed through the Ollama runtime. This approach allows developers to use powerful open-source models such as Llama, Phi, DeepSeek, and Qwen while maintaining privacy and avoiding external service dependencies. The system supports contextual queries where the AI analyzes files within a directory and generates answers based on project documentation or source code. It also detects the user’s shell environment automatically, allowing it to generate commands tailored to shells such as Bash, Zsh, or PowerShell.
    Downloads: 18 This Week
    Last Update:
    See Project
  • 25
    CoGrOO: Open|LibreOffice Grammar Checker
    CoGrOO (A LibreOffice & OpenOffice.org Grammar Checker), the only open source Portuguese grammar checker that can be used with LibreOffice & OpenOffice.org.
    Leader badge
    Downloads: 92 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB