Showing 507 open source projects for "source engine"

View related business solutions
  • Forever Free Full-Stack Observability | Grafana Cloud Icon
    Forever Free Full-Stack Observability | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • Go From AI Idea to AI App Fast Icon
    Go From AI Idea to AI App Fast

    One platform to build, fine-tune, and deploy ML models. No MLOps team required.

    Access Gemini 3 and 200+ models. Build chatbots, agents, or custom models with built-in monitoring and scaling.
    Try Free
  • 1
    llama2.c

    llama2.c

    Inference Llama 2 in one file of pure C

    llama2.c is a minimalist implementation of the Llama 2 language model architecture designed to run entirely in pure C. Created by Andrej Karpathy, this project offers an educational and lightweight framework for performing inference on small Llama 2 models without external dependencies. It provides a full training and inference pipeline: models can be trained in PyTorch and later executed using a concise 700-line C program (run.c). While it can technically load Meta’s official Llama 2...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 2
    Superduper

    Superduper

    Superduper: Integrate AI models and machine learning workflows

    ...It supports the latest technologies and techniques, including LLMs, vector-search, RAG, and multimodality as well as classical AI and ML paradigms. Developers may leverage Superduper by building compositional and declarative objects that out-source the details of deployment, orchestration versioning, and more to the Superduper engine. This allows developers to completely avoid implementing MLOps, ETL pipelines, model deployment, data migration, and synchronization. Using Superduper is simply "CAPE": Connect to your data, apply arbitrary AI to that data, package and reuse the application on arbitrary data, and execute AI-database queries and predictions on the resulting AI outputs and data.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    ViZDoom

    ViZDoom

    Doom-based AI research platform for reinforcement learning

    ...It is primarily intended for research in machine visual learning, and deep reinforcement learning, in particular. ViZDoom is based on ZDOOM, the most popular modern source-port of DOOM. This means compatibility with a huge range of tools and resources that can be used to create custom scenarios, availability of detailed documentation of the engine and tools and support of Doom community. Async and sync single-player and multi-player modes. Fast (up to 7000 fps in sync mode, single-threaded). Lightweight (few MBs). ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 4
    Aix-DB

    Aix-DB

    Based on the LangChain/LangGraph framework

    Aix-DB is an open-source intelligent data analysis platform that combines large language models with database technologies to enable conversational data exploration. The system is designed as a ChatBI solution that allows users to query datasets using natural language and receive structured insights, charts, and visualizations automatically. Built on frameworks such as LangChain and LangGraph, Aix-DB integrates retrieval-augmented generation and Text-to-SQL capabilities to convert user...
    Downloads: 1 This Week
    Last Update:
    See Project
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • 5
    Codex MCP Server

    Codex MCP Server

    MCP server wrapper for OpenAI Codex CLI

    Codex MCP Server is an open-source integration tool that allows AI development environments to access the capabilities of the OpenAI Codex command-line interface through the Model Context Protocol. The project acts as a bridge between AI assistants such as Claude Code and the Codex CLI, enabling those assistants to perform advanced coding operations using Codex as a backend engine.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 6
    SEO Machine

    SEO Machine

    A specialized Claude Code workspace for creating long-form

    SEO Machine is an AI-powered content production system built as a structured workspace for generating long-form, SEO-optimized blog content through automated workflows. It integrates research, writing, analysis, and optimization into a single pipeline, allowing users to produce high-quality articles tailored to search engine performance. The system uses specialized commands and agents to perform tasks such as keyword research, competitor analysis, content drafting, and optimization. It...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Parallax

    Parallax

    Parallax is a distributed model serving framework

    Parallax is a decentralized inference framework designed to run large language models across distributed computing resources. Instead of relying on centralized GPU clusters in data centers, the system allows multiple heterogeneous machines to collaborate in serving AI inference workloads. Parallax divides model layers across different nodes and dynamically coordinates them to form a complete inference pipeline. A two-stage scheduling architecture determines how model layers are allocated to...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    DBHub

    DBHub

    Universal database MCP server connecting to MySQL, PostgreSQL

    DBHub is a universal database gateway that implements the MCP server interface so assistants and IDEs can explore and query databases through typed tools. It supports multiple transports—stdio for desktop clients and HTTP for networked scenarios—making it flexible to embed or deploy. Configuration is environment-variable driven, with a DSN and per-engine settings covering Postgres, MySQL, MariaDB, SQL Server, and SQLite. Operational flags include read-only mode, row limits, and even SSH...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    PyBroker

    PyBroker

    Algorithmic Trading in Python with Machine Learning

    Are you looking to enhance your trading strategies with the power of Python and machine learning? Then you need to check out PyBroker! This Python framework is designed for developing algorithmic trading strategies, with a focus on strategies that use machine learning. With PyBroker, you can easily create and fine-tune trading rules, build powerful models, and gain valuable insights into your strategy’s performance.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Secure File Transfer for Windows with Cerberus by Redwood Icon
    Secure File Transfer for Windows with Cerberus by Redwood

    Protect and share files over FTP/S, SFTP, HTTPS and SCP with the #1 rated Windows file transfer server.

    Cerberus supports unlimited users and connections on a single IP, with built-in encryption, 2FA, and a browser-based web client — all deployable in under 15 minutes with a 25-day free trial.
    Try for Free
  • 10
    Core ML Tools

    Core ML Tools

    Core ML tools contain supporting tools for Core ML model conversion

    Use Core ML Tools (coremltools) to convert machine learning models from third-party libraries to the Core ML format. This Python package contains the supporting tools for converting models from training libraries. Core ML is an Apple framework to integrate machine learning models into your app. Core ML provides a unified representation for all models. Your app uses Core ML APIs and user data to make predictions, and to fine-tune models, all on the user’s device. Core ML optimizes on-device...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    PyTorch Ignite

    PyTorch Ignite

    Library to help with training and evaluating neural networks

    High-level library to help with training and evaluating neural networks in PyTorch flexibly and transparently. Less code than pure PyTorch while ensuring maximum control and simplicity. Library approach and no program's control inversion. Use ignite where and when you need. Extensible API for metrics, experiment managers, and other components. The cool thing with handlers is that they offer unparalleled flexibility (compared to, for example, callbacks). Handlers can be any function: e.g....
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    LEANN

    LEANN

    Local RAG engine for private multimodal knowledge search on devices

    LEANN is an open source system designed to enable retrieval-augmented generation (RAG) and semantic search across personal data while running entirely on local devices. It focuses on dramatically reducing the storage overhead typically required for vector search and embedding indexes, enabling efficient large-scale knowledge retrieval on consumer hardware. LEANN introduces a storage-efficient approximate nearest neighbor index combined with on-the-fly embedding recomputation to avoid storing...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    FlexLLMGen

    FlexLLMGen

    Running large language models on a single GPU

    FlexLLMGen is an open-source inference engine designed to run large language models efficiently on limited hardware resources such as a single GPU. The system focuses on high-throughput generation workloads where large batches of text must be processed quickly, such as large-scale data extraction or document analysis tasks. Instead of requiring expensive multi-GPU systems, the framework uses techniques such as memory offloading, compression, and optimized batching to run large models on commodity hardware. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    PicoLM

    PicoLM

    Run a 1-billion parameter LLM on a $10 board with 256MB RAM

    PicoLM is an open-source inference framework designed to run large language models on extremely constrained hardware environments such as inexpensive single-board computers and embedded systems. The project focuses on enabling efficient local inference by optimizing memory usage, computation, and system dependencies so that relatively large models can operate on devices with minimal RAM. It is written primarily in C and designed with a minimalist architecture that removes unnecessary...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    dataline

    dataline

    AI data analysis and visualization on CSV, Postgres, MySQL, Snowflake

    ...Once connected, users can generate tables, charts, and reports automatically based on queries produced by the AI engine. The platform is designed with a privacy-first architecture that stores data locally on the user’s device rather than sending it to external cloud services by default. It can also hide sensitive data from language models during processing, ensuring that only necessary metadata is used for query generation.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    OmAgent

    OmAgent

    Build multimodal language agents for fast prototype and production

    OmAgent is an open-source Python framework designed to simplify the development of multimodal language agents that can reason, plan, and interact with different types of data sources. The framework provides abstractions and infrastructure for building AI agents that operate on text, images, video, and audio while maintaining a relatively simple interface for developers.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    LightLLM

    LightLLM

    LightLLM is a Python-based LLM (Large Language Model) inference

    LightLLM is a high-performance inference and serving framework designed specifically for large language models, focusing on lightweight architecture, scalability, and efficient deployment. The framework enables developers to run and serve modern language models with significantly improved speed and resource efficiency compared to many traditional inference systems. Built primarily in Python, the project integrates optimization techniques and ideas from several leading open-source...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    Jovo Framework

    Jovo Framework

    The React for Voice and Chat, build apps for Alexa, Google Assistant

    The multimodal experience platform enables professional teams to build and run apps that work across smart speakers, the web, mobile, and more. Fully customizable and open source. The Jovo product ecosystem allows you to build, test, and run powerful experiences for voice, chat, and web platforms. From local development to production, Jovo allows you to build robust experiences, faster. Build across devices and platforms and use all supported modalities thanks to the Jovo output template engine. Our component and plugin architecture makes it possible to make Jovo work for your specific use case, across projects. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    LangGraph.js

    LangGraph.js

    Framework to build resilient language agents as graphs

    LangGraphJS is a JavaScript framework designed to build stateful AI applications and autonomous agents using graph-based execution models. Developed as part of the LangChain ecosystem, the framework allows developers to represent complex AI workflows as graphs where nodes represent tasks and edges define the flow of execution. This structure makes it easier to implement long-running agents, multi-step reasoning pipelines, and workflows that require persistent state. LangGraphJS supports...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    PowerInfer

    PowerInfer

    High-speed Large Language Model Serving for Local Deployment

    PowerInfer is a high-performance inference engine designed to run large language models efficiently on personal computers equipped with consumer-grade GPUs. The project focuses on improving the performance of local AI inference by optimizing how neural network computations are distributed between CPU and GPU resources. Its architecture exploits the observation that only a subset of neurons in large models are frequently activated, allowing the system to preload frequently used neurons into...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    Web Quality Skills

    Web Quality Skills

    Agent Skills for optimizing web quality based on Lighthouse

    This repository is a curated set of AI agent skills that encapsulate best practices for improving web quality, performance, accessibility, search engine optimization, and general best practices for web projects. It encodes knowledge drawn from Google Lighthouse audits, Core Web Vitals heuristics, WCAG accessibility guidelines, and real-world engineering experience, allowing coding agents to automatically assess and suggest improvements. These skills are framework-agnostic, meaning they apply...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    CTranslate2

    CTranslate2

    Fast inference engine for Transformer models

    CTranslate2 is a C++ and Python library for efficient inference with Transformer models. The project implements a custom runtime that applies many performance optimization techniques such as weights quantization, layers fusion, batch reordering, etc., to accelerate and reduce the memory usage of Transformer models on CPU and GPU. The execution is significantly faster and requires less resources than general-purpose deep learning frameworks on supported models and tasks thanks to many...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    .NET for Apache Spark

    .NET for Apache Spark

    A free, open-source, and cross-platform big data analytics framework

    .NET for Apache Spark provides high-performance APIs for using Apache Spark from C# and F#. With these .NET APIs, you can access the most popular Dataframe and SparkSQL aspects of Apache Spark, for working with structured data, and Spark Structured Streaming, for working with streaming data. .NET for Apache Spark is compliant with .NET Standard - a formal specification of .NET APIs that are common across .NET implementations. This means you can use .NET for Apache Spark anywhere you write...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 24
    Basic Memory

    Basic Memory

    Persistent AI memory using local Markdown knowledge graphs

    Basic Memory is an open source knowledge system that turns AI conversations into persistent, structured knowledge you control. Instead of losing context after each chat, it stores information as simple Markdown files on your device, allowing both you and AI to read and write to the same knowledge base. It uses the Model Context Protocol (MCP) so compatible AI tools can access, update, and build on your notes across sessions. Basic Memory creates a semantic knowledge graph by linking related...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    ChatLab

    ChatLab

    Local-first AI chat analysis tool for insights from conversation data

    ChatLab is an open source desktop application designed to help users analyze and better understand their personal chat histories through structured data exploration and AI-assisted insights. It enables users to import chat exports from multiple messaging platforms and transform them into a unified data model for consistent analysis. By combining a flexible SQL engine with AI agents, the tool allows users to query, summarize, and explore conversation patterns in a more interactive and intelligent way. ...
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB