Showing 42 open source projects for "linux pxe server"

View related business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Our Free Plans just got better! | Auth0 Icon
    Our Free Plans just got better! | Auth0

    With up to 25k MAUs and unlimited Okta connections, our Free Plan lets you focus on what you do best—building great apps.

    You asked, we delivered! Auth0 is excited to expand our Free and Paid plans to include more options so you can focus on building, deploying, and scaling applications without having to worry about your security. Auth0 now, thank yourself later.
    Try free now
  • 1
    tldw Server

    tldw Server

    Your Personal Research Multi-Tool

    tldw-server (mirror) is a mirrored distribution of an open-source backend service designed to store, process, and serve summarized information extracted from long pieces of content. The name “tldw” reflects the phrase “too long; didn’t watch,” which refers to tools that condense lengthy videos, articles, or documents into concise summaries. The server component typically acts as the core infrastructure that manages summaries, metadata, and retrieval operations for client applications or user...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    csghub-server

    csghub-server

    csghub-server is the backend server for CSGHub

    csghub-server is the backend component of the CSGHub platform, an open-source infrastructure designed to manage and operate large language models, datasets, and AI development workflows within a private deployment environment. The server acts as a centralized management layer that allows teams to store, organize, and operate AI assets such as models, datasets, and machine learning applications in a manner similar to artifact repositories used in software engineering. Built primarily in the...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Slack MCP Server

    Slack MCP Server

    The most powerful MCP Slack Server with no permission requirements

    Slack MCP Server is an open-source server implementation that connects Slack workspaces to AI systems through the Model Context Protocol (MCP). MCP is a standardized protocol that allows large language models and AI agents to securely interact with external tools and data sources such as messaging platforms, databases, or file systems. The slack-mcp-server acts as an intermediary layer that exposes Slack data and messaging functionality to AI clients while enforcing access rules and...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Google Workspace MCP Server

    Google Workspace MCP Server

    Control Gmail, Google Calendar, Docs, Sheets, Slides, Chat, Forms

    Google Workspace MCP is an open-source server that connects AI assistants to Google Workspace services through the Model Context Protocol (MCP), allowing large language models to interact directly with productivity tools. The project exposes a wide set of Google services including Gmail, Google Drive, Docs, Sheets, Slides, Calendar, Chat, and other Workspace components as structured tools that an AI system can call programmatically. By acting as a bridge between AI clients and the Google...
    Downloads: 1 This Week
    Last Update:
    See Project
  • Go From AI Idea to AI App Fast Icon
    Go From AI Idea to AI App Fast

    One platform to build, fine-tune, and deploy ML models. No MLOps team required.

    Access Gemini 3 and 200+ models. Build chatbots, agents, or custom models with built-in monitoring and scaling.
    Try Free
  • 5
    SillyTavern

    SillyTavern

    LLM Frontend for Power Users

    Mobile-friendly, Multi-API (KoboldAI/CPP, Horde, NovelAI, Ooba, OpenAI, OpenRouter, Claude, Scale), VN-like Waifu Mode, Horde SD, System TTS, WorldInfo (lorebooks), customizable UI, auto-translate, and more prompt options than you'd ever want or need. Optional Extras server for more SD/TTS options + ChromaDB/Summarize. SillyTavern is a user interface you can install on your computer (and Android phones) that allows you to interact with text generation AIs and chat/roleplay with characters...
    Downloads: 384 This Week
    Last Update:
    See Project
  • 6
    JADX-AI-MCP

    JADX-AI-MCP

    Plugin for JADX to integrate MCP server

    JADX-AI-MCP is an open-source plugin that integrates large language models into the JADX Android decompiler to assist with reverse engineering and code analysis tasks. The project connects JADX with AI assistants through the Model Context Protocol, enabling language models to interact directly with decompiled Android application code. Through this integration, AI systems can inspect classes, analyze methods, retrieve application manifests, and examine other elements of Android packages in...
    Downloads: 8 This Week
    Last Update:
    See Project
  • 7
    mistral.rs

    mistral.rs

    Fast, flexible LLM inference

    mistral.rs is a fast and flexible LLM inference engine implemented in Rust, designed to run and serve modern language models with an emphasis on performance and practical deployment. It provides multiple entry points for developers, including a CLI for running models locally and an HTTP server that exposes an OpenAI-compatible API surface for easy integration with existing clients. The project includes hardware-aware tooling that can benchmark a system and choose sensible quantization and...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 8
    wllama

    wllama

    WebAssembly binding for llama.cpp - Enabling on-browser LLM inference

    wllama is a WebAssembly-based library that enables large language model inference directly inside a web browser. Built as a binding for the llama.cpp inference engine, the project allows developers to run LLM models locally without requiring a server backend or dedicated GPU hardware. The library leverages WebAssembly SIMD capabilities to achieve efficient execution within modern browsers while maintaining compatibility across platforms. By running models locally on the user’s device, wllama...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 9
    Flowise

    Flowise

    Drag & drop UI to build your customized LLM flow

    Open source UI visual tool to build your customized LLM flow using LangchainJS, written in Node Typescript/Javascript. Conversational agent for a chat model which utilizes chat-specific prompts and buffer memory. Open source is the core of Flowise, and it will always be free for commercial and personal usage. Flowise support different environment variables to configure your instance. You can specify the following variables in the .env file inside the packages/server folder.
    Downloads: 19 This Week
    Last Update:
    See Project
  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • 10
    DevDocs by CyberAGI

    DevDocs by CyberAGI

    Completely free, private, UI based Tech Documentation MCP server

    DevDocs is an open-source documentation server designed to provide developers with a private, structured interface for browsing and interacting with technical documentation using AI tools. The system functions as a Model Context Protocol (MCP) server that allows large language models and developer assistants to access technical documentation in a structured and efficient way. Instead of sending entire documents to a language model, DevDocs organizes documentation into sections so that only...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    AingDesk

    AingDesk

    AI assistant that supports knowledge bases, model APIs

    AingDesk is an open-source desktop and server-based AI assistant platform designed to provide a user-friendly environment for interacting with language models and building AI-powered tools. The software enables users to run local AI models or connect to external model APIs through a unified interface. One of its primary goals is to simplify the process of building knowledge-based assistants by allowing users to create local knowledge bases that the AI can search and analyze. The system...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 12
    Aidea

    Aidea

    Flutter-based cross-platform app integrating major AI models

    AIdea is a comprehensive Flutter-based cross-platform app integrating major AI models—OpenAI GPT, Chinese models Tongyi Qianwen and Wenxin Yiyan, plus image models like Stable Diffusion for text-to-image, image-to-image, SDXL 1.0, super-resolution, and colorization. It includes a client app, server backend, and Docker deployment scripts for hosted setups.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Secret Llama

    Secret Llama

    Fully private LLM chatbot that runs entirely with a browser

    Secret Llama is a privacy-first large-language-model chatbot that runs entirely inside your web browser, meaning no server is required and your conversation data never leaves your device. It focuses on open-source model support, letting you load families like Llama and Mistral directly in the client for fully local inference. Because everything happens in-browser, it can work offline once models are cached, which is helpful for air-gapped environments or travel. The interface mirrors the...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 14
    ChatGLM3

    ChatGLM3

    ChatGLM3 series: Open Bilingual Chat LLMs | Open Source Bilingual Chat

    ChatGLM3 is ZhipuAI & Tsinghua KEG’s third-gen conversational model suite centered on the 6B-parameter ChatGLM3-6B. It keeps the series’ smooth dialog and low deployment cost while adding native tool use (function calling), a built-in code interpreter, and agent-style workflows. The family includes base and long-context variants (8K/32K/128K). The repo ships Python APIs, CLI and web demos (Gradio/Streamlit), an OpenAI-format API server, and a compact fine-tuning kit. Quantization (4/8-bit),...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 15
    Reader 3

    Reader 3

    Quick illustration of how one can easily read books together with LLMs

    This project is a minimalist, self-hosted EPUB reader designed to help users browse and read EPUB books one chapter at a time through a lightweight local server, making it especially easy to extract or work with chapters in external tools like large language models. It was created primarily as a simple demonstration of how to combine local book reading with LLM workflows without heavy dependencies or complicated setup, and it runs with just a small Python script and a basic HTTP server. The...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 16
    Grounded Docs

    Grounded Docs

    Open-Source Alternative to Context7, Nia, and Ref.Tools

    Grounded Docs is an open-source implementation of a Model Context Protocol server designed to expose documentation and structured information as tools that AI agents can query. The project allows language models and agent frameworks to retrieve and interact with documentation through standardized MCP interfaces. By acting as an intermediary layer between documentation sources and AI tools, the server enables models to access structured documentation in a consistent and machine-readable...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    webclaw

    webclaw

    Fast, local-first web content extraction for LLMs

    webclaw is a high-performance web content extraction tool designed specifically for AI agents and large language models, focusing on delivering clean, structured data instead of raw HTML. It is built in Rust and operates without a headless browser, using advanced techniques such as TLS fingerprinting to bypass common scraping barriers and mimic real browser behavior. The tool addresses a major inefficiency in AI workflows by removing irrelevant elements like navigation menus, ads, and...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 18
    WebLLM

    WebLLM

    Bringing large-language models and chat to web browsers

    WebLLM is a modular, customizable javascript package that directly brings language model chats directly onto web browsers with hardware acceleration. Everything runs inside the browser with no server support and is accelerated with WebGPU. We can bring a lot of fun opportunities to build AI assistants for everyone and enable privacy while enjoying GPU acceleration. WebLLM offers a minimalist and modular interface to access the chatbot in the browser. The WebLLM package itself does not come...
    Downloads: 10 This Week
    Last Update:
    See Project
  • 19
    React Native AI

    React Native AI

    Full stack framework for building cross-platform mobile AI apps

    React Native AI is a full-stack framework designed to simplify the development of AI-powered mobile applications using React Native. The project provides a ready-to-use infrastructure for building cross-platform apps that integrate large language models and other AI services. It supports real-time streaming responses from multiple AI providers and enables developers to build chat interfaces, AI-driven image generation tools, and natural language features within mobile apps. The framework...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 20
    Anyquery

    Anyquery

    Query anything (GitHub, Notion, +40 more) with SQL and let LLMs

    Anyquery is an open-source SQL query engine designed to allow users to query data from almost any source using a unified SQL interface. The system enables developers and analysts to run SQL queries on files, APIs, applications, and databases without needing separate connectors or query languages for each platform. Built on top of SQLite, the engine uses a plugin architecture that allows it to extend support to dozens of external services and data sources. Users can query structured files...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    node-llama-cpp

    node-llama-cpp

    Run AI models locally on your machine with node.js bindings for llama

    node-llama-cpp is a JavaScript and Node.js binding that allows developers to run large language models locally using the high-performance inference engine provided by llama.cpp. The library enables applications built with Node.js to interact directly with local LLM models without requiring a remote API or external service. By using native bindings and optimized model execution, the framework allows developers to integrate advanced language model capabilities into desktop applications, server...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 22
    CogVLM2

    CogVLM2

    GPT4V-level open-source multi-modal model based on Llama3-8B

    CogVLM2 is the second generation of the CogVLM vision-language model series, developed by ZhipuAI and released in 2024. Built on Meta-Llama-3-8B-Instruct, CogVLM2 significantly improves over its predecessor by providing stronger performance across multimodal benchmarks such as TextVQA, DocVQA, and ChartQA, while introducing extended context length support of up to 8K tokens and high-resolution image input up to 1344×1344. The series includes models for both image understanding and video...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 23
    MCP Router

    MCP Router

    A Unified MCP Server Management App (MCP Manager)

    MCP Router is an open-source management platform designed to simplify the deployment and coordination of Model Context Protocol (MCP) servers used by AI agents. MCP is an emerging standard that allows language models and AI assistants to connect to external tools, data sources, and services through a structured interface. The MCP Router project acts as a centralized manager that helps developers run, configure, and coordinate multiple MCP servers within a single environment. This enables AI...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 24
    Farfalle

    Farfalle

    AI search engine - self-host with local or cloud LLMs

    Farfalle is an open-source AI-powered search engine designed to provide an answer-centric search experience similar to modern conversational search systems. The project integrates large language models with multiple search APIs so that the system can gather information from external sources and synthesize responses into concise answers. It can run either with local language models or with cloud-based providers, allowing developers to deploy it privately or integrate with hosted AI services....
    Downloads: 1 This Week
    Last Update:
    See Project
  • 25
    SeaGOAT

    SeaGOAT

    local-first semantic code search engine

    SeaGOAT is an open-source semantic code search engine designed to help developers explore and understand large codebases more efficiently. Instead of relying solely on traditional keyword search, it uses vector embeddings to represent the meaning of code and queries, allowing users to perform semantic searches that find relevant code even when the exact keywords are not present. The tool runs locally on a developer’s machine and processes repositories using a combination of embedding models...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • Next
MongoDB Logo MongoDB