Showing 40 open source projects for "llama-gguf-split"

View related business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Application Monitoring That Won't Slow Your App Down Icon
    Application Monitoring That Won't Slow Your App Down

    AppSignal's Rust-based agent is lightweight and stable. Already running in thousands of production apps.

    Full APM with errors, performance, logs, and uptime monitoring. 99.999% uptime SLA on the platform itself.
    Start Free
  • 1
    Llama Coder

    Llama Coder

    Open source Claude Artifacts – built with Llama 3.1 405B

    Llama Coder is an open-source tool that lets you generate small applications (often React or web apps) from a single natural-language prompt using the Llama 3 family of models. It’s framed as an open-source “Claude Artifacts”-style experience: you describe the app you want, the tool calls an LLM hosted on Together.ai, and you get back a runnable code artifact.
    Downloads: 12 This Week
    Last Update:
    See Project
  • 2
    Secret Llama

    Secret Llama

    Fully private LLM chatbot that runs entirely with a browser

    Secret Llama is a privacy-first large-language-model chatbot that runs entirely inside your web browser, meaning no server is required and your conversation data never leaves your device. It focuses on open-source model support, letting you load families like Llama and Mistral directly in the client for fully local inference. Because everything happens in-browser, it can work offline once models are cached, which is helpful for air-gapped environments or travel.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 3
    node-llama-cpp

    node-llama-cpp

    Run AI models locally on your machine with node.js bindings for llama

    node-llama-cpp is a JavaScript and Node.js binding that allows developers to run large language models locally using the high-performance inference engine provided by llama.cpp. The library enables applications built with Node.js to interact directly with local LLM models without requiring a remote API or external service. By using native bindings and optimized model execution, the framework allows developers to integrate advanced language model capabilities into desktop applications, server software, and command-line tools. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 4
    Clippy

    Clippy

    Clippy, now with some AI

    ...The project serves as both a playful homage to the early days of personal computing and a practical demonstration of local AI inference. Clippy integrates with the llama.cpp runtime to run models directly on a user’s computer without requiring cloud-based AI services. It supports models in the GGUF format, which allows it to run many publicly available open-source LLMs efficiently on consumer hardware. Users interact with the system through a simple animated assistant interface that can answer questions, generate text, and perform conversational tasks. The application includes one-click installation support for several popular models such as Meta’s Llama, Google’s Gemma, and other open models.
    Downloads: 23 This Week
    Last Update:
    See Project
  • Our Free Plans just got better! | Auth0 Icon
    Our Free Plans just got better! | Auth0

    With up to 25k MAUs and unlimited Okta connections, our Free Plan lets you focus on what you do best—building great apps.

    You asked, we delivered! Auth0 is excited to expand our Free and Paid plans to include more options so you can focus on building, deploying, and scaling applications without having to worry about your security. Auth0 now, thank yourself later.
    Try free now
  • 5
    llama.vscode

    llama.vscode

    VS Code extension for LLM-assisted code/text completion

    llama.vscode is a Visual Studio Code extension that provides AI-assisted coding features powered primarily by locally running language models. The extension is designed to be lightweight and efficient, enabling developers to use AI tools even on consumer-grade hardware. It integrates with the llama.cpp runtime to run language models locally, eliminating the need to rely entirely on external APIs or cloud providers. The extension supports common AI development features such as code...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 6
    Jan.ai

    Jan.ai

    Open source alternative to ChatGPT that runs 100% offline

    ...It allows you to download and run LLMs (local language models) offline while also offering optional integration with cloud-based model providers—giving you full control over your data and AI interactions. Download and run LLMs (Llama, Gemma, Qwen, GPT-oss etc.) from HuggingFace. Connect to GPT models via OpenAI, Claude models via Anthropic, Mistral, Groq, and others. Create specialized AI assistants for your tasks. MCP integration for agentic capabilities.
    Downloads: 52 This Week
    Last Update:
    See Project
  • 7
    Ollama Server

    Ollama Server

    No need for Termux, you can start the Ollama service

    ...This makes it especially useful for developers who want to prototype or deploy local AI workflows on mobile hardware while maintaining compatibility with existing tooling. The application includes basic model lifecycle management such as downloading official models, uploading custom GGUF files, and controlling running instances. It also enables offline inference, allowing users to run language models entirely on-device without requiring a network connection or external infrastructure.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 8
    Eliza

    Eliza

    Autonomous agents for everyone

    ...Full support for voice, text, and media interactions. Built-in RAG memory system, document processing, media analysis, and autonomous trading capabilities. Supports multiple AI models including Llama, GPT-4, and Claude. Create custom actions, add new platform integrations, and extend functionality through a modular plugin system. Full TypeScript support.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 9
    wllama

    wllama

    WebAssembly binding for llama.cpp - Enabling on-browser LLM inference

    wllama is a WebAssembly-based library that enables large language model inference directly inside a web browser. Built as a binding for the llama.cpp inference engine, the project allows developers to run LLM models locally without requiring a server backend or dedicated GPU hardware. The library leverages WebAssembly SIMD capabilities to achieve efficient execution within modern browsers while maintaining compatibility across platforms. By running models locally on the user’s device, wllama...
    Downloads: 4 This Week
    Last Update:
    See Project
  • Forever Free Full-Stack Observability | Grafana Cloud Icon
    Forever Free Full-Stack Observability | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 10
    promptfoo

    promptfoo

    Evaluate and compare LLM outputs, catch regressions, improve prompts

    ...Use built-in metrics, LLM-graded evals, or define your own custom metrics. Compare prompts and model outputs side-by-side, or integrate the library into your existing test/CI workflow. Use OpenAI, Anthropic, and open-source models like Llama and Vicuna, or integrate custom API providers for any LLM API.
    Downloads: 7 This Week
    Last Update:
    See Project
  • 11
    critique

    critique

    TUI for reviewing git changes

    critique is a beautiful terminal-oriented user interface tool for reviewing git diffs that makes inspecting source control changes more intuitive and readable directly from the command line. The tool provides a styled, split-view diff layout with syntax highlighting and word-level diffing, which gives developers clear insight into what has changed in each file beyond simple line additions or deletions. It supports viewing diff ranges across commits, staged versus unstaged changes, and even comparisons between branches, turning the diff output into a more interactive, readable experience than traditional git diff. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Terminus

    Terminus

    A terminal for a more modern age

    Terminus is a highly configurable terminal emulator for Windows, macOS and Linux. Features an integrated SSH client and connection manager. Provides theming and color schemes, fully configurable shortcuts, and split panes. Remembers your tabs. With PowerShell (and PS Core), WSL, Git-Bash, Cygwin, Cmder and CMD support. Enables direct file transfer from/to SSH sessions via Zmodem. Full Unicode support including double-width characters. Terminus doesn't choke on fast-flowing outputs. Allows proper shell experience on Windows including tab completion (via Clink). ...
    Downloads: 29 This Week
    Last Update:
    See Project
  • 13
    Claude Canvas

    Claude Canvas

    Give Claude Code an external monitor

    Claude Canvas is a terminal-focused UI toolkit that extends Claude Code by giving it a dedicated visual interface within the terminal, allowing interactive panes for apps like email, calendar, flight bookings, and other structured interfaces directly alongside the coding agent session. Rather than limiting interactions to text prompts and responses, Claude-Canvas uses tools like tmux to spawn multiple split panes so that you can see persistent interfaces for tasks that benefit from visual context, making your CL-based AI coding workspace feel more like a rich interactive environment. It acts as a proof-of-concept that bridges traditional command-line AI interactions with more dynamic terminal user interfaces, enabling developers to craft more immersive and contextually rich workflows without leaving their development terminal.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Bee Agent Framework

    Bee Agent Framework

    The framework for building scalable agentic applications

    ...The Bee Agent Framework makes it easy to build scalable agent-based workflows with your model of choice. The framework is been designed to perform robustly with IBM Granite and Llama 3.x models, and we’re actively working on optimizing its performance with other popular LLMs. Our goal is to empower developers to adopt the latest open-source and proprietary models with minimal changes to their current agent implementation.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    dockview

    dockview

    Zero dependency docking layout manager supporting tabs

    dockview is a zero-dependency docking layout manager written in TypeScript that enables developers to create highly dynamic, IDE-like interfaces with draggable panels, tabs, and resizable layouts. It supports multiple layout paradigms, including split views, grid layouts, and dockable panels, allowing complex UI arrangements similar to tools like Visual Studio Code. The library is framework-agnostic at its core, with official bindings for React, Vue, and Angular, as well as support for vanilla TypeScript. Dockview includes advanced features such as floating panels, popout windows, and persistent layout serialization, enabling users to save and restore custom workspace configurations. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    BrowserAI

    BrowserAI

    Run local LLMs like llama, deepseek, kokoro etc. inside your browser

    BrowserAI is a cutting-edge platform that allows users to run large language models (LLMs) directly in their web browser without the need for a server. It leverages WebGPU for accelerated performance and supports offline functionality, making it a highly efficient and privacy-conscious solution. The platform provides a developer-friendly SDK with pre-configured popular models, and it allows for seamless switching between MLC and Transformer engines. Additionally, it supports features such as...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 17
    OpenWork

    OpenWork

    An open-source alternative to Claude Cowork, powered by opencode

    OpenWork is a framework for building decentralized collaborative work environments powered by AI and human contributions. At its core, the project enables contributors to define tasks, workflows, and goals that can be split, shared, and recombined across distributed nodes while agents and humans cooperate to advance progress. It offers structured templates for work items, decision logic for task allocation, and consensus mechanisms that let groups verify and validate results toward shared objectives. This project also includes moderation and reputation layers so that contributor trust and quality can be assessed and integrated into future task assignments. ...
    Downloads: 19 This Week
    Last Update:
    See Project
  • 18
    express-openapi-validator

    express-openapi-validator

    Auto-validates api requests, responses, and securities using ExpressJS

    Auto-validates api requests, responses, and securities using ExpressJS and an OpenAPI 3.x specification. Express-openapi-validator is an unopinionated library that integrates with new and existing API applications. express-openapi-validator lets you write code the way you want; it does not impose any coding convention or project layout. Simply, install the validator onto your express app, point it to your OpenAPI 3 specification, then define and implement routes the way you prefer. An...
    Downloads: 15 This Week
    Last Update:
    See Project
  • 19
    Tribe AI

    Tribe AI

    Low code tool to rapidly build and coordinate multi-agent teams

    Low code tool to rapidly build and coordinate multi-agent teams. Have you heard the saying, 'Two minds are better than one'? That's true for agents too. Tribe leverages on the langgraph framework to let you customize and coordinate teams of agents easily. By splitting up tough tasks among agents who are good at different things, each one can focus on what it does best. This makes solving problems faster and better.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    LLM Scraper

    LLM Scraper

    Extract structured data from webpages using LLM-powered scraping

    LLM Scraper is a TypeScript library designed to extract structured data from webpages using large language models. Instead of relying on fragile HTML selectors or manual parsing rules, the tool interprets webpage content with language models and converts it into structured data according to a defined schema. Developers can specify the data structure using tools such as Zod or JSON Schema, enabling the model to extract relevant information directly into typed objects. LLM Scraper integrates...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 21
    Bit

    Bit

    A tool for component-driven application development

    Bit is the platform for component-driven development. Forget monoliths and distribute app development to components. Enjoy better scale, speed, and consistency. Join 200k+ developers and start free. Empower teams in your organization to deliver features and innovate autonomously while continuously collaborating on each other's components and building world-class products together. Join the world's best teams on the enterprise component cloud. Say goodbye to monolithic web apps, and hello to...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 22
    Groq AppGen

    Groq AppGen

    Project showcasing Llama 3.3 70B HTML codegen abilities

    Groq AppGen is an interactive web application (built with Next.js and TypeScript) that uses Groq’s LLM API to generate or modify web application code based on natural-language prompts. Essentially, you tell the app what kind of web app or page you want (in plain English), and groq-appgen will produce HTML/JSX code scaffolding, layout, and optionally application logic accordingly. It supports iterative feedback: you can refine your prompt, adjust parameters or requirements, and have the app...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 23
    Generative AI for Beginners (Version 3)

    Generative AI for Beginners (Version 3)

    21 Lessons, Get Started Building with Generative AI

    Generative AI for Beginners is a 21-lesson course by Microsoft Cloud Advocates that teaches the fundamentals of building generative AI applications in a practical, project-oriented way. Lessons are split into “Learn” modules for core concepts and “Build” modules with hands-on code in Python and TypeScript, so you can jump in at any point that matches your goals. The course covers everything from model selection, prompt engineering, and chat/text/image app patterns to secure development practices and UX for AI. It also walks through modern application techniques such as function calling, RAG with vector databases, working with open source models, agents, fine-tuning, and using SLMs. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 24
    LibPDF

    LibPDF

    A modern PDF library for TypeScript

    LibPDF-js/core is a modern, TypeScript-first PDF processing library that provides a comprehensive toolkit for parsing, modifying, and generating PDF documents with a clean, intuitive API designed to handle real-world files safely and robustly. Unlike many existing JavaScript PDF libraries, it emphasizes lenient parsing that can gracefully handle malformed structures and fallback strategies where typical parsers fail, making it useful for production environments that encounter unpredictable...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    react-resizable-panels

    react-resizable-panels

    React components for resizable panel groups/layouts

    react-resizable-panels provides accessible, flexible split-view layouts for React without wrestling low-level pointer math yourself. It models layouts as “groups” of panels separated by resize handles, supporting horizontal or vertical orientation and mixed min/max constraints that prevent panels from collapsing unexpectedly. The library focuses on smooth interactions: dragging, double-clicking to reset, and keyboard resizing for users who rely on accessibility features.
    Downloads: 1 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • Next
MongoDB Logo MongoDB