Open Source JavaScript Large Language Models (LLM)

JavaScript Large Language Models (LLM)

View 361 business solutions

Browse free open source JavaScript Large Language Models (LLM) and projects below. Use the toggles on the left to filter open source JavaScript Large Language Models (LLM) by OS, license, language, programming language, and project status.

  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 1
    SillyTavern

    SillyTavern

    LLM Frontend for Power Users

    Mobile-friendly, Multi-API (KoboldAI/CPP, Horde, NovelAI, Ooba, OpenAI, OpenRouter, Claude, Scale), VN-like Waifu Mode, Horde SD, System TTS, WorldInfo (lorebooks), customizable UI, auto-translate, and more prompt options than you'd ever want or need. Optional Extras server for more SD/TTS options + ChromaDB/Summarize. SillyTavern is a user interface you can install on your computer (and Android phones) that allows you to interact with text generation AIs and chat/roleplay with characters you or the community create. SillyTavern is a fork of TavernAI 1.2.8 which is under more active development and has added many major features. At this point, they can be thought of as completely independent programs.
    Downloads: 498 This Week
    Last Update:
    See Project
  • 2
    AnythingLLM

    AnythingLLM

    The all-in-one Desktop & Docker AI application with full RAG and AI

    A full-stack application that enables you to turn any document, resource, or piece of content into a context that any LLM can use as references during chatting. This application allows you to pick and choose which LLM or Vector Database you want to use as well as supporting multi-user management and permissions. AnythingLLM is a full-stack application where you can use commercial off-the-shelf LLMs or popular open-source LLMs and vectorDB solutions to build a private ChatGPT with no compromises that you can run locally as well as host remotely and be able to chat intelligently with any documents you provide it. AnythingLLM divides your documents into objects called workspaces. A Workspace functions a lot like a thread, but with the addition of containerization of your documents. Workspaces can share documents, but they do not talk to each other so you can keep your context for each workspace clean.
    Downloads: 157 This Week
    Last Update:
    See Project
  • 3
    LandPPT

    LandPPT

    An LLM-based presentation generation platform

    LandPPT is an open-source AI platform that automatically generates professional presentation slides using large language models. The system allows users to create complete PowerPoint presentations simply by entering a topic or uploading source documents such as PDFs, Word files, or Markdown notes. Using natural language processing and structured content generation, the platform produces presentation outlines and converts them into fully formatted slide decks. The application integrates multiple AI models from providers such as OpenAI, Anthropic, Google, and locally hosted models to generate text, images, and structured presentation layouts. It also includes template systems and style options that allow presentations to be customized for different industries, visual themes, or storytelling formats.
    Downloads: 9 This Week
    Last Update:
    See Project
  • 4
    LLM Datasets

    LLM Datasets

    Curated list of datasets and tools for post-training

    LLM Datasets curates and standardizes datasets commonly used to train and fine-tune large language models, reducing the overhead of hunting down sources and normalizing formats. The repository aims to make datasets easy to inspect and transform, with scripts for downloading, deduping, cleaning, and converting to formats like JSONL that slot into training pipelines. It highlights instruction-tuning and conversation-style corpora while also pointing to code, math, or domain-specific sets for targeted capabilities. Quality is a recurring theme: examples and utilities help filter low-value samples, enforce length limits, and split train/validation consistently so results are comparable. Licensing and provenance are surfaced to encourage compliant usage and to guide dataset selection in commercial settings. For practitioners, the repo is a practical “starting pantry” that accelerates experimentation and helps keep data wrangling from dominating the project timeline.
    Downloads: 8 This Week
    Last Update:
    See Project
  • Forever Free Full-Stack Observability | Grafana Cloud Icon
    Forever Free Full-Stack Observability | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 5
    Superagent

    Superagent

    Superagent protects your AI applications

    Superagent is an open-source AI safety platform built to protect applications from prompt injections, data leaks, and harmful outputs. It embeds real-time safety directly into AI workflows, helping teams secure models before threats cause damage. Superagent provides guardrails that block jailbreaks, prompt manipulation, and sensitive data exfiltration. It includes redaction tools to remove PII, PHI, and secrets automatically from text. The platform also scans code repositories to detect AI-specific attack vectors like repo poisoning. Superagent is designed for low-latency production environments and works with any major LLM provider. It enables teams to prove compliance with modern AI security and regulatory standards.
    Downloads: 7 This Week
    Last Update:
    See Project
  • 6
    Easy DataSet

    Easy DataSet

    A powerful tool for creating datasets for LLM fine-tuning

    Easy DataSet is a comprehensive open-source tool designed to make creating high-quality datasets for large language model fine-tuning, retrieval-augmented generation (RAG), and evaluation as easy and automated as possible by providing intuitive interfaces and powerful parsing, segmentation, and labeling tools. It supports ingesting domain-specific documents in a wide range of formats — including PDF, Markdown, DOCX, EPUB, and plain text — and can intelligently segment, clean, and structure content into rich datasets tailored for downstream LLM training needs. The system includes automated question-generation capabilities, hierarchical label trees, and answer generation pipelines that use LLM APIs to produce coherent paired data with customizable templates. Beyond dataset creation, Easy-dataset also provides a built-in evaluation system with model testing and blind-test features, helping teams validate model performance using curated test sets.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 7
    Aix-DB

    Aix-DB

    Based on the LangChain/LangGraph framework

    Aix-DB is an open-source intelligent data analysis platform that combines large language models with database technologies to enable conversational data exploration. The system is designed as a ChatBI solution that allows users to query datasets using natural language and receive structured insights, charts, and visualizations automatically. Built on frameworks such as LangChain and LangGraph, Aix-DB integrates retrieval-augmented generation and Text-to-SQL capabilities to convert user questions into executable database queries. The platform supports multiple types of data sources and provides an end-to-end pipeline that includes intent recognition, SQL generation, database execution, and visual presentation of results. Its architecture includes multiple layers such as a web interface, API gateway, AI service layer, and data storage layer that support relational databases, vector stores, graph databases, and file systems.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 8
    CSGHub

    CSGHub

    CSGHub is a brand-new open-source platform for managing LLMs

    CSGHub is an open-source framework designed for collaborative scientific research and content generation. It enables researchers to utilize AI-driven tools for literature review, hypothesis generation, and automated writing assistance, streamlining the scientific discovery process.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 9
    Easy-Vibe

    Easy-Vibe

    Tutorial on Product Prototype, AI Capability Integration

    Easy-Vibe is an open-source educational project designed to teach developers, product managers, and beginners how to build AI-powered applications using the emerging concept of “vibe coding,” a development approach that relies heavily on AI-assisted programming tools. The project provides a structured curriculum that guides learners from having no programming experience to building fully functional AI-integrated applications. Instead of focusing only on theoretical concepts, Easy-Vibe emphasizes practical, step-by-step tutorials that demonstrate how to transform product ideas into working software using modern AI coding tools and development workflows. The learning path is divided into progressive stages that cover beginner concepts, full-stack development, and advanced multi-platform application development. Throughout the curriculum, learners explore topics such as prompt engineering, AI tool integration, product prototyping, and deployment strategies for AI-enabled applications.
    Downloads: 1 This Week
    Last Update:
    See Project
  • AI-generated apps that pass security review Icon
    AI-generated apps that pass security review

    Stop waiting on engineering. Build production-ready internal tools with AI—on your company data, in your cloud.

    Retool lets you generate dashboards, admin panels, and workflows directly on your data. Type something like “Build me a revenue dashboard on my Stripe data” and get a working app with security, permissions, and compliance built in from day one. Whether on our cloud or self-hosted, create the internal software your team needs without compromising enterprise standards or control.
    Try Retool free
  • 10
    GPU Hot

    GPU Hot

    Real-time NVIDIA GPU dashboard

    GPU Hot is an open-source, lightweight monitoring dashboard designed to provide real-time visibility into NVIDIA GPU performance across single machines or entire clusters. The project offers a self-hosted web interface that streams hardware metrics directly from GPU servers, enabling developers, ML engineers, and system administrators to observe GPU utilization and system behavior in real time through a browser. The dashboard collects and displays a wide range of performance metrics including temperature, memory usage, power consumption, clock speeds, fan speed, and active processes. It can scale from monitoring a single GPU workstation to large distributed environments with dozens or even hundreds of GPUs by running lightweight containers on each node and aggregating the data centrally.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    Lemon AI

    Lemon AI

    Full-stack Open-source Self-Evolving General AI Agent

    LemonAI is an open-source full-stack framework for building autonomous AI agents capable of performing complex tasks such as research, programming, data analysis, and document processing. The platform is designed to run primarily on local infrastructure, providing a privacy-focused alternative to cloud-dependent agent platforms. It integrates with local large language models through tools such as Ollama, vLLM, and other model runtimes while also allowing optional connections to external cloud models. The system includes a multi-agent architecture that supports planning, action execution, reflection, and memory, allowing the agent to reason through tasks and refine results iteratively. A key component of the framework is a virtual machine sandbox environment that safely executes code generated by the agent without affecting the host system.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 12
    AI Agents From Scratch

    AI Agents From Scratch

    Demystify AI agents by building them yourself. Local LLMs

    AI Agents from Scratch is an educational repository designed to teach developers how to build autonomous AI agents using large language models and modern AI frameworks. The project walks through the process of constructing agents step by step, beginning with simple prompt-based interactions and gradually introducing more advanced capabilities such as planning, tool use, and memory. The repository provides example implementations that demonstrate how language models can interact with external systems, perform reasoning tasks, and execute structured workflows. It focuses on explaining the architecture of agent systems rather than simply providing finished code, making it useful for developers who want to understand how AI agents actually work internally. By building agents incrementally, the project helps learners grasp concepts such as decision loops, task decomposition, and environment interaction.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    ChatLLM Web

    ChatLLM Web

    Chat with LLM like Vicuna totally in your browser with WebGPU

    Chat with LLM like Vicuna totally in your browser with WebGPU, safely, privately, and with no server. Powered By web-llm. To use this app, you need a browser that supports WebGPU, such as Chrome 113 or Chrome Canary. Chrome versions ≤ 112 are not supported. You will need a GPU with about 6.4GB of memory. If your GPU has less memory, the app will still run, but the response time will be slower. The first time you use the app, you will need to download the model. For the Vicuna-7b model that we are currently using, the download size is about 4GB. After the initial download, the model will be loaded from the browser cache for faster usage.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Generative AI for beginners with JS

    Generative AI for beginners with JS

    Join a time-traveling adventure where you meet history’s legends

    Generative AI with JavaScript is an educational repository created by Microsoft that teaches developers how to build applications powered by large language models using the JavaScript ecosystem. The project is structured as a multi-lesson curriculum that introduces the concepts, tools, and practical techniques required to create generative AI applications. Each lesson includes written explanations, hands-on exercises, quizzes, and supporting videos to help developers learn the material progressively. Topics covered include prompt engineering, building AI-powered applications, working with structured outputs, integrating retrieval-augmented generation, and enabling tool or function calling in AI systems. The repository focuses specifically on how generative AI can be integrated into web, mobile, or desktop applications using JavaScript frameworks and APIs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    LIDA

    LIDA

    Automatic Generation of Visualizations and Infographics using LLMs

    LIDA is an open-source library developed to automate the process of creating data visualizations and infographics using large language models. The system treats visualizations as executable code and uses AI to generate, modify, and interpret that code in order to transform raw datasets into meaningful charts and graphical explanations. Instead of requiring users to manually explore datasets and write plotting scripts, LIDA analyzes the data and automatically proposes visualization goals and design ideas that highlight patterns and relationships. The platform can generate visualization code compatible with a wide range of libraries, allowing it to integrate with common data science ecosystems. It also supports iterative workflows where visualizations can be edited, explained, evaluated, and repaired through AI-driven feedback loops. The system is model-agnostic and can connect to multiple language model providers, enabling flexibility across different AI infrastructures.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    LLM Course

    LLM Course

    Course to get into Large Language Models (LLMs)

    LLM Course is a hands-on, notebook-driven path for learning how large language models work in practice, from data curation to training, fine-tuning, evaluating, and deploying. It emphasizes reproducible experiments: each step is demonstrated with runnable code, clear dependencies, and references to commonly used open-source models and libraries. Learners get exposure to multiple adaptation strategies—LoRA/QLoRA, instruction fine-tuning, and alignment techniques—so they can choose approaches that fit their hardware and budgets. The materials also cover inference optimization and quantization to make serving LLMs feasible on commodity GPUs or even CPUs, which is crucial for side projects and startups. Evaluation is treated as a first-class topic, with examples of automatic and human-in-the-loop methods to catch regressions and verify quality beyond simple loss values. By the end, students have a mental model and a practical toolkit for iterating on datasets, training configs, etc.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    LangServe

    LangServe

    Helps developers deploy LangChain runnables and chains as a REST API

    LangServe is an open-source deployment framework designed to expose LangChain applications as production-ready REST APIs. The tool simplifies the process of turning language-model pipelines, chains, and agents into web services that can be accessed by external applications. Instead of manually writing API endpoints, developers can use LangServe to automatically generate a server that exposes LangChain workflows through HTTP interfaces. The framework is built on top of FastAPI and uses Pydantic for request validation and structured data handling. It also includes client libraries that allow developers to interact with deployed chains from Python or JavaScript applications. LangServe is commonly used to deploy AI applications such as chatbots, document analysis pipelines, and agent-based systems that require scalable access through APIs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    Learn Prompting

    Learn Prompting

    This website is a free, open-source guide on prompt engineering

    This website is a free, open-source guide on prompt engineering. Contributions are welcome! Harsh criticism is welcome too. We launched the first ever prompt hacking competition designed to enhance AI safety and education by challenging participants to outsmart large language models from May 5th to June 3rd! The competition featured 10 increasingly difficult levels of prompt hacking defenses and the chance to win over $35,000 in prizes. Coding is a great skill to learn alongside prompt engineering. We recommend learning Python, as it is a popular language for AI and machine learning. Be among the first to access the certification program as soon as it launches.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    NLUX

    NLUX

    The powerful Conversational AI JavaScript Library

    NLUX, short for Natural Language User Experience, is an open-source JavaScript and React library designed to simplify the creation of conversational interfaces powered by large language models. The library provides developers with prebuilt components and utilities that make it easy to integrate chat-based AI functionality into web applications. By using NLUX, developers can connect their applications to models such as ChatGPT or other LLM providers and create interactive conversational interfaces with minimal setup. The framework supports both React components and standalone JavaScript implementations, giving developers flexibility in how they integrate AI functionality into existing applications. NLUX focuses on improving the user experience of AI-powered systems by providing structured messaging interfaces, streaming responses, and customizable conversational UI elements.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    RAG from Scratch

    RAG from Scratch

    Demystify RAG by building it from scratch

    RAG From Scratch is an educational open-source project designed to teach developers how retrieval-augmented generation systems work by building them step by step. Instead of relying on complex frameworks or cloud services, the repository demonstrates the entire RAG pipeline using transparent and minimal implementations. The project walks through key concepts such as generating embeddings, building vector databases, retrieving relevant documents, and integrating the retrieved context into language model prompts. Each example is written with detailed explanations so that developers can understand the internal mechanics of semantic search and context-aware language generation. The repository emphasizes learning through direct implementation, allowing users to see how each component of the RAG architecture functions independently.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    Transformer Explainer

    Transformer Explainer

    Learn How LLM Transformer Models Work with Interactive Visualization

    Transformer Explainer is an interactive visualization tool created to help users understand how transformer-based language models operate internally. The platform runs a lightweight GPT-2 model directly in the user’s browser and allows users to experiment with text prompts while observing the model’s internal operations. Through visual diagrams and interactive interfaces, the tool reveals how tokens are processed through layers such as embeddings, attention mechanisms, and feed-forward networks. Users can observe how attention weights change as the model predicts the next token, offering insight into how transformer architectures capture relationships between words. The design of the platform emphasizes educational accessibility, allowing students, researchers, and developers to explore complex machine learning concepts without requiring specialized hardware or installations.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    Wiseflow

    Wiseflow

    Enhance any agent's browser use skill

    Wiseflow is an open-source information extraction and knowledge discovery system designed to collect, filter, and organize valuable information from large volumes of online content. The platform continuously monitors specified sources such as websites, social platforms, and other digital channels to identify relevant data according to user-defined interests or topics. By combining web crawling, content parsing, and large language model analysis, the system extracts concise insights from raw information streams and converts them into structured data that can be stored or analyzed. This automated workflow helps reduce the noise associated with large information ecosystems and highlights the most important insights for users. Wiseflow can automatically categorize extracted content, assign tags, and upload processed results into databases or knowledge systems for further use.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    browserable

    browserable

    Open source and self-hostable browser automation library for AI agents

    Browserable is an open-source browser automation framework designed specifically for AI agents that need to interact with web interfaces in a human-like way. The project provides tools that allow automated agents to navigate websites, click buttons, fill out forms, and extract information from pages without manual scripting of each step. Built primarily in JavaScript, the framework offers both a developer-friendly SDK and a REST API that allow integration with AI applications and automation pipelines. It is designed to be self-hostable, which means developers can deploy and run it on their own infrastructure without relying on third-party services. The platform enables the creation of browser-based agents capable of performing complex online workflows such as data collection, research tasks, and automated interactions with web platforms.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    gpu_poor

    gpu_poor

    Calculate token/s & GPU memory requirement for any LLM

    gpu_poor is an open-source tool designed to help developers determine whether their hardware is capable of running a specific large language model and to estimate the performance they can expect from it. The project focuses on calculating GPU memory requirements and predicted inference speed for different models, hardware configurations, and quantization strategies. By analyzing factors such as model size, context length, batch size, and GPU specifications, the system estimates how much VRAM will be required and how fast tokens can be generated during inference. The tool also provides a detailed breakdown of where GPU memory is allocated, including model weights, KV cache, activations, and other runtime overhead. This information allows developers to evaluate trade-offs between different quantization methods such as GGML, bitsandbytes, and QLoRA before attempting to deploy a model. gpu_poor is particularly useful for researchers and hobbyists.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    super-agent-party

    super-agent-party

    All-in-one AI companion! Desktop girlfriend + virtual streamer

    Super Agent Party is an open-source experimental framework designed to demonstrate collaborative multi-agent AI systems interacting within a shared environment. The project explores how multiple specialized AI agents can coordinate to solve complex tasks by communicating with each other and sharing information. Instead of relying on a single monolithic model, the framework organizes agents with different roles or capabilities that cooperate to achieve goals. Each agent may handle different responsibilities such as planning, execution, reasoning, or knowledge retrieval, allowing the system to tackle more complex problems than a single agent might handle alone. The platform is primarily intended as a research and demonstration environment for experimenting with agent collaboration strategies. Developers can use it to study coordination patterns, communication protocols, and task decomposition in multi-agent systems.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB