Open Source TypeScript Large Language Models (LLM)

TypeScript Large Language Models (LLM)

View 375 business solutions

Browse free open source TypeScript Large Language Models (LLM) and projects below. Use the toggles on the left to filter open source TypeScript Large Language Models (LLM) by OS, license, language, programming language, and project status.

  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    Access competitive interest rates on your digital assets.

    Generate interest, borrow against your crypto, and trade a range of cryptocurrencies — all in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • Go from Code to Production URL in Seconds Icon
    Go from Code to Production URL in Seconds

    Cloud Run deploys apps in any language instantly. Scales to zero. Pay only when code runs.

    Skip the Kubernetes configs. Cloud Run handles HTTPS, scaling, and infrastructure automatically. Two million requests free per month.
    Try it free
  • 1
    Langflow

    Langflow

    Low-code app builder for RAG and multi-agent AI applications

    Langflow is a low-code app builder for RAG and multi-agent AI applications. It’s Python-based and agnostic to any model, API, or database.
    Downloads: 57 This Week
    Last Update:
    See Project
  • 2
    Clippy

    Clippy

    Clippy, now with some AI

    Clippy is an open-source desktop assistant that allows users to run modern large language models locally while presenting them through a nostalgic interface inspired by Microsoft’s classic Clippy assistant from the 1990s. The project serves as both a playful homage to the early days of personal computing and a practical demonstration of local AI inference. Clippy integrates with the llama.cpp runtime to run models directly on a user’s computer without requiring cloud-based AI services. It supports models in the GGUF format, which allows it to run many publicly available open-source LLMs efficiently on consumer hardware. Users interact with the system through a simple animated assistant interface that can answer questions, generate text, and perform conversational tasks. The application includes one-click installation support for several popular models such as Meta’s Llama, Google’s Gemma, and other open models.
    Downloads: 41 This Week
    Last Update:
    See Project
  • 3
    Flowise

    Flowise

    Drag & drop UI to build your customized LLM flow

    Open source UI visual tool to build your customized LLM flow using LangchainJS, written in Node Typescript/Javascript. Conversational agent for a chat model which utilizes chat-specific prompts and buffer memory. Open source is the core of Flowise, and it will always be free for commercial and personal usage. Flowise support different environment variables to configure your instance. You can specify the following variables in the .env file inside the packages/server folder.
    Downloads: 35 This Week
    Last Update:
    See Project
  • 4
    OmniRoute

    OmniRoute

    OmniRoute is an AI gateway for multi-provider LLM

    OmniRoute is a routing and orchestration framework designed to simplify the handling of requests, workflows, or data flows across multiple services or endpoints in a unified manner. It focuses on providing a flexible abstraction layer where developers can define routing logic that dynamically directs traffic based on conditions, context, or predefined rules. The project emphasizes modularity and extensibility, allowing users to plug in different services or handlers without tightly coupling components. It is particularly useful in distributed systems where requests need to be intelligently routed between APIs, microservices, or processing pipelines. OmniRoute aims to reduce boilerplate by centralizing routing logic and providing reusable patterns for managing complex flows. Its architecture supports scalability and maintainability, making it suitable for both small applications and larger systems with multiple integrations.
    Downloads: 34 This Week
    Last Update:
    See Project
  • Go From AI Idea to AI App Fast Icon
    Go From AI Idea to AI App Fast

    One platform to build, fine-tune, and deploy ML models. No MLOps team required.

    Access Gemini 3 and 200+ models. Build chatbots, agents, or custom models with built-in monitoring and scaling.
    Try Free
  • 5
    FreedomGPT

    FreedomGPT

    React and Electron-based app that executes the FreedomGPT LLM locally

    FreedomGPT is a locally executed large language model (LLM) application built using React and Electron, allowing users to interact with AI models privately on their Mac or Windows devices. The app enables offline operation, ensuring privacy and security while providing a chat-based interface for seamless communication with the AI. It supports integration with models like Liberty Edge and offers an open-source solution for those seeking more control over their AI interactions. The app's setup is simple, and it includes clear installation guides for both macOS and Windows platforms, as well as detailed instructions for building necessary libraries like llama.cpp.
    Downloads: 26 This Week
    Last Update:
    See Project
  • 6
    Dify

    Dify

    One API for plugins and datasets, one interface for prompt engineering

    Dify is an easy-to-use LLMOps platform designed to empower more people to create sustainable, AI-native applications. With visual orchestration for various application types, Dify offers out-of-the-box, ready-to-use applications that can also serve as Backend-as-a-Service APIs. Unify your development process with one API for plugins and datasets integration, and streamline your operations using a single interface for prompt engineering, visual analytics, and continuous improvement. Out-of-the-box web sites supporting form mode and chat conversation mode A single API encompassing plugin capabilities, context enhancement, and more, saving you backend coding effort Visual data analysis, log review, and annotation for applications
    Downloads: 11 This Week
    Last Update:
    See Project
  • 7
    OpenClaude

    OpenClaude

    Claude Code opened to any LLM

    OpenClaude is an open-source alternative or extension inspired by Claude-style agent systems, designed to provide similar capabilities in a customizable and self-hosted environment. The project focuses on enabling users to run their own AI agents with full control over data, workflows, and integrations, reducing reliance on proprietary platforms. It likely includes support for executing tasks, managing context, and interacting with external tools, allowing agents to perform real-world actions beyond simple text generation. The architecture emphasizes flexibility, enabling developers to adapt the system to different use cases and integrate it with various models or APIs. It may also include modular components for extending functionality, such as plugins or skills. The project reflects a broader trend toward open and decentralized AI systems that prioritize transparency and control.
    Downloads: 9 This Week
    Last Update:
    See Project
  • 8
    Gemini Fullstack LangGraph Quickstart

    Gemini Fullstack LangGraph Quickstart

    Get started w/ building Fullstack Agents using Gemini 2.5 & LangGraph

    gemini-fullstack-langgraph-quickstart is a fullstack reference application from Google DeepMind’s Gemini team that demonstrates how to build a research-augmented conversational AI system using LangGraph and Google Gemini models. The project features a React (Vite) frontend and a LangGraph/FastAPI backend designed to work together seamlessly for real-time research and reasoning tasks. The backend agent dynamically generates search queries based on user input, retrieves information via the Google Search API, and performs reflective reasoning to identify knowledge gaps. It then iteratively refines its search until it produces a comprehensive, well-cited answer synthesized by the Gemini model. The repository provides both a browser-based chat interface and a command-line script (cli_research.py) for executing research queries directly. For production deployment, the backend integrates with Redis and PostgreSQL to manage persistent memory, streaming outputs, & background task coordination.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 9
    Hollama

    Hollama

    A minimal LLM chat app that runs entirely in your browser

    Hollama is a lightweight open-source chat application designed to run entirely within the browser while interacting with large language model servers. The project provides a minimal but powerful user interface for communicating with local or remote LLMs, including servers powered by Ollama or OpenAI-compatible APIs. Because the application runs as a static web interface, it does not require complex backend infrastructure and can be easily deployed or self-hosted. Hollama supports both text-based and multimodal interactions, allowing users to work with models that process images as well as text. The interface includes features for editing prompts, retrying responses, copying generated code snippets, and storing conversation history locally within the browser. Mathematical expressions can be rendered using KaTeX, and Markdown formatting allows code blocks and structured outputs to appear clearly within conversations.
    Downloads: 5 This Week
    Last Update:
    See Project
  • AI-generated apps that pass security review Icon
    AI-generated apps that pass security review

    Stop waiting on engineering. Build production-ready internal tools with AI—on your company data, in your cloud.

    Retool lets you generate dashboards, admin panels, and workflows directly on your data. Type something like “Build me a revenue dashboard on my Stripe data” and get a working app with security, permissions, and compliance built in from day one. Whether on our cloud or self-hosted, create the internal software your team needs without compromising enterprise standards or control.
    Try Retool free
  • 10
    Pluely

    Pluely

    The Open Source Alternative to Cluely

    Pluely is an open-source AI automation framework designed to simplify the development and deployment of AI-driven workflows across applications and services. The system focuses on orchestrating tasks performed by large language models and other AI components, allowing developers to define structured workflows where models interact with tools, APIs, and external systems. By providing a modular architecture for building AI pipelines, the platform enables developers to connect multiple processing steps such as data retrieval, prompt execution, analysis, and response generation. The project emphasizes flexibility, allowing developers to extend the platform with custom integrations and automation logic. This makes the framework suitable for building intelligent assistants, automated business workflows, and data-processing pipelines that rely on generative AI capabilities.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 11
    Vercel AI SDK

    Vercel AI SDK

    Build AI-powered applications with React, Svelte, Vue, and Solid

    The Vercel AI SDK is a library for building AI-powered streaming text and chat UIs.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 12
    node-llama-cpp

    node-llama-cpp

    Run AI models locally on your machine with node.js bindings for llama

    node-llama-cpp is a JavaScript and Node.js binding that allows developers to run large language models locally using the high-performance inference engine provided by llama.cpp. The library enables applications built with Node.js to interact directly with local LLM models without requiring a remote API or external service. By using native bindings and optimized model execution, the framework allows developers to integrate advanced language model capabilities into desktop applications, server software, and command-line tools. The system automatically detects the available hardware on a machine and selects the most appropriate compute backend, including CPU or GPU acceleration. Developers can use the library to perform tasks such as text generation, conversational chat, embedding generation, and structured output generation. Because it runs models locally, the platform is particularly useful for privacy-sensitive environments or offline AI deployments.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 13
    oh my PI

    oh my PI

    AI Coding agent for the terminal

    Oh-My-Pi is an open-source AI agent toolkit focused on creating intelligent coding assistants that operate directly from the terminal environment. The project provides a command-line coding agent capable of analyzing repositories, generating commits, editing code, and interacting with development tools through an integrated tool system. Instead of functioning as a simple prompt-based assistant, the system includes an agent architecture that can inspect Git repositories, analyze changes, and perform development actions with fine-grained control. The platform also supports tool-based workflows where the agent can run shell commands, read files, modify code, and stage changes during development tasks. It includes infrastructure for integrating different AI providers and models through a unified API layer, allowing developers to switch between models while keeping the same agent interface.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 14
    Quivr

    Quivr

    Your Second Brain supercharged by Generative AI

    Quivr, your second brain, utilizes the power of GenerativeAI to store and retrieve unstructured information. Think of it as Obsidian, but turbocharged with AI capabilities.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 15
    Reader LLM

    Reader LLM

    Convert any URL to an LLM-friendly input with a simple prefix

    Reader LLM is an open-source tool designed to convert web content into formats that are easier for large language models to process. The system works by transforming a webpage into a clean text or Markdown representation that removes unnecessary formatting and highlights the core information within the page. Developers can use a simple URL prefix to retrieve a version of a webpage that has been optimized for machine consumption, making it suitable for use in AI agents or retrieval-augmented generation pipelines. In addition to converting individual pages, the service can perform web searches and return relevant content that can be ingested directly by AI systems. The tool relies on specialized models and parsing techniques to handle complex HTML structures and extract meaningful content while preserving important context.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 16
    AI as Workspace

    AI as Workspace

    An elegant AI chat client. Full-featured, lightweight

    AI as Workspace, short for AI as Workspace, is an open-source AI client application that provides a unified interface for interacting with multiple large language models and AI tools within a single workspace environment. The platform is designed as a lightweight yet powerful desktop or web application that organizes AI interactions through structured workspaces. Instead of managing individual chat sessions separately, users can group conversations, artifacts, and tasks within customizable workspaces that support different projects or contexts. AIaW supports multiple AI providers and models through a flexible interface compatible with common API formats used by services such as OpenAI-style endpoints. The application also includes a plugin system that allows developers to extend the platform with additional capabilities such as automation tools, integrations, or custom AI utilities.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 17
    DocStrange

    DocStrange

    Extract and convert data from any document, images, pdfs, word doc

    DocStrange is an open-source document understanding and extraction library designed to convert complex files into structured, LLM-ready outputs such as Markdown, JSON, CSV, and HTML. Developed by Nanonets, the project combines OCR, layout detection, table understanding, and structured extraction into one end-to-end pipeline, which reduces the need to stitch together multiple separate services. It is built for developers who need high-quality parsing from scans, photos, PDFs, office files, and other document sources while preserving privacy and control over the processing flow. One of its key differentiators is deployment flexibility: it offers a cloud API for managed usage as well as a fully private offline mode that runs locally on a GPU. The platform also supports synchronous extraction, streaming responses, and asynchronous processing for larger documents, which makes it adaptable to both interactive workflows and heavier back-end pipelines.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 18
    LangGraph.js

    LangGraph.js

    Framework to build resilient language agents as graphs

    LangGraphJS is a JavaScript framework designed to build stateful AI applications and autonomous agents using graph-based execution models. Developed as part of the LangChain ecosystem, the framework allows developers to represent complex AI workflows as graphs where nodes represent tasks and edges define the flow of execution. This structure makes it easier to implement long-running agents, multi-step reasoning pipelines, and workflows that require persistent state. LangGraphJS supports advanced capabilities such as branching logic, loops, and conditional execution, enabling developers to build sophisticated AI systems that can adapt to dynamic conditions. The framework integrates seamlessly with language models, tools, and external APIs, allowing agents to retrieve information and perform actions across different systems. Developers can also build applications that maintain conversation history and state across multiple interactions.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 19
    Read Frog

    Read Frog

    Open Source Immersive Translate

    Read Frog is an open-source browser extension designed to transform everyday web reading into an immersive language learning experience powered by artificial intelligence. The tool integrates translation, contextual explanations, and content analysis directly into the browsing workflow so users can learn languages naturally while reading authentic online content. Instead of forcing learners to switch between translation tools and the original text, the extension displays translations alongside the source language, making comprehension immediate and continuous. The system automatically extracts the main content of an article using intelligent parsing techniques, allowing users to focus on the most relevant text without distractions. AI models are used to generate summaries, introductions, and explanations for words, phrases, and sentences based on the learner’s language level, making the experience personalized and adaptive.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 20
    Superset LLM

    Superset LLM

    Run an army of Claude Code, Codex, etc. on your machine

    Superset is a development environment and terminal-based platform designed to orchestrate multiple AI coding agents simultaneously within a single workspace. The tool enables developers to run many autonomous coding agents in parallel without the typical overhead of manually managing multiple terminals, repositories, or branches. Each agent task is isolated in its own Git worktree, ensuring that code changes from different agents do not interfere with each other while allowing developers to track their progress independently. The platform includes built-in monitoring capabilities so users can observe the activity of each agent, receive notifications when tasks are completed, and quickly review changes produced by automated coding workflows. Superset also integrates tools for reviewing code differences, editing generated outputs, and managing the development environment directly from the interface.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 21
    opensrc

    opensrc

    Fetch source code for npm packages

    OpenSrc is an open-source utility developed by Vercel Labs that retrieves and exposes the source code of npm packages so that AI coding agents can better understand how external libraries work. When large language models generate code, they often rely only on type definitions or documentation, which can limit their understanding of how a library actually behaves. OpenSrc addresses this limitation by allowing agents to fetch the underlying source code of dependencies and analyze their implementation directly. This gives AI coding assistants richer context about functions, internal logic, and architectural patterns used within external packages. The tool is designed to integrate into AI-driven developer workflows where coding agents explore repositories, inspect dependencies, and reason about how to use libraries correctly.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 22
    Langfuse

    Langfuse

    Open source LLM engineering platform: LLM Observability, metrics, etc.

    Langfuse is a logging and analytics tool for large language model (LLM) applications, providing insights into usage, performance, and troubleshooting.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 23
    MCP Router

    MCP Router

    A Unified MCP Server Management App (MCP Manager)

    MCP Router is an open-source management platform designed to simplify the deployment and coordination of Model Context Protocol (MCP) servers used by AI agents. MCP is an emerging standard that allows language models and AI assistants to connect to external tools, data sources, and services through a structured interface. The MCP Router project acts as a centralized manager that helps developers run, configure, and coordinate multiple MCP servers within a single environment. This enables AI applications to access multiple tools and knowledge sources through a unified interface rather than connecting to each service individually. The project provides infrastructure for routing requests between clients and MCP servers, enabling scalable multi-tool agent systems. Developers building AI agents can use the platform to manage tool endpoints, control service availability, and simplify agent integration workflows.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 24
    Nanocoder

    Nanocoder

    A beautiful local-first coding agent running in your terminal

    Nanocoder is an open-source, local-first coding assistant that runs in the command line and allows developers to use AI models to assist with programming tasks directly from their terminal environment. The tool is designed as a privacy-focused alternative to proprietary AI coding assistants, allowing users to run local models or connect to external APIs while keeping full control over their data and development workflow. Built with TypeScript and distributed as a CLI application, nanocoder enables developers to interact with AI agents that can read files, modify code, execute commands, and assist with debugging tasks. The platform supports multiple AI providers through OpenAI-compatible APIs and can also integrate with local model runtimes such as Ollama or LM Studio. Its architecture emphasizes extensibility through custom commands and integration with Model Context Protocol servers that allow the AI agent to access additional tools.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 25
    NativeMind Extension

    NativeMind Extension

    Your fully private, open-source, on-device AI assistant

    NativeMindExtension is an open-source browser extension that provides a private, on-device AI assistant designed to run without cloud dependencies. The project is built around a privacy-first model in which conversations, document analysis, translations, and writing assistance stay on the user’s device rather than being sent to external servers. It integrates with local model back ends such as Ollama and also supports WebLLM for quick in-browser trials, giving users a choice between stronger local setups and lighter no-install demonstrations. The extension is aimed at everyday browser workflows, offering features like multi-tab context awareness, webpage summarization, document understanding, contextual toolbars, and AI-assisted rewriting directly inside the browsing experience. Because it runs locally after setup, it is also positioned as an always-available assistant that avoids API quotas, network latency, and service outages common in cloud-based AI tools.
    Downloads: 2 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • 4
  • Next
MongoDB Logo MongoDB