35 projects for "llm api" with 1 filter applied:

  • Build on Google Cloud with $300 in Free Credit Icon
    Build on Google Cloud with $300 in Free Credit

    New to Google Cloud? Get $300 in free credit to explore Compute Engine, BigQuery, Cloud Run, Vertex AI, and 150+ other products.

    Start your next project with $300 in free Google Cloud credit. Spin up VMs, run containers, query exabytes in BigQuery, or build AI apps with Vertex AI and Gemini. Once your credits are used, keep building with 20+ products with free monthly usage, including Compute Engine, Cloud Storage, GKE, and Cloud Run functions. Sign up to start building right away.
    Start Free Trial
  • Managed MySQL, PostgreSQL, and SQL Databases on Google Cloud Icon
    Managed MySQL, PostgreSQL, and SQL Databases on Google Cloud

    Get back to your application and leave the database to us. Cloud SQL automatically handles backups, replication, and scaling.

    Cloud SQL is a fully managed relational database for MySQL, PostgreSQL, and SQL Server. We handle patching, backups, replication, encryption, and failover—so you can focus on your app. Migrate from on-prem or other clouds with free Database Migration Service. IDC found customers achieved 246% ROI. New customers get $300 in credits plus a 30-day free trial.
    Try Cloud SQL Free
  • 1
    Free LLM API resources

    Free LLM API resources

    A list of free LLM inference resources accessible via API

    Free LLM API resources repository curated by cheahjs is a community-driven index of free and open API endpoints, tools, datasets, runtimes, and utilities for working with large language models (LLMs) without cost-barriers. It collects a wide range of resources including hosted free-tier LLM APIs, documentation links, public model endpoints, open datasets useful for training or evaluation, tooling integrations, and examples showing how to interact with these services in real applications. ...
    Downloads: 12 This Week
    Last Update:
    See Project
  • 2
    Bifrost

    Bifrost

    The Fastest LLM Gateway with built in OTel observability

    Bifrost is an LLM gateway designed to provide a unified OpenAI-compatible API front for many different model providers. It abstracts away the complexity of working directly with multiple backend providers (OpenAI, Anthropic, AWS Bedrock, Google Vertex, etc.), enabling you to plug in providers and switch between them without touching your client code.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 3
    GPT4All

    GPT4All

    Run Local LLMs on Any Device. Open-source

    GPT4All is an open-source project that allows users to run large language models (LLMs) locally on their desktops or laptops, eliminating the need for API calls or GPUs. The software provides a simple, user-friendly application that can be downloaded and run on various platforms, including Windows, macOS, and Ubuntu, without requiring specialized hardware. It integrates with the llama.cpp implementation and supports multiple LLMs, allowing users to interact with AI models privately. This...
    Downloads: 199 This Week
    Last Update:
    See Project
  • 4
    Open WebUI

    Open WebUI

    User-friendly AI Interface

    Open WebUI is an extensible, feature-rich, and user-friendly self-hosted AI platform designed to operate entirely offline. It supports various LLM runners like Ollama and OpenAI-compatible APIs, with a built-in inference engine for Retrieval Augmented Generation (RAG), making it a powerful AI deployment solution. Key features include effortless setup via Docker or Kubernetes, seamless integration with OpenAI-compatible APIs, granular permissions and user groups for enhanced security,...
    Downloads: 80 This Week
    Last Update:
    See Project
  • Build AI Apps with Gemini 3 on Vertex AI Icon
    Build AI Apps with Gemini 3 on Vertex AI

    Access Google’s most capable multimodal models. Train, test, and deploy AI with 200+ foundation models on one platform.

    Vertex AI gives developers access to Gemini 3—Google’s most advanced reasoning and coding model—plus 200+ foundation models including Claude, Llama, and Gemma. Build generative AI apps with Vertex AI Studio, customize with fine-tuning, and deploy to production with enterprise-grade MLOps. New customers get $300 in free credits.
    Try Vertex AI Free
  • 5
    PasteGuard

    PasteGuard

    Masks sensitive data and secrets before they reach AI

    PasteGuard is an open-source privacy proxy that protects sensitive information like personal data and API secrets by detecting and masking them before they reach large language model APIs such as OpenAI or Anthropic Claude. It sits between an application and the LLM provider, automatically replacing names, emails, tokens, and other personally identifiable information (PII) with placeholders so that external services never see raw sensitive values, and then optionally unmasking them in the returned output. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    BricksLLM

    BricksLLM

    Enterprise-grade API gateway that helps you monitor and impose cost

    BricksLLM is an open-source framework for building and managing production-ready LLM (Large Language Model) applications. It provides tooling for prompt engineering, memory management, observability, and chaining, all in one unified developer experience. BricksLLM is designed to reduce boilerplate and increase the maintainability of LLM-based workflows.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    whisper.cpp

    whisper.cpp

    Port of OpenAI's Whisper model in C/C++

    whisper.cpp is a lightweight, C/C++ reimplementation of OpenAI’s Whisper automatic speech recognition (ASR) model—designed for efficient, standalone transcription without external dependencies. The entire high-level implementation of the model is contained in whisper.h and whisper.cpp. The rest of the code is part of the ggml machine learning library. The command downloads the base.en model converted to custom ggml format and runs the inference on all .wav samples in the folder samples....
    Downloads: 370 This Week
    Last Update:
    See Project
  • 8
    kgateway

    kgateway

    The Cloud-Native API Gateway and AI Gateway

    kgateway is a mature, cloud-native API and ingress gateway designed to provide unified API connectivity for services, microservices, serverless workloads, and AI-centric systems running on Kubernetes clusters. It implements the Kubernetes Gateway API and can operate as both a lightweight in-cluster microgateway and a centralized gateway capable of handling billions of API calls with high performance and low latency.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 9
    Memobase

    Memobase

    Fast backend for long-term AI user memory via structured profiles

    Memobase is an open source backend system that enables long-term user memory functionality for AI applications by capturing and structuring information about users across interactions. Its design centers on creating user profiles and recording event timelines, allowing AI systems to remember, understand, and evolve in their behaviour toward individual users over time. Instead of relying purely on traditional embedding-based retrieval or RAG systems, Memobase uses profile and timeline...
    Downloads: 5 This Week
    Last Update:
    See Project
  • AI-generated apps that pass security review Icon
    AI-generated apps that pass security review

    Stop waiting on engineering. Build production-ready internal tools with AI—on your company data, in your cloud.

    Retool lets you generate dashboards, admin panels, and workflows directly on your data. Type something like “Build me a revenue dashboard on my Stripe data” and get a working app with security, permissions, and compliance built in from day one. Whether on our cloud or self-hosted, create the internal software your team needs without compromising enterprise standards or control.
    Try Retool free
  • 10
    TONL

    TONL

    TONL (Token-Optimized Notation Language)

    ...It provides a serialization format that significantly reduces token usage compared with traditional JSON, which can result in lower costs and more efficient prompt size utilization in LLM-driven systems. TONL isn’t just a format — it includes a rich API for querying, indexing, modifying, and streaming data, along with tools for schema validation and TypeScript code generation. The platform comes with a complete command-line interface that supports interactive dashboards and cross-platform usage in browsers and server environments, and its high test coverage gives developers confidence in stability.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    ChatGLM-6B

    ChatGLM-6B

    ChatGLM-6B: An Open Bilingual Dialogue Language Model

    ChatGLM-6B is an open bilingual (Chinese + English) conversational language model based on the GLM architecture, with approximately 6.2 billion parameters. The project provides inference code, demos (command line, web, API), quantization support for lower memory deployment, and tools for finetuning (e.g., via P-Tuning v2). It is optimized for dialogue and question answering with a balance between performance and deployability in consumer hardware settings. Support for quantized inference...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 12
    Groq Python

    Groq Python

    The official Python Library for the Groq API

    Groq Python is the official Python SDK for the Groq REST API, giving Python developers straightforward access to Groq’s LLM, chat, audio, and other AI services. Through this library, you can call Groq’s models from Python code — for example to request chat completions, code generation, transcription, or any supported endpoint — using idiomatic Python syntax. The SDK handles authentication (via environment variable or parameter), defines proper type-safe request/response data types, and supports both synchronous and asynchronous usage patterns depending on your application needs. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Trae Agent

    Trae Agent

    LLM-based agent for general purpose software engineering tasks

    Trae Agent is an open-source, LLM-based agent system also developed by ByteDance, focused primarily on automating software engineering workflows. It provides a command-line interface (CLI) that accepts natural-language instructions (e.g. “refactor this module,” “write a unit test,” “generate a REST API skeleton”), and then orchestrates tool-based workflows — such as file editing, shell/batch commands, code generation, code formatting or refactoring — to carry out complex engineering tasks. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    BrowserAI

    BrowserAI

    Run local LLMs like llama, deepseek, kokoro etc. inside your browser

    BrowserAI is a cutting-edge platform that allows users to run large language models (LLMs) directly in their web browser without the need for a server. It leverages WebGPU for accelerated performance and supports offline functionality, making it a highly efficient and privacy-conscious solution. The platform provides a developer-friendly SDK with pre-configured popular models, and it allows for seamless switching between MLC and Transformer engines. Additionally, it supports features such as...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 15
    Open Responses

    Open Responses

    Specification for multi-provider, interoperable LLM interfaces

    Open Responses is an open-source implementation of an API compatible with the OpenAI Responses API that lets developers self-host a drop-in alternative endpoint for AI interactions while preserving compatibility with existing Agents SDKs and model workflows. It enables you to run a local or private server that speaks the standard Responses API, so tools, applications, and agents built against that API can operate without contacting OpenAI’s cloud and can instead route calls to any large...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 16
    Unsloth-MLX

    Unsloth-MLX

    Bringing the Unsloth experience to Mac users via Apple's MLX framework

    Unsloth-MLX offers developers the power of Unsloth’s efficient large language model fine-tuning experience on Apple Silicon Macs by wrapping Apple’s native MLX framework with an API fully compatible with Unsloth workflows. This project removes traditional barriers that prevent Mac users from prototyping and experimenting with LLM training locally by allowing the same code used in cloud GPU environments to run on M-series hardware, improving workflow continuity and reducing iteration costs. It supports loading and training Hugging Face models with fine-tuning strategies like SFT, DPO, ORPO, and GRPO and even handles exporting models to formats like GGUF for downstream use, although some limitations apply with quantized models. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Jesse

    Jesse

    An advanced crypto trading bot written in Python

    ...Because it’s framework-agnostic with respect to model providers, Jesse can work with different LLM backends, giving teams flexibility in cost and capability trade-offs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    MiniMax-M2.1

    MiniMax-M2.1

    MiniMax M2.1, a SOTA model for real-world dev & agents.

    MiniMax-M2.1 is an open-source, state-of-the-art agentic language model released to democratize high-performance AI capabilities. It goes beyond a simple parameter upgrade, delivering major gains in coding, tool use, instruction following, and long-horizon planning. The model is designed to be transparent, controllable, and accessible, enabling developers to build autonomous systems without relying on closed platforms. MiniMax-M2.1 excels in real-world software engineering tasks, including...
    Downloads: 10 This Week
    Last Update:
    See Project
  • 19
    Mastra

    Mastra

    The TypeScript AI agent framework

    ...At its core, Mastra provides abstractions for agents, workflows, tools, memory, retrieval, and model routing, so developers can focus on specifying behavior rather than wiring infrastructure from scratch. Model routing lets you connect to dozens of providers (OpenAI, Anthropic, Gemini, and others) through a single standardized interface, while agents orchestrate LLM calls and tools to solve open-ended tasks with internal reasoning loops. When explicit control is needed, Mastra’s workflow engine uses a graph-style API (.then(), .branch(), .parallel()) to orchestrate multi-step processes.
    Downloads: 10 This Week
    Last Update:
    See Project
  • 20
    Portkey AI Gateway

    Portkey AI Gateway

    A blazing fast AI Gateway with integrated guardrails

    Portkey AI Gateway aims to offer a blazing fast, secure, and flexible gateway for interacting with a wide variety of models and enforcing guardrails. It presents a single, friendly API through which you can route to 200+ LLMs, while applying configurable input/output guardrails to enforce policies or restrict certain content. It supports automatic retries, fallbacks, load balancing across providers or keys, and request timeouts to avoid latency spikes. The gateway is multimodal: it can...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    Groq AppGen

    Groq AppGen

    Project showcasing Llama 3.3 70B HTML codegen abilities

    Groq AppGen is an interactive web application (built with Next.js and TypeScript) that uses Groq’s LLM API to generate or modify web application code based on natural-language prompts. Essentially, you tell the app what kind of web app or page you want (in plain English), and groq-appgen will produce HTML/JSX code scaffolding, layout, and optionally application logic accordingly. It supports iterative feedback: you can refine your prompt, adjust parameters or requirements, and have the app regenerate or update the code — facilitating rapid prototyping and experimentation. ...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 22
    Poetiq

    Poetiq

    Reproduction of Poetiq's record-breaking submission to the ARC-AGI-1

    poetiq-arc-agi-solver is the open-source codebase from Poetiq that replicates their record-breaking submission to the challenging benchmark suite ARC-AGI (both ARC-AGI-1 and ARC-AGI-2). The project demonstrates a system that orchestrates large language models (LLMs) — like those from major providers — with carefully engineered prompting, reasoning workflows, and dynamic strategies, to tackle the abstract, logic-heavy problems in ARC-AGI. Instead of relying on a single prompt or fixed...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    Microsoft Learn MCP Server

    Microsoft Learn MCP Server

    Official Microsoft Learn MCP Server, powering LLMs and AI agents

    Microsoft Learn MCP Server is the official GitHub repository for the Microsoft Learn MCP (Model Context Protocol) Server, a service that implements the Model Context Protocol to provide AI assistants and tools with reliable, real-time access to Microsoft’s official documentation. Rather than relying on training data that may be outdated or incomplete, MCP servers let agents like GitHub Copilot, Claude, or other LLM-based tools search and pull context directly from up-to-date Microsoft Learn content, including Azure, .NET, and other tech docs. By connecting to the MCP endpoint, coding agents can answer questions, retrieve code examples, and offer best practices grounded in authoritative sources without requiring API keys or manual browser searches. ...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 24
    Groq TypeScript / Node.s

    Groq TypeScript / Node.s

    The official Node.js / Typescript library for the Groq API

    Groq TypeScript / Node.s (also often referred to as “groq-sdk” on npm) is the official Node.js / TypeScript client library for Groq’s REST API, enabling JavaScript/TypeScript developers to integrate LLM and AI-powered services into web backends, serverless functions, or frontend apps. It exports strongly-typed interfaces for models, chat completions, file uploads (e.g. for audio transcription), and other endpoints, allowing for better type safety and developer experience when using Groq from TypeScript. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Speech-AI-Forge

    Speech-AI-Forge

    Speech-AI-Forge is a project developed around TTS generation model

    Speech-AI-Forge is a full-stack project built around modern text-to-speech generation models, providing both an API server and a Gradio-based web UI for interactive use. At its core, it acts as a hub that wires together multiple speech-related capabilities, including TTS, speech-to-text and LLM-based control flows, behind a consistent interface. The system is designed to be deployed in several ways: you can try it online via hosted demos, spin it up in a one-click Colab environment, run it in Docker containers, or set it up locally with its environment preparation scripts. ...
    Downloads: 3 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • Next
MongoDB Logo MongoDB