• AI-generated apps that pass security review Icon
    AI-generated apps that pass security review

    Stop waiting on engineering. Build production-ready internal tools with AI—on your company data, in your cloud.

    Retool lets you generate dashboards, admin panels, and workflows directly on your data. Type something like “Build me a revenue dashboard on my Stripe data” and get a working app with security, permissions, and compliance built in from day one. Whether on our cloud or self-hosted, create the internal software your team needs without compromising enterprise standards or control.
    Try Retool free
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    More flexibility. More control.

    Generate interest, access liquidity without selling, and execute trades seamlessly. All in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 1
    Easy DataSet

    Easy DataSet

    A powerful tool for creating datasets for LLM fine-tuning

    Easy DataSet is a comprehensive open-source tool designed to make creating high-quality datasets for large language model fine-tuning, retrieval-augmented generation (RAG), and evaluation as easy and automated as possible by providing intuitive interfaces and powerful parsing, segmentation, and labeling tools. It supports ingesting domain-specific documents in a wide range of formats — including PDF, Markdown, DOCX, EPUB, and plain text — and can intelligently segment, clean, and structure content into rich datasets tailored for downstream LLM training needs. ...
    Downloads: 12 This Week
    Last Update:
    See Project
  • 2
    LangServe

    LangServe

    Helps developers deploy LangChain runnables and chains as a REST API

    LangServe is an open-source deployment framework designed to expose LangChain applications as production-ready REST APIs. The tool simplifies the process of turning language-model pipelines, chains, and agents into web services that can be accessed by external applications. Instead of manually writing API endpoints, developers can use LangServe to automatically generate a server that exposes LangChain workflows through HTTP interfaces. The framework is built on top of FastAPI and uses Pydantic for request validation and structured data handling. ...
    Downloads: 10 This Week
    Last Update:
    See Project
  • 3
    Transformer Explainer

    Transformer Explainer

    Learn How LLM Transformer Models Work with Interactive Visualization

    Transformer Explainer is an interactive visualization tool created to help users understand how transformer-based language models operate internally. The platform runs a lightweight GPT-2 model directly in the user’s browser and allows users to experiment with text prompts while observing the model’s internal operations. Through visual diagrams and interactive interfaces, the tool reveals how tokens are processed through layers such as embeddings, attention mechanisms, and feed-forward networks. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 4
    Superagent

    Superagent

    Superagent protects your AI applications

    Superagent is an open-source AI safety platform built to protect applications from prompt injections, data leaks, and harmful outputs. It embeds real-time safety directly into AI workflows, helping teams secure models before threats cause damage. Superagent provides guardrails that block jailbreaks, prompt manipulation, and sensitive data exfiltration. It includes redaction tools to remove PII, PHI, and secrets automatically from text. The platform also scans code repositories to detect...
    Downloads: 4 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 5
    Generative AI for beginners with JS

    Generative AI for beginners with JS

    Join a time-traveling adventure where you meet history’s legends

    ...Topics covered include prompt engineering, building AI-powered applications, working with structured outputs, integrating retrieval-augmented generation, and enabling tool or function calling in AI systems. The repository focuses specifically on how generative AI can be integrated into web, mobile, or desktop applications using JavaScript frameworks and APIs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    AI Agents From Scratch

    AI Agents From Scratch

    Demystify AI agents by building them yourself. Local LLMs

    ...The project walks through the process of constructing agents step by step, beginning with simple prompt-based interactions and gradually introducing more advanced capabilities such as planning, tool use, and memory. The repository provides example implementations that demonstrate how language models can interact with external systems, perform reasoning tasks, and execute structured workflows. It focuses on explaining the architecture of agent systems rather than simply providing finished code, making it useful for developers who want to understand how AI agents actually work internally. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Easy-Vibe

    Easy-Vibe

    Tutorial on Product Prototype, AI Capability Integration

    ...The learning path is divided into progressive stages that cover beginner concepts, full-stack development, and advanced multi-platform application development. Throughout the curriculum, learners explore topics such as prompt engineering, AI tool integration, product prototyping, and deployment strategies for AI-enabled applications.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    gpu_poor

    gpu_poor

    Calculate token/s & GPU memory requirement for any LLM

    gpu_poor is an open-source tool designed to help developers determine whether their hardware is capable of running a specific large language model and to estimate the performance they can expect from it. The project focuses on calculating GPU memory requirements and predicted inference speed for different models, hardware configurations, and quantization strategies.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB