Showing 344 open source projects for "lightweight programming language"

View related business solutions
  • Compliant and Reliable File Transfers Backed by Top Security Certifications Icon
    Compliant and Reliable File Transfers Backed by Top Security Certifications

    Cerberus FTP Server delivers SOC 2 Type II certified security and FIPS 140-2 validated encryption.

    Stop relying on non-certified, legacy file transfer tools that creak under the weight of modern security demands. Get full audit trails, advanced access controls and more supported by an award-winning team of experts. Start your free 25-day trial today.
    Start Free Trial
  • Go From AI Idea to AI App Fast Icon
    Go From AI Idea to AI App Fast

    One platform to build, fine-tune, and deploy ML models. No MLOps team required.

    Access Gemini 3 and 200+ models. Build chatbots, agents, or custom models with built-in monitoring and scaling.
    Try Free
  • 1
    Magicoder

    Magicoder

    Empowering Code Generation with OSS-Instruct

    Magicoder is an open-source family of large language models designed specifically for code generation and software development tasks. The project focuses on improving the quality and diversity of code generation by training models with a novel dataset construction approach known as OSS-Instruct. This technique uses open-source code repositories as a foundation for generating more realistic and diverse instruction datasets for training language models.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    LiteLLM

    LiteLLM

    lightweight package to simplify LLM API calls

    Call all LLM APIs using the OpenAI format [Anthropic, Huggingface, Cohere, Azure OpenAI etc.] liteLLM supports streaming the model response back, pass stream=True to get a streaming iterator in response. Streaming is supported for OpenAI, Azure, Anthropic, and Huggingface models.
    Downloads: 11 This Week
    Last Update:
    See Project
  • 3
    Search with Lepton

    Search with Lepton

    Lightweight demo to build a conversational AI search engine quickly

    Search with Lepton is an open source demonstration project that shows how to build a conversational search engine using the Lepton AI framework. It combines traditional web search with large language models to provide natural language answers to user queries. It retrieves information from supported search engines and uses that context to generate responses through a retrieval-augmented generation approach. The implementation is intentionally minimal, containing fewer than 500 lines of code...
    Downloads: 8 This Week
    Last Update:
    See Project
  • 4
    autoMate

    autoMate

    AI tool for automating desktop tasks via natural language input

    autoMate is an AI-powered local automation tool designed to enable users to control and automate their computers using natural language instructions instead of traditional scripting or rule-based systems. It combines large language models with computer vision techniques to interpret user intent and understand on-screen content, allowing it to interact with graphical interfaces similarly to a human user. autoMate follows an observe-decide-act workflow, where it analyzes the screen, plans...
    Downloads: 5 This Week
    Last Update:
    See Project
  • Our Free Plans just got better! | Auth0 Icon
    Our Free Plans just got better! | Auth0

    With up to 25k MAUs and unlimited Okta connections, our Free Plan lets you focus on what you do best—building great apps.

    You asked, we delivered! Auth0 is excited to expand our Free and Paid plans to include more options so you can focus on building, deploying, and scaling applications without having to worry about your security. Auth0 now, thank yourself later.
    Try free now
  • 5
    API-for-Open-LLM

    API-for-Open-LLM

    Openai style api for open large language models

    API-for-Open-LLM is a lightweight API server designed for deploying and serving open large language models (LLMs), offering a simple way to integrate LLMs into applications.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    LazyLLM

    LazyLLM

    Easiest and laziest way for building multi-agent LLMs applications

    LazyLLM is an optimized, lightweight LLM server designed for easy and fast deployment of large language models. It is fully compatible with the OpenAI API specification, enabling developers to integrate their own models into applications that normally rely on OpenAI’s endpoints. LazyLLM emphasizes low resource usage and fast inference while supporting multiple models.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 7
    ChainForge

    ChainForge

    An open-source visual programming environment

    ChainForge is an open-source visual programming environment designed to help developers systematically test, compare, and evaluate prompts and outputs across multiple large language models in a structured and scalable way. Instead of relying on isolated prompt experimentation, it introduces a dataflow-based interface that allows users to create complex prompt pipelines and evaluate them across different models, parameters, and datasets simultaneously.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 8
    Ollamac

    Ollamac

    Mac app for Ollama

    ...The application focuses on delivering a lightweight and responsive experience that integrates seamlessly with the macOS ecosystem. Because the models run locally, the system enables private AI workflows without sending data to external APIs or cloud services. Ollamac supports different Ollama models and provides features designed to improve usability such as syntax highlighting and configurable settings.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 9
    gemma.cpp

    gemma.cpp

    lightweight, standalone C++ inference engine for Google's Gemma models

    Gemma.cpp is a C++ implementation for running inference with Gemma models efficiently on CPUs and GPUs. Developed by Google, it allows running large language models (LLMs) like Gemma with minimal hardware, focusing on optimized performance and low latency. Gemma.cpp is intended for developers seeking to deploy LLMs in production environments without needing massive computational resources.
    Downloads: 4 This Week
    Last Update:
    See Project
  • Fully Managed MySQL, PostgreSQL, and SQL Server Icon
    Fully Managed MySQL, PostgreSQL, and SQL Server

    Automatic backups, patching, replication, and failover. Focus on your app, not your database.

    Cloud SQL handles your database ops end to end, so you can focus on your app.
    Try Free
  • 10
    DSPy

    DSPy

    DSPy: The framework for programming—not prompting—language models

    Developed by the Stanford NLP Group, DSPy (Declarative Self-improving Python) is a framework that enables developers to program language models through compositional Python code rather than relying solely on prompt engineering. It facilitates the construction of modular AI systems and provides algorithms for optimizing prompts and weights, enhancing the quality and reliability of language model outputs.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 11
    Fragments by E2B

    Fragments by E2B

    Open source template for AI-powered code generation apps w/ sandboxes

    Fragments is an open source template designed for building applications where artificial intelligence generates and executes code directly from natural language prompts. It provides a ready-to-use web interface where users can chat with AI models to create code snippets, applications, or scripts in various programming environments. Generated code can be executed safely in isolated sandbox environments using the E2B SDK, which helps protect the host system while running AI-produced programs. ...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 12
    GPU Hot

    GPU Hot

    Real-time NVIDIA GPU dashboard

    GPU Hot is an open-source, lightweight monitoring dashboard designed to provide real-time visibility into NVIDIA GPU performance across single machines or entire clusters. The project offers a self-hosted web interface that streams hardware metrics directly from GPU servers, enabling developers, ML engineers, and system administrators to observe GPU utilization and system behavior in real time through a browser. The dashboard collects and displays a wide range of performance metrics...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 13
    Riff

    Riff

    Nine opinionated skills for writing personal essays with Claude

    ...It also supports extensibility, allowing developers to adapt the tool to different programming environments. Overall, Riff acts as an intelligent coding assistant that improves productivity and developer experience.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 14
    llama.vim

    llama.vim

    Vim plugin for LLM-assisted code/text completion

    llama.vim is a lightweight Vim plugin that integrates large language model capabilities directly into the Vim text editor. The plugin enables developers to access AI-assisted text and code completion features without leaving their terminal-based development environment. Instead of relying on remote AI services, the plugin is designed to work with locally running LLM inference engines such as llama.cpp.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Reader 3

    Reader 3

    Quick illustration of how one can easily read books together with LLMs

    This project is a minimalist, self-hosted EPUB reader designed to help users browse and read EPUB books one chapter at a time through a lightweight local server, making it especially easy to extract or work with chapters in external tools like large language models. It was created primarily as a simple demonstration of how to combine local book reading with LLM workflows without heavy dependencies or complicated setup, and it runs with just a small Python script and a basic HTTP server. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    MiniMind-V

    MiniMind-V

    "Big Model" trains a visual multimodal VLM with 26M parameters

    ...MiniMind-V combines techniques from modern vision-language modeling but focuses on efficiency and simplicity so that individuals or small teams can explore multimodal learning without massive GPU clusters. It includes training scripts, model definitions, and associated tooling that illustrate how to build and evaluate such lightweight models. While not intended to compete with large production models, it serves as a hands-on educational resource and starting point for experimentation.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Simple Evals

    Simple Evals

    Lightweight framework for evaluating large language model performance

    simple-evals is a lightweight evaluation framework developed by OpenAI for quickly testing models against small, focused benchmarks. It is designed to help researchers and developers run targeted evaluations without the complexity of large-scale pipelines. By emphasizing simplicity, the framework makes it easy to define new tasks, run evaluations, and interpret results in a reproducible way. It is particularly useful for sanity checks, exploratory research, and comparing performance across...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 18
    MathModelAgent

    MathModelAgent

    An Agent Designed for Mathematical Modeling

    ...The platform automates the process of analyzing mathematical problems, constructing models, generating code for simulations or computations, and producing a complete research-style report. The project uses a multi-agent architecture where different specialized agents handle tasks such as problem interpretation, modeling design, programming implementation, and paper writing. Through integration with multiple large language models, the system can coordinate these components to generate structured modeling solutions and formatted research papers suitable for submission. The platform also includes a code execution environment that allows generated programs to be tested, corrected, and refined during the modeling workflow.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 19
    Easy-Vibe

    Easy-Vibe

    Tutorial on Product Prototype, AI Capability Integration

    Easy-Vibe is an open-source educational project designed to teach developers, product managers, and beginners how to build AI-powered applications using the emerging concept of “vibe coding,” a development approach that relies heavily on AI-assisted programming tools. The project provides a structured curriculum that guides learners from having no programming experience to building fully functional AI-integrated applications. Instead of focusing only on theoretical concepts, Easy-Vibe...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 20
    CodeGeeX

    CodeGeeX

    CodeGeeX: An Open Multilingual Code Generation Model (KDD 2023)

    CodeGeeX is a large-scale multilingual code generation model with 13 billion parameters, trained on 850B tokens across more than 20 programming languages. Developed with MindSpore and later made PyTorch-compatible, it is capable of multilingual code generation, cross-lingual code translation, code completion, summarization, and explanation. It has been benchmarked on HumanEval-X, a multilingual program synthesis benchmark introduced alongside the model, and achieves state-of-the-art...
    Downloads: 12 This Week
    Last Update:
    See Project
  • 21
    Nimbalyst

    Nimbalyst

    Run multiple Codex and Claude Code AI sessions

    Crystal is an open-source project focused on building a lightweight and flexible system for managing structured data, workflows, or automation pipelines, typically oriented toward developer productivity and extensible backend tooling. It is designed with modularity in mind, allowing developers to define reusable components and compose them into larger workflows that can adapt to different use cases.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 22
    LangChain Rust

    LangChain Rust

    LangChain for Rust, the easiest way to write LLM-based programs

    LangChain Rust is an open-source Rust implementation inspired by the LangChain ecosystem for building applications powered by large language models. The library aims to provide Rust developers with a structured framework for orchestrating prompts, chains, agents, and external tools within LLM-driven workflows. By adapting LangChain concepts to the Rust programming language, the project emphasizes performance, safety, and efficient memory management. Developers can use the framework to build chatbots, autonomous agents, and knowledge-augmented AI systems that interact with external data sources. ...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 23
    csghub-server

    csghub-server

    csghub-server is the backend server for CSGHub

    ...Built primarily in the Go programming language, the system enables organizations to run model inference, training, and fine-tuning tasks within a unified platform. It integrates capabilities similar to model repositories like Hugging Face while allowing enterprises to host and manage their AI assets internally for security and compliance purposes.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 24
    llama.vscode

    llama.vscode

    VS Code extension for LLM-assisted code/text completion

    llama.vscode is a Visual Studio Code extension that provides AI-assisted coding features powered primarily by locally running language models. The extension is designed to be lightweight and efficient, enabling developers to use AI tools even on consumer-grade hardware. It integrates with the llama.cpp runtime to run language models locally, eliminating the need to rely entirely on external APIs or cloud providers. The extension supports common AI development features such as code completion, conversational chat assistance, and AI-assisted code editing directly within the IDE. ...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 25
    Floneum

    Floneum

    Instant, controllable, local pre-trained AI models in Rust

    Floneum is an open-source platform for building AI-powered workflows using large language models through a visual and extensible interface. The system allows users to design complex AI pipelines using a drag-and-drop workflow builder rather than writing extensive code. It focuses on enabling developers and researchers to create language model applications that combine different tools, data sources, and AI capabilities into automated workflows.
    Downloads: 2 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB