173 projects for "instruction programming language" with 2 filters applied:

  • Stop Storing Third-Party Tokens in Your Database Icon
    Stop Storing Third-Party Tokens in Your Database

    Auth0 Token Vault handles secure token storage, exchange, and refresh for external providers so you don't have to build it yourself.

    Rolling your own OAuth token storage can be a security liability. Token Vault securely stores access and refresh tokens from federated providers and handles exchange and renewal automatically. Connected accounts, refresh exchange, and privileged worker flows included.
    Try Auth0 for Free
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    Access competitive interest rates on your digital assets.

    Generate interest, borrow against your crypto, and trade a range of cryptocurrencies — all in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 1
    Magicoder

    Magicoder

    Empowering Code Generation with OSS-Instruct

    Magicoder is an open-source family of large language models designed specifically for code generation and software development tasks. The project focuses on improving the quality and diversity of code generation by training models with a novel dataset construction approach known as OSS-Instruct. This technique uses open-source code repositories as a foundation for generating more realistic and diverse instruction datasets for training language models.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    CodeLlama

    CodeLlama

    Inference code for CodeLlama models

    Code Llama is a family of Llama-based code models optimized for programming tasks such as code generation, completion, and repair, with variants specialized for base coding, Python, and instruction following. The repo documents the sizes and capabilities (e.g., 7B, 13B, 34B) and highlights features like infilling and large input context to support real IDE workflows. It targets both general software synthesis and language-specific productivity, offering strong performance among open models at release time. ...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 3
    Huatuo-Llama-Med-Chinese

    Huatuo-Llama-Med-Chinese

    Instruction-tuning LLM with Chinese Medical Knowledge

    ...The goal of the project is to improve the reliability and domain expertise of language models when answering medical questions or assisting with healthcare-related tasks. By combining domain-specific training data with instruction-tuning techniques, the project produces models capable of generating more accurate medical responses than general-purpose models.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    BAML

    BAML

    The AI framework that adds the engineering to prompt engineering

    ...The framework enables developers to define prompt logic in a dedicated language while integrating it into applications written in various programming languages such as Python, TypeScript, Ruby, and Go. BAML also allows developers to specify which models are used for each prompt and how outputs should be validated or structured. By converting prompt engineering into a more formal programming workflow, the framework improves reliability, debugging, and maintainability of AI systems.
    Downloads: 17 This Week
    Last Update:
    See Project
  • Application Monitoring That Won't Slow Your App Down Icon
    Application Monitoring That Won't Slow Your App Down

    AppSignal's Rust-based agent is lightweight and stable. Already running in thousands of production apps.

    Full APM with errors, performance, logs, and uptime monitoring. 99.999% uptime SLA on the platform itself.
    Start Free
  • 5
    Chinese-LLaMA-Alpaca-3

    Chinese-LLaMA-Alpaca-3

    Chinese Llama-3 LLMs) developed from Meta Llama 3

    Chinese-LLaMA-Alpaca-3 is an open-source project that provides Mandarin-focused large language models based on Meta’s LLaMA-3 architecture, with both foundational and instruction-tuned variants to support high-quality Chinese natural language understanding and generation. It extends the original LLaMA models with expanded Chinese vocabularies and additional pretraining on Chinese corpora to improve semantic encoding and decoding specifically for Chinese text. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    CodeGen

    CodeGen

    Open-source model for program synthesis

    CodeGen is a family of open-source large language models designed specifically for program synthesis and code generation tasks. Developed by Salesforce Research, the models are trained on large datasets containing both natural language and programming language content. This allows them to translate natural language descriptions into functional code across a variety of programming languages.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Prometheus-Eval

    Prometheus-Eval

    Evaluate your LLM's response with Prometheus and GPT4

    Prometheus-Eval is an open-source framework designed to evaluate the outputs of large language models using specialized evaluator models known as Prometheus. The project provides tools, datasets, and scripts that allow developers and researchers to measure the quality of LLM responses through automated scoring rather than relying solely on human evaluators. It implements an “LLM-as-a-judge” approach in which a dedicated language model analyzes instruction–response pairs and assigns scores or rankings based on predefined evaluation criteria. ...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 8
    tt-metal

    tt-metal

    TT-NN operator library, and TT-Metalium low level kernel programming

    tt-metal, also referred to in its documentation as TT-Metalium, is Tenstorrent’s low-level software development kit for programming applications on Tenstorrent AI accelerators. The project is designed for developers who need direct access to the company’s Tensix processor architecture, exposing a programming model that is closer to hardware control than high-level inference frameworks. Instead of following a traditional GPU model centered on massive thread parallelism, the platform is built...
    Downloads: 25 This Week
    Last Update:
    See Project
  • 9
    CLIPS Rule Based Programming Language
    CLIPS is a forward-chaining rule-based programming language written in C that also provides procedural and object-oriented programming facilities.
    Leader badge
    Downloads: 993 This Week
    Last Update:
    See Project
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • 10
    NExT-GPT

    NExT-GPT

    Code and models for ICML 2024 paper, NExT-GPT

    NExT-GPT is an open-source research framework that implements an advanced multimodal large language model capable of understanding and generating content across multiple modalities. Unlike traditional models that primarily handle text, NExT-GPT supports input and output combinations involving text, images, video, and audio in a unified architecture. The system connects a large language model with multimodal encoders and diffusion-based decoders so it can interpret information from different sensory formats and generate responses in different media types. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    MING

    MING

    A large-scale model of medical consultation in Chinese

    MING is an open-source medical large language model designed for intelligent medical consultation and question answering in Chinese. The project focuses on building a healthcare-focused conversational system capable of responding to medical questions, analyzing case descriptions, and guiding diagnostic reasoning. It is trained using medical instruction tuning so that the model can understand patient symptoms and respond with structured explanations and clinical suggestions. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    LLM-Finetuning

    LLM-Finetuning

    LLM Finetuning with peft

    ...The repository includes step-by-step notebooks demonstrating how to fine-tune models such as LLaMA, Falcon, OPT, Vicuna, and GPT-NeoX. These tutorials show how developers can adapt pretrained models for tasks such as chatbots, classification, and instruction following. The project also illustrates how low-precision training techniques and adapter-based methods reduce memory requirements while maintaining strong model performance.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    LLM Datasets

    LLM Datasets

    Curated list of datasets and tools for post-training

    LLM Datasets curates and standardizes datasets commonly used to train and fine-tune large language models, reducing the overhead of hunting down sources and normalizing formats. The repository aims to make datasets easy to inspect and transform, with scripts for downloading, deduping, cleaning, and converting to formats like JSONL that slot into training pipelines. It highlights instruction-tuning and conversation-style corpora while also pointing to code, math, or domain-specific sets for targeted capabilities. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 14
    LLM Course

    LLM Course

    Course to get into Large Language Models (LLMs)

    LLM Course is a hands-on, notebook-driven path for learning how large language models work in practice, from data curation to training, fine-tuning, evaluating, and deploying. It emphasizes reproducible experiments: each step is demonstrated with runnable code, clear dependencies, and references to commonly used open-source models and libraries. Learners get exposure to multiple adaptation strategies—LoRA/QLoRA, instruction fine-tuning, and alignment techniques—so they can choose approaches that fit their hardware and budgets. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Vibe Vibe

    Vibe Vibe

    The First Systematic Vibe Coding Tutorial

    Vibe Vibe is an open-source educational platform and tutorial system designed to teach AI-assisted programming, also known as “vibe coding,” through a structured and beginner-friendly learning path. The project is aimed at users with little to no programming experience, guiding them from initial ideas to fully deployed applications using natural language interactions with AI tools. It introduces a paradigm shift where coding becomes a conversational process, allowing users to build software by describing their intent rather than writing every line of code manually. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    EmoLLM

    EmoLLM

    Pre & Post-training & Dataset & Evaluation & Depoly & RAG

    EmoLLM is an open-source family of large language models focused on mental health support and counseling-oriented interactions. The project is designed to help users through mental health conversations and has been fine-tuned from existing instruction-following LLMs rather than built as a base model from scratch. Its repository includes multiple model variants and training configurations spanning several underlying model families, including InternLM, Qwen, DeepSeek, Mixtral, LLaMA, and others, which shows that the initiative is structured as a broad ecosystem rather than a single release. ...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 17
    OpenAI Harmony

    OpenAI Harmony

    Renderer for the harmony response format to be used with gpt-oss

    Harmony is a response format developed by OpenAI for use with the gpt-oss model series. It defines a structured way for language models to produce outputs, including regular text, reasoning traces, tool calls, and structured data. By mimicking the OpenAI Responses API, Harmony provides developers with a familiar interface while enabling more advanced capabilities such as multiple output channels, instruction hierarchies, and tool namespaces. The format is essential for ensuring gpt-oss models operate correctly, as they are trained to rely on this structure for generating and organizing their responses. ...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 18
    AICGSecEval

    AICGSecEval

    A.S.E (AICGSecEval) is a repository-level AI-generated code security

    AICGSecEval is an open-source benchmark framework designed to evaluate the security of code generated by artificial intelligence systems. The project was developed to address concerns that AI-assisted programming tools may produce insecure code containing vulnerabilities such as injection flaws or unsafe logic. The framework constructs evaluation tasks based on real-world software repositories and known vulnerability cases derived from CVE records. By simulating realistic development...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    autoMate

    autoMate

    AI tool for automating desktop tasks via natural language input

    autoMate is an AI-powered local automation tool designed to enable users to control and automate their computers using natural language instructions instead of traditional scripting or rule-based systems. It combines large language models with computer vision techniques to interpret user intent and understand on-screen content, allowing it to interact with graphical interfaces similarly to a human user. autoMate follows an observe-decide-act workflow, where it analyzes the screen, plans...
    Downloads: 7 This Week
    Last Update:
    See Project
  • 20
    Fragments by E2B

    Fragments by E2B

    Open source template for AI-powered code generation apps w/ sandboxes

    Fragments is an open source template designed for building applications where artificial intelligence generates and executes code directly from natural language prompts. It provides a ready-to-use web interface where users can chat with AI models to create code snippets, applications, or scripts in various programming environments. Generated code can be executed safely in isolated sandbox environments using the E2B SDK, which helps protect the host system while running AI-produced programs. ...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 21
    code-act

    code-act

    Official Repo for ICML 2024 paper

    code-act is a research framework for building intelligent language-model agents that interact with their environment through executable code actions. The system proposes a unified action representation where language models produce Python code that can be executed directly, allowing the model to interact with external tools and environments in a structured way. By integrating a Python interpreter with the agent architecture, the system enables the agent to execute code, observe the results,...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    Riff

    Riff

    Nine opinionated skills for writing personal essays with Claude

    ...It also supports extensibility, allowing developers to adapt the tool to different programming environments. Overall, Riff acts as an intelligent coding assistant that improves productivity and developer experience.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 23
    Easy-Vibe

    Easy-Vibe

    Tutorial on Product Prototype, AI Capability Integration

    Easy-Vibe is an open-source educational project designed to teach developers, product managers, and beginners how to build AI-powered applications using the emerging concept of “vibe coding,” a development approach that relies heavily on AI-assisted programming tools. The project provides a structured curriculum that guides learners from having no programming experience to building fully functional AI-integrated applications. Instead of focusing only on theoretical concepts, Easy-Vibe...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 24
    MathModelAgent

    MathModelAgent

    An Agent Designed for Mathematical Modeling

    ...The platform automates the process of analyzing mathematical problems, constructing models, generating code for simulations or computations, and producing a complete research-style report. The project uses a multi-agent architecture where different specialized agents handle tasks such as problem interpretation, modeling design, programming implementation, and paper writing. Through integration with multiple large language models, the system can coordinate these components to generate structured modeling solutions and formatted research papers suitable for submission. The platform also includes a code execution environment that allows generated programs to be tested, corrected, and refined during the modeling workflow.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 25
    Xtuner

    Xtuner

    A Next-Generation Training Engine Built for Ultra-Large MoE Models

    ...Its architecture incorporates memory-efficient optimizations that allow researchers to train large models even when computational resources are limited. XTuner is also designed to integrate with modern AI ecosystems, supporting multimodal training, reinforcement learning optimization, and instruction tuning pipelines.
    Downloads: 1 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • 4
  • 5
  • Next
MongoDB Logo MongoDB