Showing 275 open source projects for "instruction programming language"

View related business solutions
  • AI-powered service management for IT and enterprise teams Icon
    AI-powered service management for IT and enterprise teams

    Enterprise-grade ITSM, for every business

    Give your IT, operations, and business teams the ability to deliver exceptional services—without the complexity. Maximize operational efficiency with refreshingly simple, AI-powered Freshservice.
    Try it Free
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build, govern, and optimize agents and models with Gemini Enterprise Agent Platform.
    Start Free
  • 1
    Anthony Fu's Skills

    Anthony Fu's Skills

    Anthony Fu's curated collection of agent skills

    Anthony Fu's Skills is an open-source collection of agent skills — modular instruction packages that teach AI coding assistants how to perform specific tasks automatically when relevant. These skills are typically simple, human-readable files that contain structured steps, rules, examples, and workflow logic, letting tools like Claude Code or Copilot CLI load and run them only when they apply to the user’s input.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    OpenSandbox

    OpenSandbox

    OpenSandbox is a general-purpose sandbox platform for AI applications

    OpenSandbox is a general purpose sandbox platform designed to securely run and isolate AI applications and untrusted workloads in controlled environments. The project focuses on providing a unified sandbox API that simplifies the process of executing code safely across different runtime backends. It supports multiple programming languages through SDKs, allowing developers to integrate sandbox capabilities into their systems without building custom isolation layers. The platform is built to...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 3
    machine_learning_examples

    machine_learning_examples

    A collection of machine learning examples and tutorials

    machine_learning_examples is an open-source repository that provides a large collection of machine learning tutorials and practical code examples. The project aims to teach machine learning concepts through hands-on programming rather than purely theoretical explanations. It includes implementations of many machine learning algorithms and neural network architectures using Python and popular libraries such as TensorFlow and NumPy. The repository covers a wide range of topics including supervised learning, unsupervised learning, reinforcement learning, and natural language processing. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    MCP UI

    MCP UI

    SDK for building interactive UI components over MCP for AI tools

    mcp-ui is a software development kit designed to bring interactive user interface capabilities to applications built on the Model Context Protocol (MCP). It enables developers to create rich, dynamic UI components that can be delivered from an MCP server and rendered seamlessly by a compatible client. Instead of returning only text responses, tools can provide structured UI resources such as HTML or remote-rendered components, allowing more engaging and functional interactions. mcp-ui...
    Downloads: 2 This Week
    Last Update:
    See Project
  • Build Securely on AWS with Proven Frameworks Icon
    Build Securely on AWS with Proven Frameworks

    Lay a foundation for success with Tested Reference Architectures developed by Fortinet’s experts. Learn more in this white paper.

    Moving to the cloud brings new challenges. How can you manage a larger attack surface while ensuring great network performance? Turn to Fortinet’s Tested Reference Architectures, blueprints for designing and securing cloud environments built by cybersecurity experts. Learn more and explore use cases in this white paper.
    Download Now
  • 5
    Gen.jl

    Gen.jl

    A general-purpose probabilistic programming system

    An open-source stack for generative modeling and probabilistic inference. Gen’s inference library gives users building blocks for writing efficient probabilistic inference algorithms that are tailored to their models, while automating the tricky math and the low-level implementation details. Gen helps users write hybrid algorithms that combine neural networks, variational inference, sequential Monte Carlo samplers, and Markov chain Monte Carlo. Gen features an easy-to-use modeling language...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 6
    Embabel Agent Framework

    Embabel Agent Framework

    Agent framework for the JVM. Pronounced Em-BAY-bel

    Embabel Agent Framework is a JVM-based framework for building advanced AI agent systems that combine structured programming with large language model interactions to execute complex, goal-oriented workflows. The framework introduces a planning-driven approach where agents dynamically determine sequences of actions required to achieve objectives, rather than relying on fixed pipelines or predefined flows. It models agent behavior through concepts such as goals, actions, and conditions, allowing systems to adapt in real time based on changing inputs and outcomes. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    ZML

    ZML

    Any model. Any hardware. Zero compromise

    ...One of its key strengths is cross-compilation, enabling developers to build once and deploy across various platforms without rewriting code. zml provides example implementations of models and workflows, demonstrating how to run inference tasks such as image classification or large language models. It is designed to handle complex distributed setups, including scenarios where model components are split across devices connected via networks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    Qwen-VL

    Qwen-VL

    Chat & pretrained large vision language model

    Qwen-VL is Alibaba Cloud’s vision-language large model family, designed to integrate visual and linguistic modalities. It accepts image inputs (with optional bounding boxes) and text, and produces text (and sometimes bounding boxes) as output. The model variants (VL-Plus, VL-Max, etc.) have been upgraded for better visual reasoning, text recognition from images, fine-grained understanding, and support for high image resolutions / extreme aspect ratios. Qwen-VL supports multilingual inputs...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 9
    Supertonic

    Supertonic

    Lightning-fast, on-device TTS, running natively via ONNX

    Supertonic is a lightning-fast, on-device text-to-speech system built around ONNX Runtime for maximum speed and portability. It focuses on running entirely locally, eliminating the need for cloud APIs and providing low latency and strong privacy guarantees, even on constrained devices like Raspberry Pi boards and e-readers. The core model is highly compact at around 66 million parameters, yet benchmarks show it can generate speech up to 167× faster than real time on modern consumer hardware...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Secure File Transfer for Windows with Cerberus by Redwood Icon
    Secure File Transfer for Windows with Cerberus by Redwood

    Protect and share files over FTP/S, SFTP, HTTPS and SCP with the #1 rated Windows file transfer server.

    Cerberus supports unlimited users and connections on a single IP, with built-in encryption, 2FA, and a browser-based web client — all deployable in under 15 minutes with a 25-day free trial.
    Try for Free
  • 10
    MiniMax-M2

    MiniMax-M2

    MiniMax-M2, a model built for Max coding & agentic workflows

    MiniMax-M2 is an open-weight large language model designed specifically for high-end coding and agentic workflows while staying compact and efficient. It uses a Mixture-of-Experts (MoE) architecture with 230 billion total parameters but only 10 billion activated per token, giving it the behavior of a very large model at a fraction of the runtime cost. The model is tuned for end-to-end developer flows such as multi-file edits, compile–run–fix loops, and test-validated repairs across real repositories and diverse programming languages. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Qwen-Audio

    Qwen-Audio

    Chat & pretrained large audio language model proposed by Alibaba Cloud

    Qwen-Audio is a large audio-language model developed by Alibaba Cloud, built to accept various types of audio input (speech, natural sounds, music, singing) along with text input, and output text. There is also an instruction-tuned version called Qwen-Audio-Chat which supports conversational interaction (multi-round), audio + text input, creative tasks and reasoning over audio.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    PyCaret

    PyCaret

    An open-source, low-code machine learning library in Python

    PyCaret is an open-source, low-code machine learning library in Python that automates machine learning workflows. It is an end-to-end machine learning and model management tool that speeds up the experiment cycle exponentially and makes you more productive. In comparison with the other open-source machine learning libraries, PyCaret is an alternate low-code library that can be used to replace hundreds of lines of code with few lines only. This makes experiments exponentially fast and...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Qwen2.5

    Qwen2.5

    Open source large language model by Alibaba

    Qwen2.5 is a series of large language models developed by the Qwen team at Alibaba Cloud, designed to enhance natural language understanding and generation across multiple languages. The models are available in various sizes, including 0.5B, 1.5B, 3B, 7B, 14B, 32B, and 72B parameters, catering to diverse computational requirements. Trained on a comprehensive dataset of up to 18 trillion tokens, Qwen2.5 models exhibit significant improvements in instruction following, long-text generation (exceeding 8,000 tokens), and structured data comprehension, such as tables and JSON formats. ...
    Downloads: 33 This Week
    Last Update:
    See Project
  • 14
    LLaVA

    LLaVA

    Visual Instruction Tuning: Large Language-and-Vision Assistant

    Visual instruction tuning towards large language and vision models with GPT-4 level capabilities.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 15
    MGIE

    MGIE

    Guiding Instruction-based Image Editing via Multimodal Large Language

    MGIE—Guiding Instruction-based Image Editing—demonstrates how a multimodal LLM can parse natural-language editing instructions and then drive image transformations accordingly. The project focuses on making edits explainable and controllable: the model interprets text guidance, reasons over image content, and outputs edits aligned with user intent. It’s positioned as an ICLR 2024 Spotlight work, with code and references that show how to connect language planning to concrete image operations. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Adrenaline

    Adrenaline

    AI tool that answers and explains programming questions interactively

    Adrenaline is an AI-powered developer tool designed to answer technical questions and help users better understand code and programming concepts. It focuses on enabling natural language interaction with codebases, documentation, and general software development topics, making it easier for developers to explore and reason about technical material. Adrenaline is presented as an issues-only repository, meaning it primarily serves as a feedback and tracking hub rather than containing the full application code. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 17
    ECLiPSe CLP

    ECLiPSe CLP

    ECLiPSe Constraint Logic Programming System

    The ECLiPSe Constraint Logic Programming System is designed for solving combinatorial optimization problems, for the development of new constraint solver technology and their hybrids, and for the teaching of modelling, solving and search techniques.
    Leader badge
    Downloads: 14 This Week
    Last Update:
    See Project
  • 18
    Qwen2.5-Coder

    Qwen2.5-Coder

    Qwen2.5-Coder is the code version of Qwen2.5, the large language model

    Qwen2.5-Coder, developed by QwenLM, is an advanced open-source code generation model designed for developers seeking powerful and diverse coding capabilities. It includes multiple model sizes—ranging from 0.5B to 32B parameters—providing solutions for a wide array of coding needs. The model supports over 92 programming languages and offers exceptional performance in generating code, debugging, and mathematical problem-solving. Qwen2.5-Coder, with its long context length of 128K tokens, is...
    Downloads: 19 This Week
    Last Update:
    See Project
  • 19
    DMNL

    DMNL

    dmnl is another attempt of A. i.

    dmnl is another attempt of A. i. written with Ada. Ada is an portable programming language. With Ada, the text after tow minus sign ('-') are a comment.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 20
    ToRA

    ToRA

    Tool-integrated Reasoning LLM Agents

    ToRA is an open-source framework developed by Microsoft for building tool-integrated reasoning agents powered by large language models. The project focuses on improving the ability of AI systems to solve complex mathematical and analytical problems by combining natural language reasoning with external computational tools. Instead of relying solely on text generation, the system dynamically invokes tools such as symbolic solvers or programming libraries when deeper computation is required. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    Firefly LLM

    Firefly LLM

    A large model training tool that supports training large models

    Firefly is an open-source framework designed to simplify the training and fine-tuning of large language models through a unified and configurable workflow. The project provides a comprehensive environment where developers can perform tasks such as model pre-training, instruction tuning, and preference optimization using widely adopted machine learning techniques. Its architecture supports both full-parameter training and parameter-efficient strategies like LoRA and QLoRA, making it suitable for environments with limited computational resources. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    GNU Prolog

    GNU Prolog

    The GNU Prolog native compiler

    GNU Prolog is a free implementation (under GPL) of the logic programming language PROLOG. It can compile to native machine code which is extremely fast in execution. Another feature is the included constraint solver.
    Leader badge
    Downloads: 95 This Week
    Last Update:
    See Project
  • 23
    DeepSeek LLM

    DeepSeek LLM

    DeepSeek LLM: Let there be answers

    The DeepSeek-LLM repository hosts the code, model files, evaluations, and documentation for DeepSeek’s LLM series (notably the 67B Chat variant). Its tagline is “Let there be answers.” The repo includes an “evaluation” folder (with results like math benchmark scores) and code artifacts (e.g. pre-commit config) that support model development and deployment. According to the evaluation files, DeepSeek LLM 67B Chat achieves strong performance on math benchmarks under both chain-of-thought (CoT)...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 24
    GLM-4-32B-0414

    GLM-4-32B-0414

    Open Multilingual Multimodal Chat LMs

    GLM-4-32B-0414 is a powerful open-source large language model featuring 32 billion parameters, designed to deliver performance comparable to leading models like OpenAI’s GPT series. It supports multilingual and multimodal chat capabilities with an extensive 32K token context length, making it ideal for dialogue, reasoning, and complex task completion. The model is pre-trained on 15 trillion tokens of high-quality data, including substantial synthetic reasoning datasets, and further enhanced with reinforcement learning and human preference alignment for improved instruction-following and function calling. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 25
    Chinese-LLaMA-Alpaca 2

    Chinese-LLaMA-Alpaca 2

    Chinese LLaMA-2 & Alpaca-2 Large Model Phase II Project

    This project is developed based on the commercially available large model Llama-2 released by Meta. It is the second phase of the Chinese LLaMA&Alpaca large model project. The Chinese LLaMA-2 base model and the Alpaca-2 instruction fine-tuning large model are open-sourced. These models expand and optimize the Chinese vocabulary on the basis of the original Llama-2, use large-scale Chinese data for incremental pre-training, and further improve the basic semantics and command understanding of...
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB