Showing 169 open source projects for "qt-based"

View related business solutions
  • Add Two Lines of Code. Get Full APM. Icon
    Add Two Lines of Code. Get Full APM.

    AppSignal installs in minutes and auto-configures dashboards, alerts, and error tracking.

    Works out of the box for Rails, Django, Express, Phoenix, and more. Monitoring exceptions and performance in no time.
    Start Free
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 1
    RecAI

    RecAI

    Bridging LLM and Recommender System

    RecAI is an open-source research platform developed by Microsoft to explore how large language models can be integrated into modern recommender systems. Traditional recommender systems rely on structured behavioral data such as user interactions and item embeddings, while large language models excel at understanding language and reasoning about user preferences. RecAI aims to bridge these two domains by creating architectures and training methods that allow LLMs to function as intelligent...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    MetaScreener

    MetaScreener

    AI-powered tool for efficient abstract and PDF screening

    ...Instead of manually reviewing hundreds or thousands of documents, researchers can use MetaScreener to apply machine learning techniques that assist with classification and prioritization of candidate papers. The platform can analyze both abstracts and full PDF documents, enabling automated filtering based on research criteria defined by the user. By incorporating natural language processing techniques, the system can identify potentially relevant studies and reduce the workload associated with manual screening.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    nano-graphrag

    nano-graphrag

    A simple, easy-to-hack GraphRAG implementation

    nano-graphrag is a lightweight implementation of the GraphRAG approach designed to simplify experimentation with graph-based retrieval-augmented generation systems. GraphRAG expands traditional RAG pipelines by constructing knowledge graphs from documents and using relationships between entities to improve the quality and reasoning of AI responses. The nano-GraphRAG project focuses on reducing complexity by providing a compact and readable codebase that preserves the core functionality of graph-based retrieval systems while remaining easy to modify and extend. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    VLMEvalKit

    VLMEvalKit

    Open-source evaluation toolkit of large multi-modality models (LMMs)

    ...Instead of requiring complex data preparation pipelines or multiple repositories for each benchmark, the system enables evaluation through simple commands that automatically handle dataset loading, model inference, and metric computation. VLMEvalKit supports generation-based evaluation methods, allowing models to produce textual responses to visual inputs while measuring performance through techniques such as exact matching or language-model-assisted answer extraction.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • 5
    LangBot

    LangBot

    Production-grade platform for building agentic IM bots

    ...LangBot combines LLM capabilities with agent logic, knowledge base orchestration, and plugin infrastructure so that bots can perform complex tasks rather than simple conversational responses. The platform includes a web-based management interface that simplifies configuration, access control, and integration with external AI services.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    LLM Telegram Bot

    LLM Telegram Bot

    A Telegram bot for Large Language Models

    LLM Telegram Bot is a self-hosted Telegram chatbot that connects messaging interactions with large language models, typically powered by Ollama or similar backends. The project is designed to provide a customizable AI assistant that can operate within Telegram conversations, supporting dynamic responses based on user input and configurable parameters. It includes features such as conversation memory, allowing the bot to maintain context across multiple messages and provide more coherent responses. The system supports multiple modes or personas, enabling users to switch between different conversational styles or use cases. It also allows fine-tuning of generation parameters such as temperature and token limits, giving users control over response behavior. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 7
    MING

    MING

    A large-scale model of medical consultation in Chinese

    ...This interactive capability makes it suitable for conversational health applications, patient triage scenarios, and educational demonstrations. The model is built on transformer-based architectures using frameworks such as PyTorch and integrates with Hugging Face tooling for training and inference workflows.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    LLM-Pruner

    LLM-Pruner

    On the Structural Pruning of Large Language Models

    ...LLM-Pruner addresses this issue by identifying and removing non-essential components within transformer architectures, such as redundant attention heads or feed-forward structures. The framework relies on gradient-based analysis to determine which parameters contribute least to model performance, enabling targeted structural pruning rather than simple weight removal. After pruning, the framework applies lightweight fine-tuning methods such as LoRA to recover performance using relatively small datasets and short training times.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    Pixeltable

    Pixeltable

    Data Infrastructure providing an approach to multimodal AI workloads

    ...Unlike traditional architectures that require multiple tools such as databases, vector stores, and workflow orchestrators, Pixeltable unifies these functions within a table-based abstraction. Developers define data transformations and AI operations using computed columns on tables, allowing pipelines to evolve incrementally as new data or models are added. The framework supports multimodal content including images, video, text, and audio, enabling applications such as retrieval-augmented generation systems, semantic search, and multimedia analytics.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Go From AI Idea to AI App Fast Icon
    Go From AI Idea to AI App Fast

    One platform to build, fine-tune, and deploy ML models. No MLOps team required.

    Access Gemini 3 and 200+ models. Build chatbots, agents, or custom models with built-in monitoring and scaling.
    Try Free
  • 10
    Agentless

    Agentless

    An agentless approach to automatically solve software development

    ...In the final stage, the framework validates potential patches by running regression tests and additional reproduction tests to confirm whether the fix resolves the original error. Based on these results, the system ranks the candidate patches and selects the most reliable solution to submit.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    MaxText

    MaxText

    A simple, performant and scalable Jax LLM

    MaxText is a high-performance, highly scalable open-source framework designed to train and fine-tune large language models using the JAX ecosystem. The project acts as both a reference implementation and a practical training library that demonstrates best practices for building and scaling transformer-based language models on modern accelerator hardware. It is optimized to run efficiently on Google Cloud TPUs and GPUs, enabling researchers and engineers to train models ranging from small experiments to extremely large distributed workloads. The framework focuses on simplicity while still supporting advanced techniques such as model sharding, distributed computation, and high-throughput training pipelines. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Lagent

    Lagent

    A lightweight framework for building LLM-based agents

    Lagent is a lightweight open-source framework designed to help developers build autonomous agents powered by large language models. The framework provides tools and abstractions that allow language models to interact with external tools, execute tasks, and perform multi-step reasoning processes. Instead of using LLMs only for text generation, Lagent enables developers to transform models into agents capable of performing actions such as retrieving data, executing code, or interacting with...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    TigerBot

    TigerBot

    TigerBot: A multi-language multi-task LLM

    ...The project focuses on building high-performance models capable of handling both English and Chinese tasks while maintaining strong reasoning and conversational abilities. TigerBot models are based on modern transformer architectures and are trained on large datasets that cover multiple domains and languages. The project provides both base models and chat-optimized variants that can be used for dialogue systems, question answering, and general language understanding tasks. In addition to model weights, the repository includes training scripts, inference tools, and configuration files that allow researchers and developers to reproduce experiments or fine-tune the models for specific applications.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    OmAgent

    OmAgent

    Build multimodal language agents for fast prototype and production

    ...Instead of forcing developers to implement complex orchestration logic manually, the system manages task scheduling, worker coordination, and node optimization behind the scenes. Its architecture uses a graph-based workflow engine where tasks are represented as nodes in a directed workflow, enabling modular composition of complex reasoning pipelines. The framework also includes support for various reasoning strategies commonly used in language agents, such as chain-of-thought prompting, self-consistency reasoning, and ReAct-style decision loops.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Skywork-R1V4

    Skywork-R1V4

    Skywork-R1V is an advanced multimodal AI model series

    Skywork-R1V is an open-source multimodal reasoning model designed to extend the capabilities of large language models into vision-language tasks that require complex logical reasoning. The project introduces a model architecture that transfers the reasoning abilities of advanced text-based models into visual domains so the system can interpret images and perform multi-step reasoning about them. Instead of retraining both language and vision models from scratch, the framework uses a lightweight visual projection layer that connects a pretrained vision backbone with a reasoning-capable language model. This design allows the model to analyze images while maintaining strong textual reasoning performance, enabling tasks such as solving visual math problems, interpreting scientific diagrams, and answering questions about images.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Torch Pruning

    Torch Pruning

    DepGraph: Towards Any Structural Pruning

    ...The library focuses on reducing the size and computational cost of neural networks by removing redundant parameters and channels while maintaining model performance. It introduces a graph-based algorithm called DepGraph that automatically identifies dependencies between layers, allowing parameters to be pruned safely across complex architectures. This dependency analysis makes it possible to prune large networks such as transformers, convolutional networks, and diffusion models without breaking the computational graph. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    NExT-GPT

    NExT-GPT

    Code and models for ICML 2024 paper, NExT-GPT

    ...Unlike traditional models that primarily handle text, NExT-GPT supports input and output combinations involving text, images, video, and audio in a unified architecture. The system connects a large language model with multimodal encoders and diffusion-based decoders so it can interpret information from different sensory formats and generate responses in different media types. This architecture allows the model to convert between modalities, such as generating images from text descriptions or producing audio or video outputs based on textual prompts. The project also introduces instruction-tuning strategies that enable the model to perform complex multimodal reasoning and generation tasks with minimal additional parameters.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    Sparrow

    Sparrow

    Structured data extraction and instruction calling with ML, LLM

    ...The system focuses on transforming complex documents such as invoices, receipts, forms, and scanned pages into structured formats like JSON that can be processed by downstream applications. It combines several components, including OCR pipelines, vision-language models, and LLM-based reasoning modules to identify and extract meaningful data fields from heterogeneous document layouts. The architecture is modular, allowing developers to build customizable processing pipelines that integrate with external tools and data extraction frameworks. Sparrow also includes workflow orchestration tools that allow multiple extraction tasks to be combined into automated pipelines for large-scale document processing.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    OM1

    OM1

    Modular AI runtime for robots

    OM1 is an open-source AI platform designed to build autonomous agents capable of interacting with digital environments and completing complex tasks. The project focuses on creating a modular architecture where language models can coordinate with external tools, APIs, and knowledge sources to accomplish multi-step objectives. Instead of operating as simple conversational systems, OM1 agents can plan actions, retrieve information, and execute tasks across different services. The framework...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 20
    SageAttention

    SageAttention

    NeurIPS2025 Spotlight] Quantized Attention

    SageAttention is an open-source optimization library designed to accelerate the attention mechanism used in transformer-based neural networks. Since attention operations are often the most computationally expensive component of modern AI models, SageAttention introduces quantization techniques that significantly reduce computational overhead while preserving model accuracy. The system achieves this by using low-precision numerical formats such as INT4, FP8, or INT8 to represent key matrices within the attention computation. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 21
    handy-ollama

    handy-ollama

    Implement CPU from scratch and play with large model deployments

    ...The repository serves as a structured tutorial that explains how to install, configure, and use Ollama to run modern language models on personal hardware without requiring advanced infrastructure. A key focus of the project is enabling users to run large models even without GPUs by leveraging optimized CPU-based inference pipelines. The project includes step-by-step guides that walk learners through tasks such as installing Ollama, managing local models, calling model APIs, and building simple AI applications on top of locally hosted models. Through hands-on exercises and practical examples, the tutorial demonstrates how developers can create applications like chat assistants or retrieval systems using locally deployed models.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    second-brain-ai-assistant-course

    second-brain-ai-assistant-course

    Learn to build your Second Brain AI assistant with LLMs

    ...The concept of a “second brain” refers to a personal knowledge repository containing notes, research, and documents that can be queried and analyzed using AI. Through a series of modules, the project explains how to design data pipelines, build retrieval-augmented generation systems, and implement agent-based reasoning workflows. The course also introduces practical techniques such as dataset generation, model fine-tuning, and deployment strategies for AI applications. Learners build a full system capable of retrieving information from stored resources and generating responses based on that data.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    RLHF-Reward-Modeling

    RLHF-Reward-Modeling

    Recipes to train reward model for RLHF

    RLHF-Reward-Modeling is an open-source research framework focused on training reward models used in reinforcement learning from human feedback for large language models. In RLHF pipelines, reward models are responsible for evaluating generated responses and assigning scores that guide the model toward outputs that better match human preferences. The repository provides training recipes and implementations for building reward and preference models using modern machine learning frameworks. It...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 24
    LLM-Aided OCR Project

    LLM-Aided OCR Project

    Enhances Tesseract OCR output using LLMs (local or API)

    ...The project addresses common OCR challenges such as distorted text, unusual fonts, historical documents, and complex layouts that often produce inaccurate results with standard OCR pipelines. The system first extracts raw text using OCR engines and then applies language models to analyze and correct recognition errors based on context. This AI-assisted correction process helps reconstruct missing characters, fix formatting mistakes, and produce more coherent text outputs. The project is particularly useful for digitizing historical documents, research papers, and scanned materials where traditional OCR often struggles. It also includes tools for processing batches of images or documents, enabling automated document digitization workflows.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 25
    Hallucination Leaderboard

    Hallucination Leaderboard

    Leaderboard Comparing LLM Performance at Producing Hallucinations

    Hallucination Leaderboard is an open research project that tracks and compares the tendency of large language models to produce hallucinated or inaccurate information when generating summaries. The project provides a standardized benchmark that evaluates different models using a dedicated hallucination detection system known as the Hallucination Evaluation Model. Each model is tested on document summarization tasks to measure how often generated responses introduce information that is not...
    Downloads: 1 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB