Search Results for "transformer design optimization"

Showing 243 open source projects for "transformer design optimization"

View related business solutions
  • Full-stack observability with actually useful AI | Grafana Cloud Icon
    Full-stack observability with actually useful AI | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 1
    Transformer Explainer

    Transformer Explainer

    Learn How LLM Transformer Models Work with Interactive Visualization

    ...Users can observe how attention weights change as the model predicts the next token, offering insight into how transformer architectures capture relationships between words. The design of the platform emphasizes educational accessibility, allowing students, researchers, and developers to explore complex machine learning concepts without requiring specialized hardware or installations.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    Transformers.jl

    Transformers.jl

    Julia Implementation of Transformer models

    Transformers.jl is a Julia library that implements Transformer models for natural language processing tasks. Inspired by architectures like BERT, GPT, and T5, the library offers a modular and flexible interface for building, training, and using transformer-based deep learning models. It supports training from scratch and fine-tuning pretrained models, and integrates with Flux.jl for automatic differentiation and optimization.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    CTranslate2

    CTranslate2

    Fast inference engine for Transformer models

    CTranslate2 is a C++ and Python library for efficient inference with Transformer models. The project implements a custom runtime that applies many performance optimization techniques such as weights quantization, layers fusion, batch reordering, etc., to accelerate and reduce the memory usage of Transformer models on CPU and GPU. The execution is significantly faster and requires less resources than general-purpose deep learning frameworks on supported models and tasks thanks to many advanced optimizations: layer fusion, padding removal, batch reordering, in-place operations, caching mechanism, etc. ...
    Downloads: 13 This Week
    Last Update:
    See Project
  • 4
    Intel Extension for Transformers

    Intel Extension for Transformers

    Build your chatbot within minutes on your favorite device

    Intel Extension for Transformers is an innovative toolkit designed to accelerate Transformer-based models on Intel platforms, including CPUs and GPUs. It offers state-of-the-art compression techniques for Large Language Models (LLMs) and provides tools to build chatbots within minutes on various devices. The extension aims to optimize the performance of Transformer-based models, making them more efficient and accessible.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Go From AI Idea to AI App Fast Icon
    Go From AI Idea to AI App Fast

    One platform to build, fine-tune, and deploy ML models. No MLOps team required.

    Access Gemini 3 and 200+ models. Build chatbots, agents, or custom models with built-in monitoring and scaling.
    Try Free
  • 5
    llm.c

    llm.c

    LLM training in simple, raw C/CUDA

    ...Its compact design makes it easy to trace execution, profile hotspots, and understand the cost of each operation. Portability is a goal: it aims to compile with common toolchains and run on modest hardware for small experiments. Rather than delivering a production-grade stack, it serves as a reference and learning scaffold for people who want to “see the metal” behind LLMs.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 6
    Heretic

    Heretic

    Fully automatic censorship removal for language models

    Heretic is an open-source Python tool that automatically removes the built-in censorship or “safety alignment” from transformer-based language models so they respond to a broader range of prompts with fewer refusals. It works by applying directional ablation techniques and a parameter optimization strategy to adjust internal model behaviors without expensive post-training or altering the core capabilities. Designed for researchers and advanced users, Heretic makes it possible to study and experiment with uncensored model responses in a reproducible, automated way. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    BitNet

    BitNet

    BitNet: Scaling 1-bit Transformers for Large Language Models

    BitNet is a machine learning research implementation that explores extremely low-precision neural network architectures designed to dramatically reduce the computational cost of large language models. The project implements the BitNet architecture described in research on scaling transformer models using extremely low-bit quantization techniques. In this approach, neural network weights are quantized to approximately one bit per parameter, allowing models to operate with far lower memory...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 8
    Axolotl

    Axolotl

    Go ahead and axolotl questions

    Axolotl is a powerful and flexible framework for fine-tuning large language models on custom datasets. Built for researchers and developers, Axolotl simplifies the process of adapting LLMs for specific tasks, including chat, code generation, and instruction following. It supports a wide variety of model architectures and offers out-of-the-box optimization strategies for efficient training.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 9
    MiniOneRec

    MiniOneRec

    Minimal reproduction of OneRec

    ...The framework provides an end-to-end pipeline for building generative recommender systems, including semantic identifier construction, supervised fine-tuning, and reinforcement learning-based optimization. Semantic IDs are created using techniques such as quantized variational autoencoders to convert item features into token sequences that can be modeled by transformer architectures. Developers can train and evaluate recommendation models using different backbone language models while benefiting from the generative framework’s parameter efficiency and scalability.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Go from Code to Production URL in Seconds Icon
    Go from Code to Production URL in Seconds

    Cloud Run deploys apps in any language instantly. Scales to zero. Pay only when code runs.

    Skip the Kubernetes configs. Cloud Run handles HTTPS, scaling, and infrastructure automatically. Two million requests free per month.
    Try it free
  • 10
    ggml

    ggml

    Tensor library for machine learning

    ...The project emphasizes portability and performance, enabling machine learning inference across a wide range of hardware environments including CPUs and specialized accelerators. It is widely used as a foundational component in projects that run large language models locally, including tools that perform inference for transformer-based models. The library also implements optimization algorithms and computation graph functionality so developers can build training and inference workflows directly on top of its tensor operations.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    SageAttention

    SageAttention

    NeurIPS2025 Spotlight] Quantized Attention

    SageAttention is an open-source optimization library designed to accelerate the attention mechanism used in transformer-based neural networks. Since attention operations are often the most computationally expensive component of modern AI models, SageAttention introduces quantization techniques that significantly reduce computational overhead while preserving model accuracy.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 12
    FlashAttention

    FlashAttention

    Fast and memory-efficient exact attention

    FlashAttention is a high-performance deep learning optimization library that reimplements the attention mechanism used in transformer models to be significantly faster and more memory-efficient than standard implementations. It achieves this by using IO-aware algorithms that minimize memory reads and writes, reducing the quadratic memory overhead typically associated with attention operations.
    Downloads: 19 This Week
    Last Update:
    See Project
  • 13
    LightLLM

    LightLLM

    LightLLM is a Python-based LLM (Large Language Model) inference

    ...Its architecture allows models to be deployed with minimal overhead while maintaining compatibility with popular transformer-based model families such as LLaMA and GPT-style architectures.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    FLUX.2-klein-4B

    FLUX.2-klein-4B

    Flux 2 image generation model pure C inference

    FLUX.2-klein-4B is a compact, high-performance C library implementation of the Flux optimization algorithm — an iterative approach for solving large-scale optimization problems common in scientific computing, machine learning, and numerical simulation. Written with a strong emphasis on simplicity, correctness, and performance, it abstracts the core logic of flux-based optimization into a minimal C API that can be embedded in broader applications without pulling in heavy dependencies. Because...
    Downloads: 9 This Week
    Last Update:
    See Project
  • 15
    SeedVR

    SeedVR

    Repo for SeedVR2 & SeedVR

    ...SeedVR’s transformer-based design allows it to handle variable frame resolutions and lengths, and its architecture is optimized to overcome traditional limitations of windowed attention in high-resolution contexts.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 16
    Janus

    Janus

    Unified Multimodal Understanding and Generation Models

    Janus is a sophisticated open-source project from DeepSeek AI that aims to unify both visual understanding and image generation in a single model architecture. Rather than having separate systems for “look and describe” and “prompt and generate”, Janus uses an autoregressive transformer framework with a decoupled visual encoder—allowing it to ingest images for comprehension and to produce images from text prompts with shared internal representations. The design tackles long-standing conflicts in multimodal models: namely that the visual encoder has to serve both analysis (understanding) and synthesis (generation) roles. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 17
    Intel LLM Library for PyTorch

    Intel LLM Library for PyTorch

    Accelerate local LLM inference and finetuning

    Intel LLM Library for PyTorch is an open-source acceleration library developed to optimize large language model inference and fine-tuning on Intel hardware platforms. Built as an extension of the PyTorch ecosystem, the library enables developers to run modern transformer models efficiently on Intel CPUs, GPUs, and specialized AI accelerators. The framework provides hardware-aware optimizations and low-precision computation techniques that significantly improve the performance of large...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    ACE-Step 1.5

    ACE-Step 1.5

    The most powerful local music generation model

    ACE-Step 1.5 is an advanced open-source foundation model for AI-driven music generation that pushes beyond traditional limitations in speed, musical coherence, and controllability by innovating in architecture and training design. It integrates cutting-edge generative techniques—such as diffusion-based synthesis combined with compressed autoencoders and lightweight transformer elements—to produce high-quality full-length music tracks with rapid inference times, capable of generating a complete song in seconds on modern GPUs while remaining efficient enough to run on consumer-grade hardware with minimal memory requirements. ...
    Downloads: 85 This Week
    Last Update:
    See Project
  • 19
    SVGO

    SVGO

    Node.js tool for optimizing SVG files

    SVG Optimizer is a Node.js-based tool for optimizing SVG vector graphics files. SVG files, in particular those exported from multiple editors, normally contain tons of redundant and useless information. This can include editor metadata, comments, hidden elements, default or non-optimal values and other stuff that can be safely removed or converted without affecting the SVG rendering result. Some options can be configured with CLI though it may be easier to have the configuration in a...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 20
    VARLET

    VARLET

    A Vue3 component library based on Material Design 2 and 3

    Varlet UI is a Material design component library developed based on Vue3, supporting mobile and desktop, developed and maintained by varletjs organization. Support Typescript, import on demand, dark mode, theme customization, internationalization, and provide VSCode plugin to ensure a good development experience.
    Downloads: 17 This Week
    Last Update:
    See Project
  • 21
    GEOFlow

    GEOFlow

    Open-source GEO content production system with AI tasks

    GEOFlow is a workflow system designed to manage and automate processes related to geographic and search optimization tasks using AI-driven pipelines. It focuses on structuring complex workflows into manageable steps, allowing users to orchestrate tasks such as content generation, analysis, and optimization. The system emphasizes modular design, enabling users to build reusable components that can be combined into larger workflows. It integrates with AI tools to enhance automation and decision-making within these pipelines. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 22
    Chirpy Jekyll Theme

    Chirpy Jekyll Theme

    A minimal, responsive and feature-rich Jekyll theme

    A minimal, responsive and feature-rich Jekyll theme for technical writing.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 23
    OpenMythos

    OpenMythos

    A theoretical reconstruction of the Claude Mythos architecture

    OpenMythos is an experimental, open-source implementation that attempts to reconstruct a hypothesized architecture behind advanced language models using a design called a Recurrent-Depth Transformer. The project explores the idea that instead of stacking hundreds of unique transformer layers, a smaller set of layers can be reused iteratively during inference to achieve deeper reasoning without increasing parameter count. It divides computation into three main stages, including a pre-processing phase, a looped recurrent reasoning block, and a final output refinement stage, creating a structured pipeline for inference. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    ModernBERT

    ModernBERT

    Bringing BERT into modernity via both architecture changes and scaling

    ModernBERT is an open-source research project that modernizes the classic BERT encoder architecture by incorporating recent advances in transformer design, training techniques, and efficiency improvements. The goal of the project is to bring BERT-style models up to date with the capabilities of modern large language models while preserving the strengths of bidirectional encoder architectures used for tasks such as classification, retrieval, and semantic search. ModernBERT introduces architectural improvements that enhance both training efficiency and inference performance, making the model more suitable for modern large-scale machine learning pipelines. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 25
    LLMs-from-scratch

    LLMs-from-scratch

    Implement a ChatGPT-like LLM in PyTorch from scratch, step by step

    ...The repository favors clear Python and NumPy or PyTorch implementations that can be run and modified without heavyweight frameworks obscuring the logic. Chapters and notebooks progress from tiny toy models to more capable transformer stacks, including sampling strategies and evaluation hooks. The focus is on readability, correctness, and experimentation, making it ideal for students and practitioners transitioning from theory to working systems. By the end, you have a grounded sense of how data pipelines, optimization, and inference interact to produce fluent text.
    Downloads: 2 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • 4
  • 5
  • Next
MongoDB Logo MongoDB