Showing 31 open source projects for "transformers"

View related business solutions
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    More flexibility. More control.

    Generate interest, access liquidity without selling, and execute trades seamlessly. All in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • Full-stack observability with actually useful AI | Grafana Cloud Icon
    Full-stack observability with actually useful AI | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 1
    Tabnine

    Tabnine

    Vim client for TabNine

    Tabnine is an AI-powered code completion extension trusted by millions of developers around the world. Whether you’re just getting started as a developer or if you’ve been doing it for decades, Tabnine will help you code twice as fast with half the keystrokes – all in your favorite IDE. Whether you call it IntelliSense, intelliCode, autocomplete, AI-assisted code completion, AI-powered code completion, AI copilot, AI code snippets, code suggestion, code prediction, code hinting, or...
    Downloads: 21 This Week
    Last Update:
    See Project
  • 2
    Sparse Attention

    Sparse Attention

    "Generating Long Sequences with Sparse Transformers" examples

    Sparse Attention is OpenAI’s code release for the Sparse Transformer model, introduced in the paper Generating Long Sequences with Sparse Transformers. It explores how modifying the self-attention mechanism with sparse patterns can reduce the quadratic scaling of standard transformers, making it possible to model much longer sequences efficiently. The repository provides implementations of sparse attention layers, training code, and evaluation scripts for benchmark datasets. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 3
    DETR

    DETR

    End-to-end object detection with transformers

    PyTorch training code and pretrained models for DETR (DEtection TRansformer). We replace the full complex hand-crafted object detection pipeline with a Transformer, and match Faster R-CNN with a ResNet-50, obtaining 42 AP on COCO using half the computation power (FLOPs) and the same number of parameters. Inference in 50 lines of PyTorch. What it is. Unlike traditional computer vision techniques, DETR approaches object detection as a direct set prediction problem. It consists of a set-based...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    PEG.js

    PEG.js

    PEG.js is a parser generator for JavaScript

    PEG.js is a simple parser generator for JavaScript that produces fast parsers with excellent error reporting. You can use it to process complex data or computer languages and build transformers, interpreters, compilers and other tools easily. PEG.js is still very much work in progress. There are no compatibility guarantees until version 1.0. Based on parsing expression grammar formalism, more powerful than traditional LL(k) and LR(k) parsers. Usable from your browser, from the command line, or via JavaScript API.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Go From AI Idea to AI App Fast Icon
    Go From AI Idea to AI App Fast

    One platform to build, fine-tune, and deploy ML models. No MLOps team required.

    Access Gemini 3 and 200+ models. Build chatbots, agents, or custom models with built-in monitoring and scaling.
    Try Free
  • 5
    Annotate your POJOs to create a fully-functional application, with table support, collection filters, and persistence hooks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    The Cocoon Adaption Kit is a NetKernel module which enables the use of Cocoon Components (Generators, Transformers, Serializer, Actions) from within the NetKernel XML Application Server.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB