Showing 30 open source projects for "transformers"

View related business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Fully Managed MySQL, PostgreSQL, and SQL Server Icon
    Fully Managed MySQL, PostgreSQL, and SQL Server

    Automatic backups, patching, replication, and failover. Focus on your app, not your database.

    Cloud SQL handles your database ops end to end, so you can focus on your app.
    Try Free
  • 1
    Sparse Attention

    Sparse Attention

    "Generating Long Sequences with Sparse Transformers" examples

    Sparse Attention is OpenAI’s code release for the Sparse Transformer model, introduced in the paper Generating Long Sequences with Sparse Transformers. It explores how modifying the self-attention mechanism with sparse patterns can reduce the quadratic scaling of standard transformers, making it possible to model much longer sequences efficiently. The repository provides implementations of sparse attention layers, training code, and evaluation scripts for benchmark datasets. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 2
    DETR

    DETR

    End-to-end object detection with transformers

    PyTorch training code and pretrained models for DETR (DEtection TRansformer). We replace the full complex hand-crafted object detection pipeline with a Transformer, and match Faster R-CNN with a ResNet-50, obtaining 42 AP on COCO using half the computation power (FLOPs) and the same number of parameters. Inference in 50 lines of PyTorch. What it is. Unlike traditional computer vision techniques, DETR approaches object detection as a direct set prediction problem. It consists of a set-based...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    PEG.js

    PEG.js

    PEG.js is a parser generator for JavaScript

    PEG.js is a simple parser generator for JavaScript that produces fast parsers with excellent error reporting. You can use it to process complex data or computer languages and build transformers, interpreters, compilers and other tools easily. PEG.js is still very much work in progress. There are no compatibility guarantees until version 1.0. Based on parsing expression grammar formalism, more powerful than traditional LL(k) and LR(k) parsers. Usable from your browser, from the command line, or via JavaScript API.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Annotate your POJOs to create a fully-functional application, with table support, collection filters, and persistence hooks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Full-stack observability with actually useful AI | Grafana Cloud Icon
    Full-stack observability with actually useful AI | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 5
    The Cocoon Adaption Kit is a NetKernel module which enables the use of Cocoon Components (Generators, Transformers, Serializer, Actions) from within the NetKernel XML Application Server.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB