Showing 3 open source projects for "python library"

View related business solutions
  • Defend your online revenue Icon
    Defend your online revenue

    Remove counterfeits, stop piracy, and monitor sellers automatically to safeguard your brand's future

    Red Points is the world’s first brand intelligence platform. By combining online brand protection, copyright enforcement, and distributor monitoring capabilities, Red Points gives you full visibility into brands’ presence online. Over 650 brands rely on Red Points to fight fakes, piracy and distribution abuse online, allowing them to maintain control, improve their brand value, and increase revenues.
    Learn More
  • Point Of Sale Software for SMBs | Bravo Icon
    Point Of Sale Software for SMBs | Bravo

    Simplify Business with All In One Point of Sale for Specialty Retailers

    Bravo is changing the point of sale software game for small business specialty retailers with our easy to use, all in one solution. Say goodbye to working in multiple systems. Say hello to Bravo.
    Learn More
  • 1
    rwkv.cpp

    rwkv.cpp

    INT4/INT5/INT8 and FP16 inference on CPU for RWKV language model

    Besides the usual FP32, it supports FP16, quantized INT4, INT5 and INT8 inference. This project is focused on CPU, but cuBLAS is also supported. RWKV is a novel large language model architecture, with the largest model in the family having 14B parameters. In contrast to Transformer with O(n^2) attention, RWKV requires only state from the previous step to calculate logits. This makes RWKV very CPU-friendly on large context lengths.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 2
    fairseq2

    fairseq2

    FAIR Sequence Modeling Toolkit 2

    fairseq2 is a modern, modular sequence modeling framework developed by Meta AI Research as a complete redesign of the original fairseq library. Built from the ground up for scalability, composability, and research flexibility, fairseq2 supports a broad range of language, speech, and multimodal content generation tasks, including instruction fine-tuning, reinforcement learning from human feedback (RLHF), and large-scale multilingual modeling. Unlike the original fairseq—which evolved into a...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    FlashMLA

    FlashMLA

    FlashMLA: Efficient Multi-head Latent Attention Kernels

    FlashMLA is a high-performance decoding kernel library designed especially for Multi-Head Latent Attention (MLA) workloads, targeting NVIDIA Hopper GPU architectures. It provides optimized kernels for MLA decoding, including support for variable-length sequences, helping reduce latency and increase throughput in model inference systems using that attention style. The library supports both BF16 and FP16 data types, and includes a paged KV cache implementation with a block size of 64 to...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next