Showing 13 open source projects for "key value"

View related business solutions
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • Our Free Plans just got better! | Auth0 Icon
    Our Free Plans just got better! | Auth0

    With up to 25k MAUs and unlimited Okta connections, our Free Plan lets you focus on what you do best—building great apps.

    You asked, we delivered! Auth0 is excited to expand our Free and Paid plans to include more options so you can focus on building, deploying, and scaling applications without having to worry about your security. Auth0 now, thank yourself later.
    Try free now
  • 1
    vLLM

    vLLM

    A high-throughput and memory-efficient inference and serving engine

    vLLM is a fast and easy-to-use library for LLM inference and serving. High-throughput serving with various decoding algorithms, including parallel sampling, beam search, and more.
    Downloads: 23 This Week
    Last Update:
    See Project
  • 2
    R-KV

    R-KV

    Redundancy-aware KV Cache Compression for Reasoning Models

    R-KV is an open-source research project that focuses on improving the efficiency of large language model inference through key-value cache compression techniques. Modern transformer models rely heavily on KV caches during autoregressive decoding, which store intermediate attention states to accelerate generation. However, these caches can consume large amounts of memory, especially in reasoning-oriented models with long context windows. R-KV introduces a method for compressing the KV cache during decoding, allowing models to maintain reasoning performance while reducing memory consumption and computational overhead. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    CAG

    CAG

    Cache-Augmented Generation: A Simple, Efficient Alternative to RAG

    ...Traditional retrieval-augmented generation systems rely on real-time retrieval of documents from databases or vector stores during inference. CAG proposes a different approach by preloading relevant knowledge into the model’s context window and precomputing the model’s key-value cache before queries are processed. This strategy allows the model to generate responses using the cached context directly, eliminating the need for repeated retrieval operations during runtime. As a result, the approach can significantly reduce latency and simplify system architecture compared with traditional RAG pipelines. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 4
    UCCL

    UCCL

    UCCL is an efficient communication library for GPUs

    ...UCCL is designed to work with heterogeneous hardware environments, allowing GPUs from different vendors and network interfaces to communicate efficiently without vendor lock-in. The system also supports specialized workloads such as reinforcement learning weight transfers, key-value cache sharing, and expert parallelism for mixture-of-experts models. Its architecture emphasizes flexibility and extensibility so that developers can implement custom communication protocols tailored to specific machine learning workloads.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Application Monitoring That Won't Slow Your App Down Icon
    Application Monitoring That Won't Slow Your App Down

    AppSignal's Rust-based agent is lightweight and stable. Already running in thousands of production apps.

    Full APM with errors, performance, logs, and uptime monitoring. 99.999% uptime SLA on the platform itself.
    Start Free
  • 5
    KVCache-Factory

    KVCache-Factory

    Unified KV Cache Compression Methods for Auto-Regressive Models

    KVCache-Factory is an open-source research framework designed to explore and implement unified key-value cache compression techniques for autoregressive transformer models. In large language models, the key-value cache stores intermediate attention states that enable efficient token generation during inference, but these caches can consume large amounts of GPU memory when handling long contexts. KVCache-Factory provides a platform for implementing and evaluating multiple compression strategies that reduce memory usage while preserving model performance. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    node-red-contrib-custom-chatgpt
    A Node-RED node that interacts with OpenAI machine learning models like "ChatGPT". Install with the built-in Node-RED Palette manager. When editing the properties of the node, to get your OPENAI_API_KEY log in to ChatGPT. Create a new secret key" then copy and paste the "API key" into the node API_KEY property value. msg.payload should be a well-written prompt that provides enough information for the model to know what you want and how it should respond. Its success generally depends on the complexity of the task and quality of your prompt. A good rule of thumb is to think about how you would write a word problem for a middle schooler to solve. msg.payload should be a well-written prompt that provides enough information for the model to know what you want and how it should respond.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    LMCache

    LMCache

    Supercharge Your LLM with the Fastest KV Cache Layer

    LMCache is an extension layer for LLM serving engines that accelerates inference, especially with long contexts, by storing and reusing key-value (KV) attention caches across requests. Instead of rebuilding KV states for repeated or shared text segments, LMCache persists and retrieves them from multiple tiers—GPU memory, CPU DRAM, and local disk—then injects them into subsequent requests to reduce TTFT and increase throughput. Its design supports reuse beyond strict prefix matching and enables sharing across serving instances, improving efficiency under real multi-tenant traffic. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 8
    HelixDB

    HelixDB

    Graph-vector database for building unified AI backends fast

    ...HelixDB is built from scratch in Rust and uses LMDB as its storage engine, enabling high performance and low-latency query execution. HelixDB also supports additional data formats such as key-value, document, and relational data, making it flexible for a wide range of backend architectures. A central feature of the project is its custom query language, HelixQL, which is fully type-safe and compiled to ensure reliability and correctness in production environments. HelixDB includes built-in capabilities for embeddings, vector search, keyword search, and graph traversal, which are particularly useful for retrieval-augmented generation and agent-based systems.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    Mooncake

    Mooncake

    Mooncake is the serving platform for Kimi

    ...This engine enables efficient movement of tensors and model data across heterogeneous environments such as GPU memory, system memory, and distributed storage systems. Mooncake also introduces distributed key-value cache storage that allows inference systems to reuse previously computed attention states, significantly improving throughput in large-scale deployments. The system supports advanced networking technologies such as RDMA and NVMe over Fabric, enabling high-speed communication across clusters.
    Downloads: 0 This Week
    Last Update:
    See Project
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 10
    Audio AI Timeline

    Audio AI Timeline

    A timeline of the latest AI models for audio generation

    ...Its value comes from presenting progress in a chronological and thematic way, which makes trends, breakthroughs, and shifts in research focus easier to see.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    ML for Trading

    ML for Trading

    Code for machine learning for algorithmic trading, 2nd edition

    On over 800 pages, this revised and expanded 2nd edition demonstrates how ML can add value to algorithmic trading through a broad range of applications. Organized in four parts and 24 chapters, it covers the end-to-end workflow from data sourcing and model development to strategy backtesting and evaluation. Covers key aspects of data sourcing, financial feature engineering, and portfolio management. The design and evaluation of long-short strategies based on a broad range of ML algorithms, how to extract tradeable signals from financial text data like SEC filings, earnings call transcripts or financial news. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 12
    imgaug

    imgaug

    Image augmentation for machine learning experiments

    ...It supports a wide range of augmentation techniques, allows to easily combine these and to execute them in random order or on multiple CPU cores, has a simple yet powerful stochastic interface and can not only augment images but also key points/landmarks, bounding boxes, heatmaps and segmentation maps. Affine transformations, perspective transformations, contrast changes, gaussian noise, dropout of regions, hue/saturation changes, cropping/padding, blurring, etc. Rotate image and segmentation map on it by the same value sampled. Convert keypoints to distance maps, extract pixels within bounding boxes from images, clip polygon to the image plane, etc. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Perico is a dialogue management library. It allows you to specify a dialogue agent: a computer program that can interact with a user (or other agents) using (natural language) messages.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB