+
+

Related Products

  • KrakenD
    71 Ratings
    Visit Website
  • LM-Kit.NET
    23 Ratings
    Visit Website
  • Convesio
    53 Ratings
    Visit Website
  • Vertex AI
    783 Ratings
    Visit Website
  • RunPod
    205 Ratings
    Visit Website
  • Sogolytics
    864 Ratings
    Visit Website
  • StackAI
    48 Ratings
    Visit Website
  • Retool
    567 Ratings
    Visit Website
  • Cloudflare
    1,915 Ratings
    Visit Website
  • Google AI Studio
    11 Ratings
    Visit Website

About

LMCache is an open source Knowledge Delivery Network (KDN) designed as a caching layer for large language model serving that accelerates inference by reusing KV (key-value) caches across repeated or overlapping computations. It enables fast prompt caching, allowing LLMs to “prefill” recurring text only once and then reuse those stored KV caches, even in non-prefix positions, across multiple serving instances. This approach reduces time to first token, saves GPU cycles, and increases throughput in scenarios such as multi-round question answering or retrieval augmented generation. LMCache supports KV cache offloading (moving cache from GPU to CPU or disk), cache sharing across instances, and disaggregated prefill, which separates the prefill and decoding phases for resource efficiency. It is compatible with inference engines like vLLM and TGI and supports compressed storage, blending techniques to merge caches, and multiple backend storage options.

About

RAGFlow is an open source Retrieval-Augmented Generation (RAG) engine that enhances information retrieval by combining Large Language Models (LLMs) with deep document understanding. It offers a streamlined RAG workflow suitable for businesses of any scale, providing truthful question-answering capabilities backed by well-founded citations from various complex formatted data. Key features include template-based chunking, compatibility with heterogeneous data sources, and automated RAG orchestration.

Platforms Supported

Windows
Mac
Linux
Cloud
On-Premises
iPhone
iPad
Android
Chromebook

Platforms Supported

Windows
Mac
Linux
Cloud
On-Premises
iPhone
iPad
Android
Chromebook

Audience

AI engineers and infrastructure teams looking for a tool to lower latency, reduce compute cost, and scale throughput

Audience

AI developers seeking a solution to integrate retrieval-augmented generation capabilities into their applications, enhancing accuracy and traceability in AI-driven tasks

Support

Phone Support
24/7 Live Support
Online

Support

Phone Support
24/7 Live Support
Online

API

Offers API

API

Offers API

Screenshots and Videos

Screenshots and Videos

Pricing

Free
Free Version
Free Trial

Pricing

Free
Free Version
Free Trial

Reviews/Ratings

Overall 0.0 / 5
ease 0.0 / 5
features 0.0 / 5
design 0.0 / 5
support 0.0 / 5

This software hasn't been reviewed yet. Be the first to provide a review:

Review this Software

Reviews/Ratings

Overall 0.0 / 5
ease 0.0 / 5
features 0.0 / 5
design 0.0 / 5
support 0.0 / 5

This software hasn't been reviewed yet. Be the first to provide a review:

Review this Software

Training

Documentation
Webinars
Live Online
In Person

Training

Documentation
Webinars
Live Online
In Person

Company Information

LMCache
United States
lmcache.ai/

Company Information

RAGFlow
ragflow.io

Alternatives

Alternatives

DeepSeek-V2

DeepSeek-V2

DeepSeek
PrimoCache

PrimoCache

Romex Software

Categories

Categories

Integrations

Docker
Elestio

Integrations

Docker
Elestio
Claim LMCache and update features and information
Claim LMCache and update features and information
Claim RAGFlow and update features and information
Claim RAGFlow and update features and information