Skip to content
View xiaguan's full-sized avatar
🎯
Focusing
🎯
Focusing

Block or report xiaguan

Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
xiaguan/README.md

Pinned Loading

  1. kvcache-ai/Mooncake kvcache-ai/Mooncake Public

    Mooncake is the serving platform for Kimi, a leading LLM service provided by Moonshot AI.

    C++ 5.2k 720

  2. sgl-project/sglang sgl-project/sglang Public

    SGLang is a high-performance serving framework for large language models and multimodal models.

    Python 26.9k 5.7k

  3. LMCache/LMCache LMCache/LMCache Public

    Supercharge Your LLM with the Fastest KV Cache Layer

    Python 8.2k 1.1k

  4. foyer-rs/foyer foyer-rs/foyer Public

    Hybrid in-memory and disk cache in Rust

    Rust 1.7k 83

  5. novitalabs/pegaflow novitalabs/pegaflow Public

    High-performance KV cache storage for LLM inference — GPU offloading, SSD caching, and cross-node sharing via RDMA. Works with vLLM and SGLang.

    Rust 51 9

  6. pegainfer pegainfer Public

    Pure Rust + CUDA LLM inference engine

    Rust 297 28