Skip to yearly menu bar Skip to main content


Poster Thu, Dec 4, 2025 • 11:00 AM – 2:00 PM PST

ChunkKV: Semantic-Preserving KV Cache Compression for Efficient Long-Context LLM Inference

Xiang Liu ⋅ Zhenheng Tang ⋅ Peijie Dong ⋅ Zeyu Li ⋅ Liuyue ⋅ Bo Li ⋅ Xuming Hu ⋅ Xiaowen Chu

Abstract

Video

Chat is not available.