[Paper Note] Sparse Indexing Large Scale, Inline Deduplication Using Sampling and Locality
Motivation
If the data deduplication system stores each deduplicated chunk in it’s index, the index size will soon exceeds the memory capacity due to the super large dataset size. Consider, for example, a store that contains 10 TB of unique data and uses 4 KB chunks. Then there are 2.7 × 109 unique chunks. Assuming that every hash entry in the index consumes 40 bytes, we need 100 GB of storage for the full index.
To overcome memory capacity limitation, the index has to be offloaded into disk. However, offloading the index into disk, we need one disk IO per chunk lookup. If an IO takes 4ms, the offloaded index can only achieve 250 lookup per second.
![Featured image for [Paper Note] Sparse indexing Large scale, inline deduplication using sampling and locality](/posts/sparse-indexing-large-scale-inline-deduplication-using-sampling-and-locality/images/pasted-image-20250911230506.png)
![Featured image for [Paper Note] IMPRESS An Importance-Informed Multi-Tier Prefix KV Storage System for Large Language Model Inference](/posts/impress-an-importance-informed-multi-tier-prefix-kv-storage-system-for-large-language-model-inference/images/pasted-image-20250907153049.png)
![Featured image for [Paper Note] Attentionstore Cost-effective attention reuse across multi-turn conversations in large language model serving](/posts/attentionstore-cost-effective-attention-reuse-across-multi-turn-conversations-in-large-language-model-serving/images/pasted-image-20250703034132.png)
![Featured image for [Paper Note] Orca A Distributed Serving System for Transformer-Based Generative Models](/posts/orca-a-distributed-serving-system-for-transformer-based-generative-models/images/pasted-image-20250702050548.png)

![Featured image for [Paper Note] CacheGen KV Cache Compression and Streaming for Fast Large Language Model Serving](/posts/cachegen-kv-cache-compression-and-streaming-for-fast-large-language-model-serving/images/cachegen-kv-cache-architecture.png)
![Featured image for [Paper Note] SGLang Efficient Execution of Structured Language Model Programs](/posts/sglang-efficient-execution-of-structured-language-model-programs/images/pasted-image-20250607093516.png)
