Abstract

Synchronization of replicated data and program state is an essential aspect of application fault-tolerance. Current solutions use virtual memory mapping to identify page writes and replicate them at the destination. This approach has limitations because the granularity is restricted to a minimum of 4KiB per page, which may result in more data being replicated. Motivated by the emerging CXL hardware, we expand on the work Waddington, et al. [SoCC 22] by evaluating popular compression algorithms on VM snapshot data at cache line granularity. We measure the compression ratio vs. the compression time and present our conclusions.

Keywords:
Computer science Cache Granularity Snapshot (computer storage) Data compression Compression ratio Replicate Parallel computing CPU cache Operating system Algorithm

Metrics

2
Cited By
0.88
FWCI (Field Weighted Citation Impact)
1
Refs
0.61
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Advanced Data Storage Technologies
Physical Sciences →  Computer Science →  Computer Networks and Communications
Parallel Computing and Optimization Techniques
Physical Sciences →  Computer Science →  Hardware and Architecture
Distributed systems and fault tolerance
Physical Sciences →  Computer Science →  Computer Networks and Communications
© 2026 ScienceGate Book Chapters — All rights reserved.