Local Decode and Update for Big Data Compression

Research output: Contribution to journalArticlepeer-review

Abstract

This paper investigates data compression that simultaneously allows local decoding and local update. The main result is a universal compression scheme for memoryless sources with the following features. The rate can be made arbitrarily close to the entropy of the underlying source, contiguous fragments of the source can be recovered or updated by probing or modifying a number of codeword bits that is on average linear in the size of the fragment, and the overall encoding and decoding complexity is quasilinear in the blocklength of the source. In particular, the local decoding or update of a single message symbol can be performed by probing or modifying on average a constant number of codeword bits. This latter part improves over previous best known results for which local decodability or update efficiency grows logarithmically with blocklength.

Original languageEnglish
Article number9115085
Pages (from-to)5790-5805
Number of pages16
JournalIEEE Transactions on Information Theory
Volume66
Issue number9
DOIs
Publication statusPublished - 1 Sept 2020

Keywords

  • Data compression
  • big data applications
  • compression algorithms
  • source coding

Fingerprint

Dive into the research topics of 'Local Decode and Update for Big Data Compression'. Together they form a unique fingerprint.

Cite this