Academia.edu no longer supports Internet Explorer.
To browse Academia.edu and the wider internet faster and more securely, please take a few seconds to upgrade your browser.
Variable-length codes can provide compression for data communication. Such codes may be used not only when the source statistics is known but also when we do not know the source probability distribution, and a source with equal symbol probabilities (equiprobable symbols) can or has to be assumed. This paper presents variable-length codes with code words that differ in length by at most one code symbol. Such codes suit the efficient encoding of sources with equiprobable symbols. We accommodate non-binary codes and present an iterative algorithm for the construction of such codes. We also calculate the average codeword length for such codes, which extends Krichevski's result for binary codes [5]. Finally, we propose a scheme that allows the code to be communicated efficiently from transmitter to receiver.
ACM Transactions on Multimedia Computing, Communications, and Applications, 2018
For the entropy coding of independent and identically distributed (i.i.d.) binary sources, variable-to-variable length (V2V) codes are an interesting alternative to arithmetic coding. Such a V2V code translates variable length words of the source into variable length code words by employing two prefix-free codes. In this article, several properties of V2V codes are studied, and new concepts are developed. In particular, it is shown that the redundancy of a V2V code cannot be zero for a binary i.i.d. source {X} with 0 < p X (1) < 0.5. Furthermore, the concept of prime and composite V2V codes is proposed, and it is shown why composite V2V codes can be disregarded in the search for particular classes of minimum redundancy codes. Moreover, a canonical representation for V2V codes is proposed, which identifies V2V codes that have the same average code length function. It is shown how these concepts can be employed to greatly reduce the complexity of a search for minimum redundancy ...
IEEE Transactions on Information Theory, 1994
We consider the problem of source coding. We investigate the cases of known and unknown statistics. The efficiency of the compression codes can be estimated by three characteristics: 1) the rebundancy (r), defined as the maximal difference between the average codeword length and Shannon entropy in case the letters are generated by a Bernoulli source;
Discrete Applied Mathematics, 1996
Several measures are defined and investigated, which allow the comparison of codes as to their robustness against errors. Then new universal and complete sequences of variable-length codewords are proposed, based on representing the integers in a binary Fibonacci numeration system. Each sequence is constant and needs not be generated for every probability distribution. These codes can be used as alternatives to Huffman codes when the optimal compression of the latter is not required, and simplicity, faster processing and robustness are preferred. The codes are compared on several "real-life" examples.
2007
We offer novel algorithms for efficient encoding/decoding of variable-to-fixed length codes, requiring at most quadratic amount of space: O(L), where L is the depth of a coding tree. This is a major improvement compared to exponential O(2) usage of space by conventional techniques using complete representations of coding trees in computer’s memory. These savings are achieved by utilizing algebraic properties of VF coding trees constructed by using Tunstall or Khodak algorithms, and employing combinatorial enumeration techniques for encoding/decoding of codewords. The encoding/decoding complexity of our algorithms is linear with the number of symbols they process. As a side product, we also derive an exact formulae for the average redundancy of such codes under memoryless sources, and show its usefulness for analysis and design of codes with small number of codewords. 1 Definitions Consider a memoryless source S producing symbols from an input alphabet A = {a1, . . . , am} (2 6 m <...
Rissanen's lower bound on the worstcase redundancy over the maximum likelihood (ML) coding length was obtained by analyzing the performance of two-part codes. We consider a class of particularly simple two-part codes for Bernoulli sequences that quantize the ML parameter estimate into one of K bins, describe that bin with log 2 (K) bits, and encode the input sequence according to the quantized parameter estimate. In addition to their simple structure, these codes have an appealing application to universal distributed lossless source coding of Bernoulli sequences. We propose several schemes for quantizing the parameter estimate; the best worstcase redundancy for the entire sequence attainable in our class of codes is 1.047 bits above Rissanen's bound.
ArXiv, 2021
We consider the issue of direct access to any letter of a sequence encoded with a variable length code and stored in the computer’s memory, which is a special case of the random access problem to compressed memory. The characteristics according to which methods are evaluated are the access time to one letter and the memory used. The proposed methods, with various trade-offs between the characteristics, outperform the known ones. keywords: source coding, data compression, random access, direct access, compressed memory.
IEEE Communications Letters, 2000
We show how low-density parity-check (LDPC) codes can be used to compress close to the Slepian-Wolf limit for correlated binary sources. Focusing on the asymmetric case of compression of an equiprobable memoryless binary source with side information at the decoder, the approach is based on viewing the correlation as a channel and applying the syndrome concept. The encoding and decoding procedures are explained in detail. The performance achieved is seen to be better than recently published results using turbo codes and very close to the Slepian-Wolf limit.
IEEE Transactions on Information Theory, 1984
2013
X is a finite discrete memoryless random source whose output probability distribution is known. A binary variable length encoder (like a huffman encoder or an arithmetic encoder) encodes the source symbols into corresponding binary variable length codes (vlc) of lengths respectively. These information words have to be transmitted over a discrete memoryless noisy channel. So to provide sufficient protection against channel errors, these variable length information words are converted into fixed-length channel codewords. The partial Hamming distance for such variable length codes is defined and then a recursive algorithm to find the extended fixed-length channel codewords is formulated.
1998
This paper calculates new bounds on the size of the performance gap between random codes and the best possible codes. The first result shows that, for large block sizes, the ratio of the error probability of a random code to the sphere-packing lower bound on the error probability of every code on the binary symmetric channel (BSC) is small for a wide range of useful crossover probabilities. Thus even far from capacity, random codes have nearly the same error performance as the best possible long codes. The paper also demonstrates that a small reduction k 0 ~ k in the number of information bits conveyed by a codeword will make the error performance of an (n; ~ k) random code better than the sphere-packing lower bound for an (n; k) code as long as the channel crossover probability is somewhat greater than a critical probability. For example, the sphere-packing lower bound for a long (n; k), rate 1=2, code will exceed the error probability of an (n; ~ k) random code if k0 ~ k?10 and th...
2007
Abstract We consider the problem of choosing a block-length to achieve a desired probability of error for fixed-length lossless source coding and channel coding of a finite amount of payload data. This is closely related to the issue of redundancy. While Baron, et al.
2007
We offer novel algorithms for efficient encoding/decoding of variable-to-fixed length codes, requiring at most quadratic amount of space: O(L 2), where L is the depth of a coding tree. This is a major improvement compared to exponential O(2 L) usage of space by conventional techniques using complete representations of coding trees in computer's memory. These savings are achieved by utilizing algebraic properties of VF coding trees constructed by using Tunstall or Khodak algorithms, and employing combinatorial enumeration techniques for encoding/decoding of codewords. The encoding/decoding complexity of our algorithms is linear with the number of symbols they process. As a side product, we also derive an exact formulae for the average redundancy of such codes under memoryless sources, and show its usefulness for analysis and design of codes with small number of codewords.
Problems of Information Transmission, 2012
The compression-complexity trade-off of lossy compression algorithms that are based on a random codebook or a random database is examined. Motivated, in part, by recent results of Gupta-Verdú-Weissman (GVW) and their underlying connections with the pattern-matching scheme of Kontoyiannis' lossy Lempel-Ziv algorithm, we introduce a non-universal version of the lossy Lempel-Ziv method (termed LLZ). The optimality of LLZ for memoryless sources is established, and its performance is compared to that of the GVW divide-and-conquer approach. Experimental results indicate that the GVW approach often yields better compression than LLZ, but at the price of much higher memory requirements. To combine the advantages of both, we introduce a hybrid algorithm (HYB) that utilizes both the divide-and-conquer idea of GVW and the single-database structure of LLZ. It is proved that HYB shares with GVW the exact same rate-distortion performance and implementation complexity, while, like LLZ, requiring less memory, by a factor which may become unbounded, depending on the choice of the relevant design parameters. Experimental results are also presented, illustrating the performance of all three methods on data generated by simple discrete memoryless sources. In particular, the HYB algorithm is shown to outperform existing schemes for the compression of some simple discrete sources with respect to the Hamming distortion criterion. 2 codes, turbo codes, and local message-passing decoding algorithms; see, e.g., [33][48][34], the texts [16][32][44], and the references therein.
1995
Two construction techniques for variable-length error-correcting (VLEC) codes are given. The first uses fixed-length linear codes and anticodes to build new VLEC codes, whereas the second uses a heuristic algorithm to perform a computer search for good VLEC codes. VLEC codes may be used for combined source and channel coding. It is shown that over an additive white Gaussian noise channel the codes so constructed can perform better than standard cascaded source and channel codes with similar parameters.
Communications of the ACM, 1987
2009
We establish a general framework for construction of small ensembles of capacity achieving linear codes for a wide range of (not necessarily memoryless) discrete symmetric channels, and in particular, the binary erasure and symmetric channels. The main tool used in our constructions is the notion of randomness extractors and lossless condensers that are regarded as central tools in theoretical computer science. Same as random codes, the resulting ensembles preserve their capacity achieving properties under any change of basis. Using known explicit constructions of condensers, we obtain specific ensembles whose size is as small as polynomial in the block length. By applying our construction to Justesen's concatenation scheme (Justesen, 1972) we obtain explicit capacity achieving codes for BEC (resp., BSC) with almost linear time encoding and almost linear time (resp., quadratic time) decoding and exponentially small error probability.
arXiv: Information Theory, 2015
We propose almost instantaneous fixed-to-variable-length (AIFV) codes such that two (resp. K − 1) code trees are used if code symbols are binary (resp. K-ary for K ≥ 3), and source symbols are assigned to incomplete internal nodes in addition to leaves. Although the AIFV codes are not instantaneous codes, they are devised such that the decoding delay is at most two bits (resp. one code symbol) in the case of binary (resp. K-ary) code alphabet. The AIFV code can attain better average compression rate than the Huffman code at the expenses of a little decoding delay and a little large memory size to store multiple code trees. We also show for the binary and ternary AIFV codes that the optimal AIFV code can be obtained by solving 0-1 integer programming problems. Index Terms AIFV code, Huffman code, FV code, code tree, Kraft inequality, Integer programming I. INTRODUCTION Lossless source codes are classified into fixed-to-variable-length (FV) codes and variable-to-fixed-length (VF) codes, which can be represented by code trees and parse trees, respectively. It is well known that the Huffman coding [1] and Tunstall coding [2] can attain the best compression rate in FV codes and VF codes, respectively, for stationary memoryless sources if a single code tree or a single parse tree is used. But, Yamamoto and Yokoo [3] showed that the AIVF (almost instantaneous variable-to-fixed length) coding can attain better compression rate than the Tunstall coding. An AIVF code uses |X |−1 parse trees for a source alphabet X and codewords are assigned to incomplete internal nodes in addition to leaves in each parse tree. Although instantaneous encoding is not possible since incomplete internal nodes are used for encoding, the AIVF code is devised such that the encoding delay is at most one source symbol, and hence the code is called almost instantaneous. Furthermore, Yoshida and Kida [4][5] showed that any AIVF code can be encoded and decoded by a single virtual multiple parse tree and the total number of nodes can be considerably reduced by the integration. In the case of FV codes, it is well known by Kraft and McMillan Theorems [6][7][8] that any uniquely decodable FV code must satisfy Kraft's inequality, and such a code can be realized by an instantaneous FV
Loading Preview
Sorry, preview is currently unavailable. You can download the paper by clicking the button above.