Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Igal Sason is active.

Publication


Featured researches published by Igal Sason.


IEEE Transactions on Information Theory | 2002

Variations on the Gallager bounds, connections, and applications

Shlomo Shamai; Igal Sason

There has been renewed interest in deriving tight bounds on the error performance of specific codes and ensembles, based on their distance spectrum. We discuss many reported upper bounds on the maximum-likelihood (ML) decoding error probability and demonstrate the underlying connections that exist between them. In addressing the Gallager bounds and their variations, we focus on the Duman and Salehi (see IEEE Trans. Commun., vol.46, p.717-723, 1998)variation, which originates from the standard Gallager bound. A large class of efficient bounds (or their Chernoff versions) is demonstrated to be a special case of the generalized second version of the Duman and Salehi bounds. Implications and applications of these observations are pointed out, including the fully interleaved fading channel, resorting to either matched or mismatched decoding. The proposed approach can be generalized to geometrically uniform nonbinary codes, finite-state channels, bit interleaved coded modulation systems, and it can be also used for the derivation of upper bounds on the conditional decoding error probability.


IEEE Transactions on Information Theory | 2003

Parity-check density versus performance of binary linear block codes over memoryless symmetric channels

Igal Sason; Rüdiger L. Urbanke

We derive lower bounds on the density of parity-check matrices of binary linear codes which are used over memoryless binary-input output-symmetric (MBIOS) channels. The bounds are expressed in terms of the gap between the rate of these codes for which reliable communications is achievable and the channel capacity; they are valid for every sequence of binary linear block codes if there exists a decoding algorithm under which the average bit-error probability vanishes. For every MBIOS channel, we construct a sequence of ensembles of regular low-density parity-check (LDPC) codes, so that an upper bound on the asymptotic density of their parity-check matrices scales similarly to the lower bound. The tightness of the lower bound is demonstrated for the binary erasure channel by analyzing a sequence of ensembles of right-regular LDPC codes which was introduced by Shokrollahi, and which is known to achieve the capacity of this channel. Under iterative message-passing decoding, we show that this sequence of ensembles is asymptotically optimal (in a sense to be defined in this paper), strengthening a result of Shokrollahi. Finally, we derive lower bounds on the bit-error probability and on the gap to capacity for binary linear block codes which are represented by bipartite graphs, and study their performance limitations over MBIOS channels. The latter bounds provide a quantitative measure for the number of cycles of bipartite graphs which represent good error-correction codes.


IEEE Transactions on Information Theory | 2000

Improved upper bounds on the ML decoding error probability of parallel and serial concatenated turbo codes via their ensemble distance spectrum

Igal Sason; Shlomo Shamai

The ensemble performance of parallel and serial concatenated turbo codes is considered, where the ensemble is generated by a uniform choice of the interleaver and of the component codes taken from the set of time-varying recursive systematic convolutional codes. Following the derivation of the input-output weight enumeration functions of the ensembles of random parallel and serial concatenated turbo codes, the tangential sphere upper bound is employed to provide improved upper bounds on the block and bit error probabilities of these ensembles of codes for the binary-input additive white Gaussian noise (AWGN) channel, based on coherent detection of equi-energy antipodal signals and maximum-likelihood decoding. The influence of the interleaver length and the memory length of the component codes is investigated. The improved bounding technique proposed here is compared to the conventional union bound and to a alternative bounding technique by Duman and Salehi (1998) which incorporates modified Gallager bounds. The advantage of the derived bounds is demonstrated for a variety of parallel and serial concatenated coding schemes with either fixed or random recursive systematic convolutional component codes, and it is especially pronounced in the region exceeding the cutoff rate, where the performance of turbo codes is most appealing. These upper bounds are also compared to simulation results of the iterative decoding algorithm.


Foundations and Trends in Communications and Information Theory | 2006

Performance analysis of linear codes under maximum-likelihood decoding: a tutorial

Igal Sason; Shlomo Shamai

This article is focused on the performance evaluation of linear codes under optimal maximum-likelihood (ML) decoding. Though the ML decoding algorithm is prohibitively complex for most practical codes, their performance analysis under ML decoding allows to predict their performance without resorting to computer simulations. It also provides a benchmark for testing the sub-optimality of iterative (or other practical) decoding algorithms. This analysis also establishes the goodness of linear codes (or ensembles), determined by the gap between their achievable rates under optimal ML decoding and information theoretical limits. In this article, upper and lower bounds on the error probability of linear codes under ML decoding are surveyed and applied to codes and ensembles of codes on graphs. For upper bounds, we discuss various bounds where focus is put on Gallager bounding techniques and their relation to a variety of other reported bounds. Within the class of lower bounds, we address de Caens based bounds and their improvements, and also consider sphere-packing bounds with their recent improvements targeting codes of moderate block lengths.


Foundations and Trends in Communications and Information Theory | 2013

Concentration of Measure Inequalities in Information Theory, Communications, and Coding

Maxim Raginsky; Igal Sason

Concentration inequalities have been the subject of exciting developments during the last two decades, and have been intensively studied and used as a powerful tool in various areas. These include convex geometry, functional analysis, statistical physics, mathematical statistics, pure and applied probability theory (e.g., concentration of measure phenomena in random graphs, random matrices, and percolation), information theory, theoretical computer science, learning theory, and dynamical systems.This monograph focuses on some of the key modern mathematical tools that are used for the derivation of concentration inequalities, on their links to information theory, and on their various applications to communications and coding. In addition to being a survey, this monograph also includes various new recent results derived by the authors.The first part of the monograph introduces classical concentration inequalities for martingales, aswell as some recent refinements and extensions. The power and versatility of the martingale approach is exemplified in the context of codes defined on graphs and iterative decoding algorithms, as well as codes for wireless communication.The second part of the monograph introduces the entropy method, an information-theoretic technique for deriving concentration inequalities for functions of many independent random variables. The basic ingredients of the entropy method are discussed first in conjunction with the closely related topic of logarithmic Sobolev inequalities, which are typical of the so-called functional approach to studying the concentration of measure phenomenon. The discussion on logarithmic Sobolev inequalities is complemented by a related viewpoint based on probability in metric spaces. This viewpoint centers around the so-called transportation-cost inequalities, whose roots are in information theory. Some representative results on concentration for dependent random variables are briefly summarized, with emphasis on their connections to the entropy method. Finally, we discuss several applications of the entropy method and related information-theoretic tools to problems in communications and coding. These include strong converses, empirical distributions of good channel codes with non-vanishing error probability, and an information-theoretic converse for concentration of measure.


IEEE Transactions on Information Theory | 2001

On improved bounds on the decoding error probability of block codes over interleaved fading channels, with applications to turbo-like codes

Igal Sason; Shlomo Shamai

We derive here improved upper bounds on the decoding error probability of block codes which are transmitted over fully interleaved Rician fading channels, coherently detected and maximum-likelihood (ML) decoded. We assume that the fading coefficients during each symbol are statistically independent (due to a perfect channel interleaver), and that perfect estimates of these fading coefficients are provided to the receiver. The improved upper bounds on the block and bit error probabilities are derived for fully interleaved fading channels with various orders of space diversity, and are found by generalizing some previously introduced upper bounds for the binary-input additive white Gaussian nose (AWGN) channel. The advantage of these bounds over the ubiquitous union bound is demonstrated for some ensembles of turbo codes and low-density parity-check (LDPC) codes, and it is especially pronounced in a portion of the rate region exceeding the cutoff rate. Our generalization of the Duman and Salehi bound (Duman and Salehi 1998, Duman 1998) which is based on certain variations of Gallagers (1965) bounding technique, is demonstrated to be the tightest reported upper bound. We therefore apply it to calculate numerically upper bounds on the thresholds of some ensembles of turbo-like codes, referring to the optimal ML decoding. For certain ensembles of uniformly interleaved turbo codes, the upper bounds derived here also indicate good match with computer simulation results of efficient iterative decoding algorithms.


IEEE Transactions on Information Theory | 2008

An Improved Sphere-Packing Bound for Finite-Length Codes Over Symmetric Memoryless Channels

Gil Wiechman; Igal Sason

This paper derives an improved sphere-packing (ISP) bound for finite-length error-correcting codes whose transmission takes place over symmetric memoryless channels, and the codes are decoded with an arbitrary list decoder. We first review classical results, i.e., the 1959 sphere-packing (SP59) bound of Shannon for the Gaussian channel, and the 1967 sphere-packing (SP67) bound of Shannon et al. for discrete memoryless channels. An improvement on the SP67 bound, as suggested by Valembois and Fossorier, is also discussed. These concepts are used for the derivation of a new lower bound on the error probability of list decoding (referred to as the ISP bound) which is uniformly tighter than the SP67 bound and its improved version. The ISP bound is applicable to symmetric memoryless channels, and some of its applications are presented. Its tightness under maximum-likelihood (ML) decoding is studied by comparing the ISP bound to previously reported upper and lower bounds on the ML decoding error probability, and also to computer simulations of iteratively decoded turbo-like codes. This paper also presents a technique which performs the entire calculation of the SP59 bound in the logarithmic domain, thus facilitating the exact calculation of this bound for moderate to large block lengths without the need for the asymptotic approximations provided by Shannon.


IEEE Transactions on Information Theory | 1999

On interleaved, differentially encoded convolutional codes

Michael Peleg; Igal Sason; Shlomo Shamai; Avner Elia

We study a serially interleaved concatenated code construction, where the outer code is a standard convolutional code, and the inner code is a recursive convolutional code of rate 1. We focus on the ubiquitous inner differential encoder (used, in particular, to resolve phase ambiguities), double differential encoder (used to resolve both phase and frequency ambiguities), and another rate 1 recursive convolutional code of memory 2. We substantiate analytically the rather surprising result, that the error probabilities corresponding to a maximum-likelihood (ML) coherently detected antipodal modulation over the additive white Gaussian noise (AWGN) channel for this construction are advantageous as compared to the stand-alone outer convolutional code. This is in spite of the fact that the inner code is of rate 1. The analysis is based on the tangential sphere upper bound of an ML decoder, incorporating the ensemble weight distribution (WD) of the concatenated code, where the ensemble is generated by all random and uniform interleavers. This surprising result is attributed to the WD thinning observed for the concatenated scheme which shapes the WD of the outer convolutional code to resemble more closely the binomial distribution (typical of a fully random code of the same length and rate). This gain is maintained regardless of a rather dramatic decrease, as demonstrated here, in the minimum distance of the concatenated scheme as compared to the minimum distance of the outer stand-alone convolutional code. The advantage of the examined serially interleaved concatenated code, given in terms of bit and/or block error probability which is decoded by a practical suboptimal decoder, over the optimally decoded standard convolutional code is demonstrated by simulations, and some insights into the performance of the iterative decoding algorithm are also discussed. Though we have investigated only specific constructions of constituent inner (rate 1) and outer codes, we trust, hinging on the rational of the arguments here, that these results extend to many other constituent convolutional outer codes and rate 1 inner recursive convolutional codes.


IEEE Transactions on Information Theory | 2007

Accumulate–Repeat–Accumulate Codes: Capacity-Achieving Ensembles of Systematic Codes for the Erasure Channel With Bounded Complexity

Henry D. Pfister; Igal Sason

This paper introduces ensembles of systematic accumulate-repeat-accumulate (ARA) codes which asymptotically achieve capacity on the binary erasure channel (BEC) with bounded complexity, per information bit, of encoding and decoding. It also introduces symmetry properties which play a central role in the construction of new capacity-achieving ensembles for the BEC. The results here improve on the tradeoff between performance and complexity provided by previous constructions of capacity-achieving code ensembles defined on graphs. The superiority of ARA codes with moderate to large block length is exemplified by computer simulations which compare their performance with those of previously reported capacity-achieving ensembles of low-density parity-check (LDPC) and irregular repeat-accumulate (IRA) codes. ARA codes also have the advantage of being systematic.


IEEE Transactions on Information Theory | 2002

On the asymptotic input-output weight distributions and thresholds of convolutional and turbo-like encoders

Igal Sason; I. Emre Telatar; Rüdiger L. Urbanke

We present a general method for computing the asymptotic input-output weight distribution of convolutional encoders. In some instances, one can derive explicit analytic expressions. In general, though, to determine the growth rate of the input-output weight distribution for a particular normalized input weight /spl kappa/ and output weight /spl omega/, a system of polynomial equations has to be solved. This method is then used to determine the asymptotic weight distribution of various concatenated code ensembles and to derive lower bounds on the thresholds of these ensembles under maximum-likelihood (ML) decoding.

Collaboration


Dive into the Igal Sason's collaboration.

Top Co-Authors

Avatar

Shlomo Shamai

Technion – Israel Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Gil Wiechman

Technion – Israel Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Eran Hof

Technion – Israel Institute of Technology

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Rüdiger L. Urbanke

École Polytechnique Fédérale de Lausanne

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Boaz Shuval

Technion – Israel Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Eshed Ram

Technion – Israel Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Idan Goldenberg

Technion – Israel Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Moshe Twitto

Technion – Israel Institute of Technology

View shared research outputs
Researchain Logo
Decentralizing Knowledge