A Z-channel is a channel with binary input and binary output, where each 0 bit is transmitted correctly, but each 1 bit has probability p of being transmitted incorrectly as a 0, and probability 1–p of being transmitted correctly as a 1. In other words, if X and Y are the random variables describing the probability distributions of the input and the output of the channel, respectively, then the crossovers of the channel are characterized by the conditional probabilities:1
The channel capacity c a p ( Z ) {\displaystyle {\mathsf {cap}}(\mathbb {Z} )} of the Z-channel Z {\displaystyle \mathbb {Z} } with the crossover 1 → 0 probability p, when the input random variable X is distributed according to the Bernoulli distribution with probability α {\displaystyle \alpha } for the occurrence of 0, is given by the following equation:
where s ( p ) = H ( p ) 1 − p {\displaystyle {\mathsf {s}}(p)={\frac {{\mathsf {H}}(p)}{1-p}}} for the binary entropy function H ( ⋅ ) {\displaystyle {\mathsf {H}}(\cdot )} .
This capacity is obtained when the input variable X has Bernoulli distribution with probability α {\displaystyle \alpha } of having value 0 and 1 − α {\displaystyle 1-\alpha } of value 1, where:
For small p, the capacity is approximated by
as compared to the capacity 1 − H ( p ) {\displaystyle 1{-}{\mathsf {H}}(p)} of the binary symmetric channel with crossover probability p.
For any p > 0 {\displaystyle p>0} , α > 0.5 {\displaystyle \alpha >0.5} (i.e. more 0s should be transmitted than 1s) because transmitting a 1 introduces noise. As p → 1 {\displaystyle p\rightarrow 1} , the limiting value of α {\displaystyle \alpha } is 1 − 1 e {\displaystyle 1-{\frac {1}{e}}} .2
Define the following distance function d A ( x , y ) {\displaystyle {\mathsf {d}}_{A}(\mathbf {x} ,\mathbf {y} )} on the words x , y ∈ { 0 , 1 } n {\displaystyle \mathbf {x} ,\mathbf {y} \in \{0,1\}^{n}} of length n transmitted via a Z-channel
Define the sphere V t ( x ) {\displaystyle V_{t}(\mathbf {x} )} of radius t around a word x ∈ { 0 , 1 } n {\displaystyle \mathbf {x} \in \{0,1\}^{n}} of length n as the set of all the words at distance t or less from x {\displaystyle \mathbf {x} } , in other words,
A code C {\displaystyle {\mathcal {C}}} of length n is said to be t-asymmetric-error-correcting if for any two codewords c ≠ c ′ ∈ { 0 , 1 } n {\displaystyle \mathbf {c} \neq \mathbf {c} '\in \{0,1\}^{n}} , one has V t ( c ) ∩ V t ( c ′ ) = ∅ {\displaystyle V_{t}(\mathbf {c} )\cap V_{t}(\mathbf {c} ')=\emptyset } . Denote by M ( n , t ) {\displaystyle M(n,t)} the maximum number of codewords in a t-asymmetric-error-correcting code of length n.
The Varshamov bound. For n≥1 and t≥1,
The constant-weight code bound. For n > 2t ≥ 2, let the sequence B0, B1, ..., Bn-2t-1 be defined as
Then M ( n , t ) ≤ B n − 2 t − 1 . {\displaystyle M(n,t)\leq B_{n-2t-1}.}
MacKay (2003), p. 148. - MacKay, David J.C. (2003). Information Theory, Inference, and Learning Algorithms. Cambridge University Press. ISBN 0-521-64298-1. http://www.inference.phy.cam.ac.uk/mackay/itila/book.html ↩
MacKay (2003), p. 159. - MacKay, David J.C. (2003). Information Theory, Inference, and Learning Algorithms. Cambridge University Press. ISBN 0-521-64298-1. http://www.inference.phy.cam.ac.uk/mackay/itila/book.html ↩