Channel polarization: A method for constructing capacity-achieving codes for symmetric binary-input memoryless channels

Size: px
Start display at page:

Download "Channel polarization: A method for constructing capacity-achieving codes for symmetric binary-input memoryless channels"

Transcription

1 Channel polarization: A method for constructing capacity-achieving codes for symmetric binary-input memoryless channels Erdal Arıkan, Senior Member, IEEE arxiv:007397v5 [csit] 0 Jul 009 Abstract A method is proposed, called channel polarization, to construct code sequences that achieve the symmetric capacity I() of any given binary-input discrete memoryless channel (B- DMC) The symmetric capacity is the highest rate achievable subject to using the input letters of the channel with equal probability Channel polarization refers to the fact that it is possible to synthesize, out of independent copies of a given B-DMC, a second set of binary-input channels { (i) : i } such that, as becomes large, the fraction of indices i for which I( (i) ) is near approaches I() and the fraction for which I( (i) ) is near 0 approaches I() The polarized channels { (i) } are well-conditioned for channel coding: one need only send data at rate through those with capacity near and at rate 0 through the remaining Codes constructed on the basis of this idea are called polar codes The paper proves that, given any B-DMC with I() > 0 and any target rate R < I(), there exists a sequence of polar codes {C n; n } such that C n has block-length n, rate R, and probability of block error under successive cancellation decoding bounded as P e(, R) O( ) independently of the code rate This performance is achievable by encoders and decoders with complexity O( log ) for each Index Terms Capacity-achieving codes, channel capacity, channel polarization, Plotkin construction, polar codes, Reed- Muller codes, successive cancellation decoding I ITRODUCTIO AD OVERVIE A fascinating aspect of Shannon s proof of the noisy channel coding theorem is the random-coding method that he used to show the existence of capacity-achieving code sequences without exhibiting any specific such sequence [] Explicit construction of provably capacity-achieving code sequences with low encoding and decoding complexities has since then been an elusive goal This paper is an attempt to meet this goal for the class of B-DMCs e will give a description of the main ideas and results of the paper in this section First, we give some definitions and state some basic facts that are used throughout the paper A Preliminaries e write : X Y to denote a generic B-DMC with input alphabet X, output alphabet Y, and transition E Arıkan is with the Department of Electrical-Electronics Engineering, Bilkent University, Ankara, 0600, Turkey ( arikan@eebilkentedutr) This work was supported in part by The Scientific and Technological Research Council of Turkey (T UBITAK) under Project 07E6 and in part by the European Commission FP7 etwork of Excellence ECOM under contract 675 probabilities (y x), x X, y Y The input alphabet X will always be {0, }, the output alphabet and the transition probabilities may be arbitrary e write to denote the channel corresponding to uses of ; thus, : X Y with (y x ) i (y i x i ) Given a B-DMC, there are two channel parameters of primary interest in this paper: the symmetric capacity I() (y x)log (y x) y Y x X (y 0) (y ) and the Bhattacharyya parameter Z() y Y (y 0)(y ) These parameters are used as measures of rate and reliability, respectively I() is the highest rate at which reliable communication is possible across using the inputs of with equal frequency Z() is an upper bound on the probability of maximum-likelihood (ML) decision error when is used only once to transmit a 0 or It is easy to see that Z() takes values in [0, ] Throughout, we will use base- logarithms; hence, I() will also take values in [0, ] The unit for code rates and channel capacities will be bits Intuitively, one would expect that I() iff Z() 0, and I() 0 iff Z() The following bounds, proved in the Appendix, make this precise Proposition : For any B-DMC, we have I() log Z(), () I() Z() () The symmetric capacity I() equals the Shannon capacity when is a symmetric channel, ie, a channel for which there exists a permutation π of the output alphabet Y such that (i) π π and (ii) (y ) (π(y) 0) for all y Y The binary symmetric channel (BSC) and the binary erasure channel (BEC) are examples of symmetric channels A BSC is a B-DMC with Y {0, }, (0 0) ( ), and ( 0) (0 ) A B-DMC is called a BEC if for each y Y, either (y 0)(y ) 0 or (y 0) (y ) In the latter case, y is said to be an erasure symbol The sum of (y 0) over all erasure symbols y is called the erasure probability of the BEC

2 e denote random variables (RVs) by upper-case letters, such as X, Y, and their realizations (sample values) by the corresponding lower-case letters, such as x, y For X a RV, P X denotes the probability assignment on X For a joint ensemble of RVs (X, Y ), P X,Y denotes the joint probability assignment e use the standard notation I(X; Y ), I(X; Y Z) to denote the mutual information and its conditional form, respectively e use the notation a as shorthand for denoting a row vector (a,,a ) Given such a vector a, we write a j i, i, j, to denote the subvector (a i,, a j ); if j < i, a j i is regarded as void Given a and A {,, }, we write a A to denote the subvector (a i : i A) e write a j,o to denote the subvector with odd indices (a k : k j; k odd) e write a j,e to denote the subvector with even indices (a k : k j; k even) For example, for a 5 (5,, 6,, ), we have a (, 6, ), a5,e (, ), a,o (5, 6) The notation 0 is used to denote the all-zero vector Code constructions in this paper will be carried out in vector spaces over the binary field GF() Unless specified otherwise, all vectors, matrices, and operations on them will be over GF() In particular, for a, b vectors over GF(), we write a b to denote their componentwise mod- sum The Kronecker product of an m-by-n matrix A [A ij ] and an r-by-s matrix B [B ij ] is defined as A B A B A n B A m B A mn B, which is an mr-by-ns matrix The Kronecker power A n is defined as A A (n ) for all n e will follow the convention that A 0 [] e write A to denote the number of elements in a set A e write A to denote the indicator function of a set A; thus, A (x) equals if x A and 0 otherwise e use the standard Landau notation O(), o(), ω() to denote the asymptotic behavior of functions B Channel polarization Channel polarization is an operation by which one manufactures out of independent copies of a given B-DMC a second set of channels { (i) : i } that show a polarization effect in the sense that, as becomes large, the symmetric capacity terms {I( (i) )} tend towards 0 or for all but a vanishing fraction of indices i This operation consists of a channel combining phase and a channel splitting phase ) Channel combining: This phase combines copies of a given B-DMC in a recursive manner to produce a vector channel : X Y, where can be any power of two, n, n 0 The recursion begins at the 0-th level (n 0) with only one copy of and we set The first level (n ) of the recursion combines two independent copies of as shown in Fig and obtains the channel : X Y with the transition probabilities (y, y u, u ) (y u u )(y u ) (3) u u Fig The channel x x The next level of the recursion is shown in Fig where two independent copies of are combined to create the channel : X Y with transition probabilities (y u ) (y u u, u 3 u ) (y 3 u, u ) u u u 3 u R v v v 3 v x x 3 y y x x Fig The channel and its relation to and In Fig, R is the permutation operation that maps an input (s, s, s 3, s ) to v (s, s 3, s, s ) The mapping u x from the input of to the input of can be written as x u G with G [ y y y 3 y ] Thus, we have the relation (y u ) (y u G ) between the transition probabilities of and those of The general form of the recursion is shown in Fig 3 where two independent copies of / are combined to produce the channel The input vector u to is first transformed into s so that s i u i u i and s i u i for i / The operator R in the figure is a permutation, known as the reverse shuffle operation, and acts on its input s to produce v (s, s 3,, s, s, s,, s ), which becomes the input to the two copies of / as shown in the figure e observe that the mapping u v is linear over GF() It follows by induction that the overall mapping u x, from the input of the synthesized channel to the input of the underlying raw channels, is also linear and may be represented by a matrix G so that x u G e call G

3 3 u s v y u s v y u / s / v / / y / u / s / v / y / 3) Channel polarization: Theorem : For any B-DMC, the channels { (i) } polarize in the sense that, for any fixed δ (0, ), as goes to infinity through powers of two, the fraction of indices i {,, } for which I( (i) ) ( δ, ] goes to I() and the fraction for which I( (i) ) [0, δ) goes to I() This theorem is proved in Sect IV 09 u / s / R v / / y / u / s / v / y / u s v y Symmetric capacity u s v y Fig 3 Recursive construction of from two copies of / the generator matrix of size The transition probabilities of the two channels and are related by (y u ) (y u G ) () for all y Y, u X e will show in Sect VII that G equals B F n for any n, n 0, where B is a permutation matrix known as bit-reversal and F [ 0 ] ote that the channel combining operation is fully specified by the matrix F Also note that G and F n have the same set of rows, but in a different (bit-reversed) order; we will discuss this topic more fully in Sect VII ) Channel splitting: Having synthesized the vector channel out of, the next step of channel polarization is to split back into a set of binary-input coordinate channels (i) : X Y X i, i, defined by the transition probabilities (i) (y, u i u i ) (y u ), (5) u i X i where (y, u i ) denotes the output of (i) and u i its input To gain an intuitive understanding of the channels { (i) }, consider a genie-aided successive cancellation decoder in which the ith decision element estimates u i after observing y and the past channel inputs u i (supplied correctly by the genie regardless of any decision errors at earlier stages) If u is a-priori uniform on X, then (i) is the effective channel seen by the ith decision element in this scenario Fig Channel index Plot of I( (i) ) vs i,, 0 for a BEC with ǫ 05 The polarization effect is illustrated in Fig for the case is a BEC with erasure probability ǫ 05 The numbers )} have been computed using the recursive relations {I( (i) I( (i ) ) I( (i) / ), I( (i) ) I( (i) / ) I( (i) / ), with I( () ) ǫ This recursion is valid only for BECs and it is proved in Sect III o efficient algorithm is known for calculation of {I( (i) )} for a general B-DMC Figure shows that I( (i) ) tends to be near 0 for small i and near for large i However, I( (i) ) shows an erratic behavior for an intermediate range of i For general B-DMCs, determining the subset of indices i for which I( (i) ) is above a given threshold is an important computational problem that will be addressed in Sect IX ) Rate of polarization: For proving coding theorems, the speed with which the polarization effect takes hold as a function of is important Our main result in this regard is given in terms of the parameters Z( (i) ) y Y u i X i (6) (i) (y, ui 0) (i) (y, ui ) Theorem : For any B-DMC with I() > 0, and any fixed R < I(), there exists a sequence of sets A {,, }, {,,, n, }, such that A R and Z( (i) ) O( 5/ ) for all i A This theorem is proved in Sect IV-B (7)

4 e stated the polarization result in Theorem in terms {Z( (i) (i) )} rather than {I( )} because this form is better suited to the coding results that we will develop A rate of polarization result in terms of {I( (i) )} can be obtained from Theorem with the help of Prop C Polar coding e take advantage of the polarization effect to construct codes that achieve the symmetric channel capacity I() by a method we call polar coding The basic idea of polar coding is to create a coding system where one can access individually and send data only through those for which Z( (i) ) is near 0 ) G -coset codes: e first describe a class of block codes that contain polar codes the codes of main interest as a special case The block-lengths for this class are restricted to powers of two, n for some n 0 For a given, each code in the class is encoded in the same manner, namely, each coordinate channel (i) x u G () where G is the generator matrix of order, defined above For A an arbitrary subset of {,,}, we may write () as x u AG (A) u A cg (A c ) (9) where G (A) denotes the submatrix of G formed by the rows with indices in A If we now fix A and u A c, but leave u A as a free variable, we obtain a mapping from source blocks u A to codeword blocks x This mapping is a coset code: it is a coset of the linear block code with generator matrix G (A), with the coset determined by the fixed vector u A cg (A c ) e will refer to this class of codes collectively as G -coset codes Individual G -coset codes will be identified by a parameter vector (, K, A, u A c), where K is the code dimension and specifies the size of A The ratio K/ is called the code rate e will refer to A as the information set and to u A c X K as frozen bits or vector For example, the (,, {, }, (, 0)) code has the encoder mapping x u G (u, u ) [ ] [ ] (, 0) (0) 0 0 For a source block (u, u ) (, ), the coded block is x (,, 0, ) Polar codes will be specified shortly by giving a particular rule for the selection of the information set A ) A successive cancellation decoder: Consider a G -coset code with parameter (, K, A, u A c) Let u be encoded into a codeword x, let x be sent over the channel, and let a channel output y be received The decoder s task is to generate an estimate û of u, given knowledge of A, u A c, and y Since the decoder can avoid errors in the frozen part e include the redundant parameter K in the parameter set because often we consider an ensemble of codes with K fixed and A free by setting û A c u A c, the real decoding task is to generate an estimate û A of u A The coding results in this paper will be given with respect to a specific successive cancellation (SC) decoder, unless some other decoder is mentioned Given any (, K, A, u A c) G - coset code, we will use a SC decoder that generates its decision û by computing û i { u i, if i A c h i (y, ûi ), if i A () in the order i from to, where h i : Y X i X, i A, are decision functions defined as h i (y, ûi ) 0, if (i) (y,ûi 0) (i) (y,ûi ) (), otherwise for all y Y, û i X i e will say that a decoder block error occurred if û u or equivalently if û A u A The decision functions {h i } defined above resemble ML decision functions but are not exactly so, because they treat the future frozen bits (u j : j > i, j A c ) as RVs, rather than as known bits In exchange for this suboptimality, {h i } can be computed efficiently using recursive formulas, as we will show in Sect II Apart from algorithmic efficiency, the recursive structure of the decision functions is important because it renders the performance analysis of the decoder tractable Fortunately, the loss in performance due to not using true ML decision functions happens to be negligible: I() is still achievable 3) Code performance: The notation P e (, K, A, u A c) will denote the probability of block error for a (, K, A, u A c) code, assuming that each data vector u A X K is sent with probability K and decoding is done by the above SC decoder More precisely, P e (, K, A, u A c) u A X K K y Y : û (y ) u (y u ) The average of P e (, K, A, u A c) over all choices for u A c will be denoted by P e (, K, A): P e (, K, A) u A c X K K P e(, K, A, u A c) A key bound on block error probability under SC decoding is the following Proposition : For any B-DMC and any choice of the parameters (, K, A), P e (, K, A) i A Z( (i) ) (3) Hence, for each (, K, A), there exists a frozen vector u A c such that P e (, K, A, u A c) i A Z( (i) ) () This is proved in Sect V-B This result suggests choosing A from among all K-subsets of {,,} so as to minimize

5 5 the RHS of (3) This idea leads to the definition of polar codes ) Polar codes: Given a B-DMC, a G -coset code with parameter (, K, A, u A c) will be called a polar code for if the information set A is chosen as a K-element subset of {,,} such that Z( (i) (j) ) Z( ) for all i A, j A c Polar codes are channel-specific designs: a polar code for one channel may not be a polar code for another The main result of this paper will be to show that polar coding achieves the symmetric capacity I() of any given B-DMC An alternative rule for polar code definition would be to specify A as a K-element subset of {,,} such that I( (i) ) I( (j) ) for all i A, j Ac This alternative rule would also achieve I() However, the rule based on the Bhattacharyya parameters has the advantage of being connected with an explicit bound on block error probability The polar code definition does not specify how the frozen vector u A c is to be chosen; it may be chosen at will This degree of freedom in the choice of u A c simplifies the performance analysis of polar codes by allowing averaging over an ensemble However, it is not for analytical convenience alone that we do not specify a precise rule for selecting u A c, but also because it appears that the code performance is relatively insensitive to that choice In fact, we prove in Sect VI-B that, for symmetric channels, any choice for u A c is as good as any other 5) Coding theorems: Fix a B-DMC and a number R 0 Let P e (, R) be defined as P e (, R, A) with A selected in accordance with the polar coding rule for Thus, P e (, R) is the probability of block error under SC decoding for polar coding over with block-length and rate R, averaged over all choices for the frozen bits u A c The main coding result of this paper is the following: Theorem 3: For any given B-DMC and fixed R < I(), block error probability for polar coding under successive cancellation decoding satisfies P e (, R) O( ) (5) This theorem follows as an easy corollary to Theorem and the bound (3), as we show in Sect V-B For symmetric channels, we have the following stronger version of Theorem 3 Theorem : For any symmetric B-DMC and any fixed R < I(), consider any sequence of G -coset codes (, K, A, u A c) with increasing to infinity, K R, A chosen in accordance with the polar coding rule for, and u A c fixed arbitrarily The block error probability under successive cancellation decoding satisfies P e (, K, A, u A c) O( ) (6) This is proved in Sect VI-B ote that for symmetric channels I() equals the Shannon capacity of 6) Complexity: An important issue about polar coding is the complexity of encoding, decoding, and code construction The recursive structure of the channel polarization construction leads to low-complexity encoding and decoding algorithms for the class of G -coset codes, and in particular, for polar codes Theorem 5: For the class of G -coset codes, the complexity of encoding and the complexity of successive cancellation decoding are both O( log ) as functions of code blocklength This theorem is proved in Sections VII and VIII otice that the complexity bounds in Theorem 5 are independent of the code rate and the way the frozen vector is chosen The bounds hold even at rates above I(), but clearly this has no practical significance As for code construction, we have found no low-complexity algorithms for constructing polar codes One exception is the case of a BEC for which we have a polar code construction algorithm with complexity O() e discuss the code construction problem further in Sect IX and suggest a lowcomplexity statistical algorithm for approximating the exact polar code construction D Relations to previous work This paper is an extension of work begun in [], where channel combining and splitting were used to show that improvements can be obtained in the sum cutoff rate for some specific DMCs However, no recursive method was suggested there to reach the ultimate limit of such improvements As the present work progressed, it became clear that polar coding had much in common with Reed-Muller (RM) coding [3], [] Indeed, recursive code construction and SC decoding, which are two essential ingredients of polar coding, appear to have been introduced into coding theory by RM codes According to one construction of RM codes, for any n, n 0, and 0 K, an RM code with blocklength and dimension K, denoted RM(, K), is defined as a linear code whose generator matrix G RM (, K) is obtained by deleting ( K) of the rows of F n so that none of the deleted rows has a larger Hamming weight (number of s in that row) than any of the remaining ] K rows For instance, G RM (, ) F [ and G RM (, ) [ 0 0 ] This construction brings out the similarities between RM codes and polar codes Since G and F n have the same set of rows (only in a different order) for any n, it is clear that RM codes belong to the class of G -coset codes For example, RM(, ) is the G -coset code with parameter (,, {, }, (0, 0)) So, RM coding and polar coding may be regarded as two alternative rules for selecting the information set A of a G -coset code of a given size (, K) Unlike polar coding, RM coding selects the information set in a channelindependent manner; it is not as fine-tuned to the channel polarization phenomenon as polar coding is e will show in Sect X that, at least for the class of BECs, the RM rule for information set selection leads to asymptotically unreliable codes under SC decoding So, polar coding goes beyond RM coding in a non-trivial manner by paying closer attention to channel polarization Another connection to existing work can be established by noting that polar codes are multi-level u u v codes, which are a class of codes originating from Plotkin s method for code combining [5] This connection is not surprising in

6 6 view of the fact that RM codes are also multi-level u u v codes [6, pp -5] However, unlike typical multi-level code constructions where one begins with specific small codes to build larger ones, in polar coding the multi-level code is obtained by expurgating rows of a full-order generator matrix, G, with respect to a channel-specific criterion The special structure of G ensures that, no matter how expurgation is done, the resulting code is a multi-level u u v code In essence, polar coding enjoys the freedom to pick a multi-level code from an ensemble of such codes so as to suit the channel at hand, while conventional approaches to multi-level coding do not have this degree of flexibility Finally, we wish to mention a spectral interpretation of polar codes which is similar to Blahut s treatment of BCH codes [7, Ch 9]; this type of similarity has already been pointed out by Forney [, Ch ] in connection with RM codes From the spectral viewpoint, the encoding operation () is regarded as a transform of a frequency domain information vector u to a time domain codeword vector x The transform is invertible with G G The decoding operation is regarded as a spectral estimation problem in which one is given a time domain observation y, which is a noisy version of x, and asked to estimate u To aid the estimation task, one is allowed to freeze a certain number of spectral components of u This spectral interpretation of polar coding suggests that it may be possible to treat polar codes and BCH codes in a unified framework The spectral interpretation also opens the door to the use of various signal processing techniques in polar coding; indeed, in Sect VII, we exploit some fast transform techniques in designing encoders for polar codes E Paper outline The rest of the paper is organized as follows Sect II explores the recursive properties of the channel splitting operation In Sect III, we focus on how I() and Z() get transformed through a single step of channel combining and splitting e extend this to an asymptotic analysis in Sect IV and complete the proofs of Theorem and Theorem This completes the part of the paper on channel polarization; the rest of the paper is mainly about polar coding Section V develops an upper bound on the block error probability of polar coding under SC decoding and proves Theorem 3 Sect VI considers polar coding for symmetric B-DMCs and proves Theorem Sect VII gives an analysis of the encoder mapping G, which results in efficient encoder implementations In Sect VIII, we give an implementation of SC decoding with complexity O( log ) In Sect IX, we discuss the code construction complexity and propose an O( log ) statistical algorithm for approximate code construction In Sect X, we explain why RM codes have a poor asymptotic performance under SC decoding In Sect XI, we point out some generalizations of the present work, give some complementary remarks, and state some open problems II RECURSIVE CHAEL TRASFORMATIOS e have defined a blockwise channel combining and splitting operation by () and (5) which transformed independent copies of into (),, () The goal in this section is to show that this blockwise channel transformation can be broken recursively into single-step channel transformations e say that a pair of binary-input channels : X Ỹ and : X Ỹ X are obtained by a single-step transformation of two independent copies of a binary-input channel : X Y and write (, ) (, ) iff there exists a one-to-one mapping f : Y Ỹ such that (f(y, y ) u ) u (y u u )(y u ), (7) (f(y, y ), u u ) (y u u )(y u ) () for all u, u X, y, y Y According to this, we can write (, ) ( (), () ) for any given B-DMC because () (y u ) u (y u ) () (y, u u ) (y u ) u (y u u )(y u ), (9) (y u u )(y u ), (0) which are in the form of (7) and () by taking f as the identity mapping It turns out we can write, more generally,, (i) (i ) ) (, (i) ) () ( (i) This follows as a corollary to the following: Proposition 3: For any n 0, n, i, and (i ) (y, ui u i ) (i) (y, u,o i u,e i u i u i ) u i (i) (y, ui,e u i) () (i) (y, ui u i ) (i) (y, u,o i u,e i u i u i ) (i) (y, ui,e u i) (3) This proposition is proved in the Appendix The transform relationship () can now be justified by noting that () and (3) are identical in form to (7) and (), respectively, after the following substitutions: (i), (i ), (i), u u i, u u i, y (y, ui,o u,e i ), y (y, u,e i ), f(y, y ) (y, u i )

7 7 Fig 5 () (5) (3) (7) () (6) () () () (3) () () () (3) () () () () () () () () () () The channel transformation process with channels ( (), () Thus, we have shown that the blockwise channel transformation from to ( () (),, ) breaks at a local level into single-step channel transformations of the form () The full set of such transformations form a fabric as shown in Fig 5 for Reading from right to left, the figure starts with four copies of the transformation (, ) ) and continues in butterfly patterns, each representing a channel transformation of the form ( (j), (j) i ) i ( (j ), (j) i ) The two channels at the right end-points i of the butterflies are always identical and independent At the rightmost level there are independent copies of ; at the next level to the left, there are independent copies of () and () each; and so on Each step to the left doubles the number of channel types, but halves the number of independent copies III TRASFORMATIO OF RATE AD RELIABILITY e now investigate how the rate and reliability parameters, I( (i) (i) ) and Z( ), change through a local (single-step) transformation () By understanding the local behavior, we will be able to reach conclusions about the overall transformation from to ( (),, () ) Proofs of the results in this section are given in the Appendix A Local transformation of rate and reliability Proposition : Suppose (, ) (, ) for some set of binary-input channels Then, I( ) I( ) I(), () with equality iff I() equals 0 or I( ) I( ) (5) The equality () indicates that the single-step channel transform preserves the symmetric capacity The inequality (5) together with () implies that the symmetric capacity remains unchanged under a single-step transform, I( ) I( ) I(), iff is either a perfect channel or a completely noisy one If is neither perfect nor completely noisy, the single-step transform moves the symmetric capacity away from the center in the sense that I( ) < I() < I( ), thus helping polarization Proposition 5: Suppose (, ) (, ) for some set of binary-input channels Then, Z( ) Z(), (6) Z( ) Z() Z(), (7) Z( ) Z() Z( ) () Equality holds in (7) iff is a BEC e have Z( ) Z( ) iff Z() equals 0 or, or equivalently, iff I() equals or 0 This result shows that reliability can only improve under a single-step channel transform in the sense that Z( ) Z( ) Z() (9) with equality iff is a BEC Since the BEC plays a special role wrt extremal behavior of reliability, it deserves special attention Proposition 6: Consider the channel transformation (, ) (, ) If is a BEC with some erasure probability ǫ, then the channels and are BECs with erasure probabilities ǫ ǫ and ǫ, respectively Conversely, if or is a BEC, then is BEC B Rate and reliability for (i) e now return to the context at the end of Sect II Proposition 7: For any B-DMC, n, n 0, i, the transformation ( (i), (i) (i ) ) (, (i) ) is rate-preserving and reliability-improving in the sense that I( (i ) Z( (i ) ) I( (i) ) Z( (i) (i) ) I( ), (30) (i) ) Z( ), (3) with equality in (3) iff is a BEC Channel splitting moves the rate and reliability away from the center in the sense that I( (i ) Z( (i ) ) I( (i) ) Z( (i) (i) ) I( ), (3) (i) ) Z( ), (33) with equality in (3) and (33) iff I() equals 0 or The reliability terms further satisfy ) Z( (i) (i) ) Z( ), (3) Z( (i) (i) ) Z( ), (35) Z( (i ) with equality in (3) iff is a BEC The cumulative rate and reliability satisfy I( (i) ) I(), (36) i i Z( (i) ) Z(), (37)

8 with equality in (37) iff is a BEC This result follows from Prop and Prop 5 as a special case and no separate proof is needed The cumulative relations (36) and (37) follow by repeated application of (30) and (3), respectively The conditions for equality in Prop are stated in terms of rather than (i) ; this is possible because: (i) by Prop, I() {0, } iff I( (i) ) {0, }; and (ii) is a BEC iff (i) is a BEC, which follows from Prop 6 by induction For the special case that is a BEC with an erasure probability ǫ, it follows from Prop and Prop 6 that the parameters {Z( (i) )} can be computed through the recursion Z( (j ) ) Z( (j) (j) / ) Z( / ), Z( (j) ) Z( (j) / ), (3) with Z( () ) ǫ The parameter Z( (i) ) equals the erasure probability of the channel (i) The recursive relations (6) follow from (3) by the fact that I( (i) (i) ) Z( ) for a BEC 0 () 000 () 00 () 00 () 0 (3) 00 () 0 () 0 (5) 00 (3) 0 (6) 0 () (7) 0 () () IV CHAEL POLARIZATIO e prove the main results on channel polarization in this section The analysis is based on the recursive relationships depicted in Fig 5; however, it will be more convenient to resketch Fig 5 as a binary tree as shown in Fig 6 The root node of the tree is associated with the channel The root gives birth to an upper channel () and a lower channel (), which are associated with the two nodes at level The channel () in turn gives birth to the channels () and (), and so on The channel (i) n is located at level n of the tree at node number i counting from the top There is a natural indexing of nodes of the tree in Fig 6 by bit sequences The root node is indexed with the null sequence The upper node at level is indexed with 0 and the lower node with Given a node at level n with index b b b n, the upper node emanating from it has the label b b b n 0 and the lower node b b b n According to this labeling, the channel (i) is situated at the node b b n b n with i n j b j n j e denote the channel (i) n located at node b b b n alternatively as bb n e define a random tree process, denoted {K n ; n 0}, in connection with Fig 6 The process begins at the root of the tree with K 0 For any n 0, given that K n b b n, K n equals b b n0 or b b n with probability / each Thus, the path taken by {K n } through the channel tree may be thought of as being driven by a sequence of iid Bernoulli RVs {B n ; n,, } where B n equals 0 or with equal probability Given that B,, B n has taken on a sample value b,, b n, the random channel process takes the value K n b b n In order to keep track of the rate and reliability parameters of the random sequence of channels K n, we define the random processes I n I(K n ) and Z n Z(K n ) For a more precise formulation of the problem, we consider the probability space (Ω, F, P) where Ω is the space of all Fig 6 The tree process for the recursive channel construction binary sequences (b, b, ) {0, }, F is the Borel field (BF) generated by the cylinder sets S(b,, b n ) {ω Ω : ω b,,ω n b n }, n, b,, b n {0, }, and P is the probability measure defined on F such that P(S(b,,b n )) / n For each n, we define F n as the BF generated by the cylinder sets S(b,,b i ), i n, b,,b i {0, } e define F 0 as the trivial BF consisting of the null set and Ω only Clearly, F 0 F F The random processes described above can now be formally defined as follows For ω (ω, ω, ) Ω and n, define B n (ω) ω n, K n (ω) ω ω n, I n (ω) I(K n (ω)), and Z n (ω) Z(K n (ω)) For n 0, define K 0, I 0 I(), Z 0 Z() It is clear that, for any fixed n 0, the RVs B n, K n, I n, and Z n are measurable with respect to the BF F n A Proof of Theorem e will prove Theorem by considering the stochastic convergence properties of the random sequences {I n } and {Z n } Proposition : The sequence of random variables and Borel fields {I n, F n ; n 0} is a martingale, ie, F n F n and I n is F n -measurable, (39) E[ I n ] <, (0) I n E[I n F n ] () Furthermore, the sequence {I n ; n 0} converges ae to a random variable I such that E[I ] I 0 Proof: Condition (39) is true by construction and (0) by the fact that 0 I n To prove (), consider a cylinder

9 9 set S(b,, b n ) F n and use Prop 7 to write E[I n S(b,, b n )] I( b b n0) I( b b n) I( b b n ) Since I( b b n ) is the value of I n on S(b,,b n ), () follows This completes the proof that {I n, F n } is a martingale Since {I n, F n } is a uniformly integrable martingale, by general convergence results about such martingales (see, eg, [9, Theorem 96]), the claim about I follows It should not be surprising that the limit RV I takes values ae in {0, }, which is the set of fixed points of I() under the transformation (, ) ( (), () ), as determined by the condition for equality in (5) For a rigorous proof of this statement, we take an indirect approach and bring the process {Z n ; n 0} also into the picture Proposition 9: The sequence of random variables and Borel fields {Z n, F n ; n 0} is a supermartingale, ie, F n F n and Z n is F n -measurable, () E[ Z n ] <, (3) Z n E[Z n F n ] () Furthermore, the sequence {Z n ; n 0} converges ae to a random variable Z which takes values ae in {0, } Proof: Conditions () and (3) are clearly satisfied To verify (), consider a cylinder set S(b,, b n ) F n and use Prop 7 to write E[Z n S(b,, b n )] Z( b b n0) Z( b b n) Z( b b n ) Since Z( b b n ) is the value of Z n on S(b,, b n ), () follows This completes the proof that {Z n, F n } is a supermartingale For the second claim, observe that the supermartingale {Z n, F n } is uniformly integrable; hence, it converges ae and in L to a RV Z such that E[ Z n Z ] 0 (see, eg, [9, Theorem 95]) It follows that E[ Z n Z n ] 0 But, by Prop 7, Z n Zn with probability /; hence, E[ Z n Z n ] (/)E[Z n ( Z n )] 0 Thus, E[Z n ( Z n )] 0, which implies E[Z ( Z )] 0 This, in turn, means that Z equals 0 or ae Proposition 0: The limit RV I takes values ae in the set {0, }: P(I ) I 0 and P(I 0) I 0 Proof: The fact that Z equals 0 or ae, combined with Prop, implies that I Z ae Since E[I ] I 0, the rest of the claim follows As a corollary to Prop 0, we can conclude that, as tends to infinity, the symmetric capacity terms {I( (i) : i } cluster around 0 and, except for a vanishing fraction This completes the proof of Theorem It is interesting that the above discussion gives a new interpretation to I 0 I() as the probability that the random process {Z n ; n 0} converges to zero e may use this to strengthen the lower bound in () (This stronger form is given as a side result and will not be used in the sequel) Proposition : For any B-DMC, we have I() Z() with equality iff is a BEC This result can be interpreted as saying that, among all B- DMCs, the BEC presents the most favorable rate-reliability trade-off: it minimizes Z() (maximizes reliability) among all channels with a given symmetric capacity I(); equivalently, it minimizes I() required to achieve a given level of reliability Z() Proof: Consider two channels and with Z() Z( ) z 0 Suppose that is a BEC Then, has erasure probability z 0 and I( ) z 0 Consider the random processes {Z n } and {Z n } corresponding to and, respectively By the condition for equality in (3), the process {Z n } is stochastically dominated by {Z n} in the sense that P(Z n z) P(Z n z) for all n, 0 z Thus, the probability of {Z n } converging to zero is lowerbounded by the probability that {Z n} converges to zero, ie, I() I( ) This implies I() Z() B Proof of Theorem e will now prove Theorem, which strengthens the above polarization results by specifying a rate of polarization Consider the probability space (Ω, F, P) For ω Ω, i 0, by Prop 7, we have Z i (ω) Z i (ω) if B i(ω) and Z i (ω) Z i (ω) Z i (ω) Z i (ω) if B i (ω) 0 For ζ 0 and m 0, define T m (ζ) {ω Ω : Z i (ω) ζ for all i m} For ω T m (ζ) and i m, we have which implies Z i (ω) Z i (ω) Z n (ω) ζ n m n im {, if B i (ω) 0 ζ, if B i (ω) (ζ/) Bi(ω), ω T m (ζ), n > m For n > m 0 and 0 < η < /, define n U m,n (η) {ω Ω : B i (ω) > (/ η)(n m)} im Then, we have [ Z n (ω) ζ η ζ η] n m, ω Tm (ζ) U m,n (η); from which, by putting ζ 0 and η 0 /0, we obtain Z n (ω) 5(n m)/, ω T m (ζ 0 ) U m,n (η 0 ) (5) ow, we show that (5) occurs with sufficiently high probability First, we use the following result, which is proved in the Appendix Lemma : For any fixed ζ > 0, δ > 0, there exists a finite integer m 0 (ζ, δ) such that P [T m0 (ζ)] I 0 δ/ Second, we use Chernoff s bound [0, p 53] to write P [U m,n (η)] (n m)[ H(/ η)] (6)

10 0 where H is the binary entropy function Define n 0 (m, η, δ) as the smallest n such that the RHS of (6) is greater than or equal to δ/; it is clear that n 0 (m, η, δ) is finite for any m 0, 0 < η < /, and δ > 0 ow, with m m (δ) m 0 (ζ 0, δ) and n n (δ) n 0 (m, η 0, δ), we obtain the desired bound: P[T m (ζ 0 ) U m,n(η 0 )] I 0 δ, n n Finally, we tie the above analysis to the claim of Theorem Define c 5m/ and and, note that V n {ω Ω : Zn (ω) c 5n/ }, n 0; T m (ζ 0 ) U m,n(η 0 ) V n, n n So, P(V n ) I 0 δ for n n On the other hand, P(V n ) n{z( ω n) c 5n/ } ω n X n A where A {i {,,} : Z( (i) ) c 5/ } with n e conclude that A (I 0 δ) for n n (δ) This completes the proof of Theorem Given Theorem, it is an easy exercise to show that polar coding can achieve rates approaching I(), as we will show in the next section It is clear from the above proof that Theorem gives only an ad-hoc result on the asymptotic rate of channel polarization; this result is sufficient for proving a capacity theorem for polar coding; however, finding the exact asymptotic rate of polarization remains an important goal for future research V PERFORMACE OF POLAR CODIG e show in this section that polar coding can achieve the symmetric capacity I() of any B-DMC The main technical task will be to prove Prop e will carry out the analysis over the class of G -coset codes before specializing the discussion to polar codes Recall that individual G -coset codes are identified by a parameter vector (, K, A, u A c) In the analysis, we will fix the parameters (, K, A) while keeping u A c free to take any value over X K In other words, the analysis will be over the ensemble of K G - coset codes with a fixed (, K, A) The decoder in the system will be the SC decoder described in Sect I-C A A probabilistic setting for the analysis Let (X Y, P) be a probability space with the probability assignment P({(u, y )}) (y u ) (7) for all (u, y ) X Y On this probability space, we define an ensemble of random vectors (U, X, Y, Û ) that represent, respectively, the input to the synthetic channel A recent result in this direction is discussed in Sect XI-A, the input to the product-form channel, the output of (and also of ), and the decisions by the decoder For each sample point (u, y ) X Y, the first three vectors take on the values U (u, y ) u, X (u, y ) u G, and Y (u, y ) y, while the decoder output takes on the value Û (u, y ) whose coordinates are defined recursively as Û i (u, y ) { u i, i A c h i (y, Ûi (u, y )), i A () for i,, A realization u X for the input random vector U corresponds to sending the data vector u A together with the frozen vector u A c As random vectors, the data part U A and the frozen part U A c are uniformly distributed over their respective ranges and statistically independent By treating U A c as a random vector over X K, we obtain a convenient method for analyzing code performance averaged over all codes in the ensemble (, K, A) The main event of interest in the following analysis is the block error event under SC decoding, defined as E {(u, y ) X Y : ÛA(u, y ) u A } (9) Since the decoder never makes an error on the frozen part of U, ie, Û A c equals U A c with probability one, that part has been excluded from the definition of the block error event The probability of error terms P e (, K, A) and P e (, K, A, u A c) that were defined in Sect I-C3 can be expressed in this probability space as P e (, K, A) P(E), P e (, K, A, u A c) P(E {U A c u A c}), (50) where {U A c u A c} denotes the event {(ũ, y ) X Y : ũ A c u A c} B Proof of Proposition e may express the block error event as E i A B i where B i {(u, y ) X Y : u i Û i (u, y ), u i Ûi(u, y )} (5) is the event that the first decision error in SC decoding occurs at stage i e notice that B i {(u, y ) X Y : u i Ûi (u, y ), u i h i (y, Ûi (u, y ))} {(u, y ) X Y : u i Ûi (u, y ), u i h i (y, u i )} {(u, y ) X Y : u i h i (y, u i )} E i where E i {(u, y ) X Y : (i ) (y, u i u i ) (i ) (y, ui u i )} (5)

11 Thus, we have E i A E i, P(E) i A P(E i ) For an upper bound on P(E i ), note that P(E i ) (y u ) Ei (u, y ) u,y u,y (y u ) (i) (y, ui u i ) (i) (y, ui u i ) Z( (i) ) (53) e conclude that P(E) i A Z( (i) ), which is equivalent to (3) This completes the proof of Prop The main coding theorem of the paper now follows readily C Proof of Theorem 3 By Theorem, for any given rate R < I(), there exists a sequence of information sets A with size A R such that Z( (i) ) max {Z( (i) i A )} O( ) (5) i A In particular, the bound (5) holds if A is chosen in accordance with the polar coding rule because by definition this rule minimizes the sum in (5) Combining this fact about the polar coding rule with Prop, Theorem 3 follows D A numerical example Although we have established that polar codes achieve the symmetric capacity, the proofs have been of an asymptotic nature and the exact asymptotic rate of polarization has not been found It is of interest to understand how quickly the polarization effect takes hold and what performance can be expected of polar codes under SC decoding in the nonasymptotic regime To investigate these, we give here a numerical study Let be a BEC with erasure probability / Figure 7 shows the rate vs reliability trade-off for using polar codes with block-lengths { 0, 5, 0 } This figure is obtained by using codes whose information sets are of the form A(η) {i {,,} : Z( (i) ) < η}, where 0 η is a variable threshold parameter There are two sets of three curves in the plot The solid lines are plots of R(η) A(η) / vs B(η) (i) i A(η) Z( ) The dashed lines are plots of R(η) vs L(η) max i A(η) {Z( (i) )} The parameter η is varied over a subset of [0, ] to obtain the curves The parameter R(η) corresponds to the code rate The significance of B(η) is also clear: it is an upper-bound on P e (η), the probability of block-error for polar coding at rate Bounds on probability of block error Rate (bits) Fig 7 Rate vs reliability for polar coding and SC decoding at block-lengths 0, 5, and 0 on a BEC with erasure probability / R(η) under SC decoding The parameter L(η) is intended to serve as a lower bound to P e (η) This example provides empirical evidence that polar coding achieves channel capacity as the block-length is increased a fact already established theoretically More significantly, the example also shows that the rate of polarization is too slow to make near-capacity polar coding under SC decoding feasible in practice 5 VI SYMMETRIC CHAELS The main goal of this section is to prove Theorem, which is a strengthened version of Theorem 3 for symmetric channels A Symmetry under channel combining and splitting Let : X Y be a symmetric B-DMC with X {0, } and Y arbitrary By definition, there exists a a permutation π on Y such that (i) π π and (ii) (y ) (π (y) 0) for all y Y Let π 0 be the identity permutation on Y Clearly, the permutations (π 0, π ) form an abelian group under function composition For a compact notation, we will write x y to denote π x (y), for x X, y Y Observe that (y x a) (a y x) for all a, x X, y Y This can be verified by exhaustive study of possible cases or by noting that (y x a) ((x a) y 0) (x (a y) 0) (a y x) Also observe that (y x a) (x y a) as is a commutative operation on X For x X, y Y, let x y 0 (x y,, x y ) (55) This associates to each element of X a permutation on Y Proposition : If a B-DMC is symmetric, then is also symmetric in the sense that (y x a ) (x y a ) (56) for all x, a X, y Y The proof is immediate and omitted

12 Proposition 3: If a B-DMC is symmetric, then the channels and (i) are also symmetric in the sense that (y u ) (a G y u a ), (57) (i) (y, ui u i ) (i) (a G y, u i a i u i a i ) (5) for all u, a X, y Y, n, n 0, i Proof: Let x u G and observe that (y u ) i (y i x i ) i (x i y i 0) (x y 0 ) ow, let b a G, and use the same reasoning to see that (b y u a ) ((x b ) (b y ) 0 ) (x y 0 ) This proves the first claim To prove the second claim, we use the first result (i) (y, u i u i ) u i u i (y u ) (a G y u a ) (a G y, ui a i u i a i ) where we used the fact that the sum over u i X i can be replaced with a sum over u i a i for any fixed a since {u i a i : u i X i } X i B Proof of Theorem e return to the analysis in Sect V and consider a code ensemble (, K, A) under SC decoding, only this time assuming that is a symmetric channel e first show that the error events {E i } defined by (5) have a symmetry property Proposition : For a symmetric B-DMC, the event E i has the property that (u, y ) E i iff (a u, a G y ) E i (59) for each i, (u, y ) X Y, a X Proof: This follows directly from the definition of E i by using the symmetry property (5) of the channel (i) ow, consider the transmission of a particular source vector u A and a frozen vector u A c, jointly forming an input vector u for the channel This event is denoted below as {U u } instead of the more formal {u } Y Corollary : For a symmetric B-DMC, for each i and u X, the events E i and {U u } are independent; hence, P(E i ) P(E i {U u }) Proof: For (u, y ) X Y and x u G, we have P(E i {U u }) y (y u ) E i (u, y ) (x y 0 ) E i (0, x y ) y (60) P(E i {U 0 }) (6) Equality follows in (60) from (57) and (59) by taking a u, and in (6) from the fact that {x y : y Y } Y for any fixed x X The rest of the proof is immediate ow, by (53), we have, for all u X, P(E i {U u }) Z( (i) ) (6) and, since E i A E i, we obtain P(E {U u }) i A Z( (i) ) (63) This implies that, for every symmetric B-DMC and every (, K, A, u A c) code, P e (, K, A, u A c) u A X K K P(E {U u }) i A Z( (i) ) (6) This bound on P e (, K, A, u A c) is independent of the frozen vector u A c Theorem is now obtained by combining Theorem with Prop, as in the proof of Theorem 3 ote that although we have given a bound on P(E {U u }) that is independent of u, we stopped short of claiming that the error event E is independent of U because our decision functions {h i } break ties always in favor of û i 0 If this bias were removed by randomization, then E would become independent of U C Further symmetries of the channel (i) e may use the degrees of freedom in the choice of a in (5) to explore the symmetries inherent in the channel (i) For a given (y, u i ), we may select a with a i u i to obtain (i) (y, u i u i ) (i) (a G y, 0 i 0) (65) So, if we were to prepare a look-up table for the transition probabilities { (i) (y, ui u i ) : y Y, u i X i }, it would suffice to store only the subset of probabilities { (i) (y, 0 i 0) : y Y } The size of the look-up table can be reduced further by using the remaining degrees of freedom in the choice of a i Let Xi {a X : a i 0i }, i Then, for any i, a Xi, and y Y, we have (i) (y, 0i 0) (i) (a G y, 0i 0) (66) which follows from (65) by taking u i 0 i on the left hand side To explore this symmetry further, let Xi y {a G y : a Xi } The set Xi y is the orbit of y under the action group Xi The orbits Xi y over variation of y partition the space Y into equivalence classes Let Yi be a set formed by taking one representative from each equivalence class The output alphabet of the channel (i) can be represented effectively by the set Yi For example, suppose is a BSC with Y {0, } Each orbit Xi y has i elements and there are i orbits In particular, the channel () has effectively two outputs, and being symmetric, it has to be a BSC This is a great

13 3 simplification since () has an apparent output alphabet size of Likewise, while (i) has an apparent output alphabet size of i, due to symmetry, the size shrinks to i Further output alphabet size reductions may be possible by exploiting other properties specific to certain B-DMCs For example, if is a BEC, the channels { (i) } are known to be BECs, each with an effective output alphabet size of three The symmetry properties of { (i) } help simplify the computation of the channel parameters Proposition 5: For any symmetric B-DMC, the parameters {Z( (i) )} given by (7) can be calculated by the simplified formula Z( (i) ) i y Y i X i y (i) (y, 0i 0) (i) (y, 0i ) e omit the proof of this result For the important example of a BSC, this formula becomes Z( (i) ) (i) (y, 0i 0) (i) (y, 0i ) u u u / u / u / u R u v y u 3 v y u / u u u / v / y / v / v / v / y / y / y y Y i This sum for Z( (i) ) has i terms, as compared to i terms in (7) Fig An alternative realization of the recursive construction for VII ECODIG In this section, we will consider the encoding of polar codes and prove the part of Theorem 5 about encoding complexity e begin by giving explicit algebraic expressions for G, the generator matrix for polar coding, which so far has been defined only in a schematic form by Fig 3 The algebraic forms of G naturally point at efficient implementations of the encoding operation x u G In analyzing the encoding operation G, we exploit its relation to fast transform methods in signal processing; in particular, we use the bit-indexing idea of [] to interpret the various permutation operations that are part of G A Formulas for G In the following, assume n for some n 0 Let I k denote the k-dimensional identity matrix for any k e begin by translating the recursive definition of G as given by Fig 3 into an algebraic form: G (I / F)R (I G / ), for, with G I Either by verifying algebraically that (I / F)R R (F I / ) or by observing that channel combining operation in Fig 3 can be redrawn equivalently as in Fig, we obtain a second recursive formula G R (F I / )(I G / ) R (F G / ), (67) valid for This form appears more suitable to derive a recursive relationship e substitute G / R / (F G / ) back into (67) to obtain G R ( F ( R/ ( F G/ ))) R ( I R / ) ( F G / ) (6) where (6) is obtained by using the identity (AC) (BD) (A B)(C D) with A I, B R /, C F, D F G / Repeating this, we obtain G B F n (69) where B R (I R / )(I R / ) (I / R ) It can seen by simple manipulations that B R (I B / ) (70) e can see that B is a permutation matrix by the following induction argument Assume that B / is a permutation matrix for some ; this is true for since B I Then, B is a permutation matrix because it is the product of two permutation matrices, R and I B / In the following, we will say more about the nature of B as a permutation B Analysis by bit-indexing To analyze the encoding operation further, it will be convenient to index vectors and matrices with bit sequences Given a vector a with length n for some n 0, we denote

Polarization codes and the rate of polarization

Polarization codes and the rate of polarization Polarization codes and the rate of polarization Erdal Arıkan, Emre Telatar Bilkent U., EPFL Sept 10, 2008 Channel Polarization Given a binary input DMC W, i.i.d. uniformly distributed inputs (X 1,...,

More information

Linear Codes. Chapter 3. 3.1 Basics

Linear Codes. Chapter 3. 3.1 Basics Chapter 3 Linear Codes In order to define codes that we can encode and decode efficiently, we add more structure to the codespace. We shall be mainly interested in linear codes. A linear code of length

More information

I. GROUPS: BASIC DEFINITIONS AND EXAMPLES

I. GROUPS: BASIC DEFINITIONS AND EXAMPLES I GROUPS: BASIC DEFINITIONS AND EXAMPLES Definition 1: An operation on a set G is a function : G G G Definition 2: A group is a set G which is equipped with an operation and a special element e G, called

More information

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS Systems of Equations and Matrices Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a

More information

Lecture 13 - Basic Number Theory.

Lecture 13 - Basic Number Theory. Lecture 13 - Basic Number Theory. Boaz Barak March 22, 2010 Divisibility and primes Unless mentioned otherwise throughout this lecture all numbers are non-negative integers. We say that A divides B, denoted

More information

INDISTINGUISHABILITY OF ABSOLUTELY CONTINUOUS AND SINGULAR DISTRIBUTIONS

INDISTINGUISHABILITY OF ABSOLUTELY CONTINUOUS AND SINGULAR DISTRIBUTIONS INDISTINGUISHABILITY OF ABSOLUTELY CONTINUOUS AND SINGULAR DISTRIBUTIONS STEVEN P. LALLEY AND ANDREW NOBEL Abstract. It is shown that there are no consistent decision rules for the hypothesis testing problem

More information

Quotient Rings and Field Extensions

Quotient Rings and Field Extensions Chapter 5 Quotient Rings and Field Extensions In this chapter we describe a method for producing field extension of a given field. If F is a field, then a field extension is a field K that contains F.

More information

Lecture 3: Finding integer solutions to systems of linear equations

Lecture 3: Finding integer solutions to systems of linear equations Lecture 3: Finding integer solutions to systems of linear equations Algorithmic Number Theory (Fall 2014) Rutgers University Swastik Kopparty Scribe: Abhishek Bhrushundi 1 Overview The goal of this lecture

More information

Notes on Determinant

Notes on Determinant ENGG2012B Advanced Engineering Mathematics Notes on Determinant Lecturer: Kenneth Shum Lecture 9-18/02/2013 The determinant of a system of linear equations determines whether the solution is unique, without

More information

Continued Fractions and the Euclidean Algorithm

Continued Fractions and the Euclidean Algorithm Continued Fractions and the Euclidean Algorithm Lecture notes prepared for MATH 326, Spring 997 Department of Mathematics and Statistics University at Albany William F Hammond Table of Contents Introduction

More information

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS 1. SYSTEMS OF EQUATIONS AND MATRICES 1.1. Representation of a linear system. The general system of m equations in n unknowns can be written a 11 x 1 + a 12 x 2 +

More information

How To Prove The Dirichlet Unit Theorem

How To Prove The Dirichlet Unit Theorem Chapter 6 The Dirichlet Unit Theorem As usual, we will be working in the ring B of algebraic integers of a number field L. Two factorizations of an element of B are regarded as essentially the same if

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES Contents 1. Random variables and measurable functions 2. Cumulative distribution functions 3. Discrete

More information

This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.

This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination. IEEE/ACM TRANSACTIONS ON NETWORKING 1 A Greedy Link Scheduler for Wireless Networks With Gaussian Multiple-Access and Broadcast Channels Arun Sridharan, Student Member, IEEE, C Emre Koksal, Member, IEEE,

More information

A Practical Scheme for Wireless Network Operation

A Practical Scheme for Wireless Network Operation A Practical Scheme for Wireless Network Operation Radhika Gowaikar, Amir F. Dana, Babak Hassibi, Michelle Effros June 21, 2004 Abstract In many problems in wireline networks, it is known that achieving

More information

THIS paper deals with a situation where a communication

THIS paper deals with a situation where a communication IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 44, NO. 3, MAY 1998 973 The Compound Channel Capacity of a Class of Finite-State Channels Amos Lapidoth, Member, IEEE, İ. Emre Telatar, Member, IEEE Abstract

More information

Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay

Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay Lecture - 17 Shannon-Fano-Elias Coding and Introduction to Arithmetic Coding

More information

NOTES ON LINEAR TRANSFORMATIONS

NOTES ON LINEAR TRANSFORMATIONS NOTES ON LINEAR TRANSFORMATIONS Definition 1. Let V and W be vector spaces. A function T : V W is a linear transformation from V to W if the following two properties hold. i T v + v = T v + T v for all

More information

Vector and Matrix Norms

Vector and Matrix Norms Chapter 1 Vector and Matrix Norms 11 Vector Spaces Let F be a field (such as the real numbers, R, or complex numbers, C) with elements called scalars A Vector Space, V, over the field F is a non-empty

More information

Weakly Secure Network Coding

Weakly Secure Network Coding Weakly Secure Network Coding Kapil Bhattad, Student Member, IEEE and Krishna R. Narayanan, Member, IEEE Department of Electrical Engineering, Texas A&M University, College Station, USA Abstract In this

More information

Lecture 1: Course overview, circuits, and formulas

Lecture 1: Course overview, circuits, and formulas Lecture 1: Course overview, circuits, and formulas Topics in Complexity Theory and Pseudorandomness (Spring 2013) Rutgers University Swastik Kopparty Scribes: John Kim, Ben Lund 1 Course Information Swastik

More information

MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.

MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1. MATH10212 Linear Algebra Textbook: D. Poole, Linear Algebra: A Modern Introduction. Thompson, 2006. ISBN 0-534-40596-7. Systems of Linear Equations Definition. An n-dimensional vector is a row or a column

More information

Optimal Design of Sequential Real-Time Communication Systems Aditya Mahajan, Member, IEEE, and Demosthenis Teneketzis, Fellow, IEEE

Optimal Design of Sequential Real-Time Communication Systems Aditya Mahajan, Member, IEEE, and Demosthenis Teneketzis, Fellow, IEEE IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 55, NO. 11, NOVEMBER 2009 5317 Optimal Design of Sequential Real-Time Communication Systems Aditya Mahajan, Member, IEEE, Demosthenis Teneketzis, Fellow, IEEE

More information

Cartesian Products and Relations

Cartesian Products and Relations Cartesian Products and Relations Definition (Cartesian product) If A and B are sets, the Cartesian product of A and B is the set A B = {(a, b) :(a A) and (b B)}. The following points are worth special

More information

Gambling and Data Compression

Gambling and Data Compression Gambling and Data Compression Gambling. Horse Race Definition The wealth relative S(X) = b(x)o(x) is the factor by which the gambler s wealth grows if horse X wins the race, where b(x) is the fraction

More information

Lemma 5.2. Let S be a set. (1) Let f and g be two permutations of S. Then the composition of f and g is a permutation of S.

Lemma 5.2. Let S be a set. (1) Let f and g be two permutations of S. Then the composition of f and g is a permutation of S. Definition 51 Let S be a set bijection f : S S 5 Permutation groups A permutation of S is simply a Lemma 52 Let S be a set (1) Let f and g be two permutations of S Then the composition of f and g is a

More information

Sphere-Bound-Achieving Coset Codes and Multilevel Coset Codes

Sphere-Bound-Achieving Coset Codes and Multilevel Coset Codes 820 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL 46, NO 3, MAY 2000 Sphere-Bound-Achieving Coset Codes and Multilevel Coset Codes G David Forney, Jr, Fellow, IEEE, Mitchell D Trott, Member, IEEE, and Sae-Young

More information

4.5 Linear Dependence and Linear Independence

4.5 Linear Dependence and Linear Independence 4.5 Linear Dependence and Linear Independence 267 32. {v 1, v 2 }, where v 1, v 2 are collinear vectors in R 3. 33. Prove that if S and S are subsets of a vector space V such that S is a subset of S, then

More information

I. INTRODUCTION. of the biometric measurements is stored in the database

I. INTRODUCTION. of the biometric measurements is stored in the database 122 IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, VOL 6, NO 1, MARCH 2011 Privacy Security Trade-Offs in Biometric Security Systems Part I: Single Use Case Lifeng Lai, Member, IEEE, Siu-Wai

More information

So let us begin our quest to find the holy grail of real analysis.

So let us begin our quest to find the holy grail of real analysis. 1 Section 5.2 The Complete Ordered Field: Purpose of Section We present an axiomatic description of the real numbers as a complete ordered field. The axioms which describe the arithmetic of the real numbers

More information

Notes 11: List Decoding Folded Reed-Solomon Codes

Notes 11: List Decoding Folded Reed-Solomon Codes Introduction to Coding Theory CMU: Spring 2010 Notes 11: List Decoding Folded Reed-Solomon Codes April 2010 Lecturer: Venkatesan Guruswami Scribe: Venkatesan Guruswami At the end of the previous notes,

More information

Testing LTL Formula Translation into Büchi Automata

Testing LTL Formula Translation into Büchi Automata Testing LTL Formula Translation into Büchi Automata Heikki Tauriainen and Keijo Heljanko Helsinki University of Technology, Laboratory for Theoretical Computer Science, P. O. Box 5400, FIN-02015 HUT, Finland

More information

1 if 1 x 0 1 if 0 x 1

1 if 1 x 0 1 if 0 x 1 Chapter 3 Continuity In this chapter we begin by defining the fundamental notion of continuity for real valued functions of a single real variable. When trying to decide whether a given function is or

More information

6.3 Conditional Probability and Independence

6.3 Conditional Probability and Independence 222 CHAPTER 6. PROBABILITY 6.3 Conditional Probability and Independence Conditional Probability Two cubical dice each have a triangle painted on one side, a circle painted on two sides and a square painted

More information

Chapter 7: Products and quotients

Chapter 7: Products and quotients Chapter 7: Products and quotients Matthew Macauley Department of Mathematical Sciences Clemson University http://www.math.clemson.edu/~macaule/ Math 42, Spring 24 M. Macauley (Clemson) Chapter 7: Products

More information

Chapter 3. Cartesian Products and Relations. 3.1 Cartesian Products

Chapter 3. Cartesian Products and Relations. 3.1 Cartesian Products Chapter 3 Cartesian Products and Relations The material in this chapter is the first real encounter with abstraction. Relations are very general thing they are a special type of subset. After introducing

More information

Less naive Bayes spam detection

Less naive Bayes spam detection Less naive Bayes spam detection Hongming Yang Eindhoven University of Technology Dept. EE, Rm PT 3.27, P.O.Box 53, 5600MB Eindhoven The Netherlands. E-mail:h.m.yang@tue.nl also CoSiNe Connectivity Systems

More information

Two classes of ternary codes and their weight distributions

Two classes of ternary codes and their weight distributions Two classes of ternary codes and their weight distributions Cunsheng Ding, Torleiv Kløve, and Francesco Sica Abstract In this paper we describe two classes of ternary codes, determine their minimum weight

More information

Coding and decoding with convolutional codes. The Viterbi Algor

Coding and decoding with convolutional codes. The Viterbi Algor Coding and decoding with convolutional codes. The Viterbi Algorithm. 8 Block codes: main ideas Principles st point of view: infinite length block code nd point of view: convolutions Some examples Repetition

More information

Online Appendix to Stochastic Imitative Game Dynamics with Committed Agents

Online Appendix to Stochastic Imitative Game Dynamics with Committed Agents Online Appendix to Stochastic Imitative Game Dynamics with Committed Agents William H. Sandholm January 6, 22 O.. Imitative protocols, mean dynamics, and equilibrium selection In this section, we consider

More information

Formal Languages and Automata Theory - Regular Expressions and Finite Automata -

Formal Languages and Automata Theory - Regular Expressions and Finite Automata - Formal Languages and Automata Theory - Regular Expressions and Finite Automata - Samarjit Chakraborty Computer Engineering and Networks Laboratory Swiss Federal Institute of Technology (ETH) Zürich March

More information

ECON 459 Game Theory. Lecture Notes Auctions. Luca Anderlini Spring 2015

ECON 459 Game Theory. Lecture Notes Auctions. Luca Anderlini Spring 2015 ECON 459 Game Theory Lecture Notes Auctions Luca Anderlini Spring 2015 These notes have been used before. If you can still spot any errors or have any suggestions for improvement, please let me know. 1

More information

Mathematics Course 111: Algebra I Part IV: Vector Spaces

Mathematics Course 111: Algebra I Part IV: Vector Spaces Mathematics Course 111: Algebra I Part IV: Vector Spaces D. R. Wilkins Academic Year 1996-7 9 Vector Spaces A vector space over some field K is an algebraic structure consisting of a set V on which are

More information

3. Linear Programming and Polyhedral Combinatorics

3. Linear Programming and Polyhedral Combinatorics Massachusetts Institute of Technology Handout 6 18.433: Combinatorial Optimization February 20th, 2009 Michel X. Goemans 3. Linear Programming and Polyhedral Combinatorics Summary of what was seen in the

More information

Introduction to Algebraic Coding Theory

Introduction to Algebraic Coding Theory Introduction to Algebraic Coding Theory Supplementary material for Math 336 Cornell University Sarah A. Spence Contents 1 Introduction 1 2 Basics 2 2.1 Important code parameters..................... 4

More information

Communication on the Grassmann Manifold: A Geometric Approach to the Noncoherent Multiple-Antenna Channel

Communication on the Grassmann Manifold: A Geometric Approach to the Noncoherent Multiple-Antenna Channel IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 48, NO. 2, FEBRUARY 2002 359 Communication on the Grassmann Manifold: A Geometric Approach to the Noncoherent Multiple-Antenna Channel Lizhong Zheng, Student

More information

No: 10 04. Bilkent University. Monotonic Extension. Farhad Husseinov. Discussion Papers. Department of Economics

No: 10 04. Bilkent University. Monotonic Extension. Farhad Husseinov. Discussion Papers. Department of Economics No: 10 04 Bilkent University Monotonic Extension Farhad Husseinov Discussion Papers Department of Economics The Discussion Papers of the Department of Economics are intended to make the initial results

More information

1 Solving LPs: The Simplex Algorithm of George Dantzig

1 Solving LPs: The Simplex Algorithm of George Dantzig Solving LPs: The Simplex Algorithm of George Dantzig. Simplex Pivoting: Dictionary Format We illustrate a general solution procedure, called the simplex algorithm, by implementing it on a very simple example.

More information

Geometric Transformations

Geometric Transformations Geometric Transformations Definitions Def: f is a mapping (function) of a set A into a set B if for every element a of A there exists a unique element b of B that is paired with a; this pairing is denoted

More information

Chapter 17. Orthogonal Matrices and Symmetries of Space

Chapter 17. Orthogonal Matrices and Symmetries of Space Chapter 17. Orthogonal Matrices and Symmetries of Space Take a random matrix, say 1 3 A = 4 5 6, 7 8 9 and compare the lengths of e 1 and Ae 1. The vector e 1 has length 1, while Ae 1 = (1, 4, 7) has length

More information

You know from calculus that functions play a fundamental role in mathematics.

You know from calculus that functions play a fundamental role in mathematics. CHPTER 12 Functions You know from calculus that functions play a fundamental role in mathematics. You likely view a function as a kind of formula that describes a relationship between two (or more) quantities.

More information

Coding Theorems for Turbo-Like Codes Abstract. 1. Introduction.

Coding Theorems for Turbo-Like Codes Abstract. 1. Introduction. Coding Theorems for Turbo-Like Codes Dariush Divsalar, Hui Jin, and Robert J. McEliece Jet Propulsion Laboratory and California Institute of Technology Pasadena, California USA E-mail: dariush@shannon.jpl.nasa.gov,

More information

GRAPH THEORY LECTURE 4: TREES

GRAPH THEORY LECTURE 4: TREES GRAPH THEORY LECTURE 4: TREES Abstract. 3.1 presents some standard characterizations and properties of trees. 3.2 presents several different types of trees. 3.7 develops a counting method based on a bijection

More information

Basic Probability Concepts

Basic Probability Concepts page 1 Chapter 1 Basic Probability Concepts 1.1 Sample and Event Spaces 1.1.1 Sample Space A probabilistic (or statistical) experiment has the following characteristics: (a) the set of all possible outcomes

More information

6.4 Normal Distribution

6.4 Normal Distribution Contents 6.4 Normal Distribution....................... 381 6.4.1 Characteristics of the Normal Distribution....... 381 6.4.2 The Standardized Normal Distribution......... 385 6.4.3 Meaning of Areas under

More information

Lecture 15 An Arithmetic Circuit Lowerbound and Flows in Graphs

Lecture 15 An Arithmetic Circuit Lowerbound and Flows in Graphs CSE599s: Extremal Combinatorics November 21, 2011 Lecture 15 An Arithmetic Circuit Lowerbound and Flows in Graphs Lecturer: Anup Rao 1 An Arithmetic Circuit Lower Bound An arithmetic circuit is just like

More information

3. Mathematical Induction

3. Mathematical Induction 3. MATHEMATICAL INDUCTION 83 3. Mathematical Induction 3.1. First Principle of Mathematical Induction. Let P (n) be a predicate with domain of discourse (over) the natural numbers N = {0, 1,,...}. If (1)

More information

COMBINATORIAL PROPERTIES OF THE HIGMAN-SIMS GRAPH. 1. Introduction

COMBINATORIAL PROPERTIES OF THE HIGMAN-SIMS GRAPH. 1. Introduction COMBINATORIAL PROPERTIES OF THE HIGMAN-SIMS GRAPH ZACHARY ABEL 1. Introduction In this survey we discuss properties of the Higman-Sims graph, which has 100 vertices, 1100 edges, and is 22 regular. In fact

More information

Some Polynomial Theorems. John Kennedy Mathematics Department Santa Monica College 1900 Pico Blvd. Santa Monica, CA 90405 rkennedy@ix.netcom.

Some Polynomial Theorems. John Kennedy Mathematics Department Santa Monica College 1900 Pico Blvd. Santa Monica, CA 90405 rkennedy@ix.netcom. Some Polynomial Theorems by John Kennedy Mathematics Department Santa Monica College 1900 Pico Blvd. Santa Monica, CA 90405 rkennedy@ix.netcom.com This paper contains a collection of 31 theorems, lemmas,

More information

CHAPTER 2 Estimating Probabilities

CHAPTER 2 Estimating Probabilities CHAPTER 2 Estimating Probabilities Machine Learning Copyright c 2016. Tom M. Mitchell. All rights reserved. *DRAFT OF January 24, 2016* *PLEASE DO NOT DISTRIBUTE WITHOUT AUTHOR S PERMISSION* This is a

More information

Mathematical finance and linear programming (optimization)

Mathematical finance and linear programming (optimization) Mathematical finance and linear programming (optimization) Geir Dahl September 15, 2009 1 Introduction The purpose of this short note is to explain how linear programming (LP) (=linear optimization) may

More information

Outline 2.1 Graph Isomorphism 2.2 Automorphisms and Symmetry 2.3 Subgraphs, part 1

Outline 2.1 Graph Isomorphism 2.2 Automorphisms and Symmetry 2.3 Subgraphs, part 1 GRAPH THEORY LECTURE STRUCTURE AND REPRESENTATION PART A Abstract. Chapter focuses on the question of when two graphs are to be regarded as the same, on symmetries, and on subgraphs.. discusses the concept

More information

Lecture 22: November 10

Lecture 22: November 10 CS271 Randomness & Computation Fall 2011 Lecture 22: November 10 Lecturer: Alistair Sinclair Based on scribe notes by Rafael Frongillo Disclaimer: These notes have not been subjected to the usual scrutiny

More information

CHAPTER 5. Number Theory. 1. Integers and Division. Discussion

CHAPTER 5. Number Theory. 1. Integers and Division. Discussion CHAPTER 5 Number Theory 1. Integers and Division 1.1. Divisibility. Definition 1.1.1. Given two integers a and b we say a divides b if there is an integer c such that b = ac. If a divides b, we write a

More information

Linear Algebra Notes for Marsden and Tromba Vector Calculus

Linear Algebra Notes for Marsden and Tromba Vector Calculus Linear Algebra Notes for Marsden and Tromba Vector Calculus n-dimensional Euclidean Space and Matrices Definition of n space As was learned in Math b, a point in Euclidean three space can be thought of

More information

a 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2.

a 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2. Chapter 1 LINEAR EQUATIONS 1.1 Introduction to linear equations A linear equation in n unknowns x 1, x,, x n is an equation of the form a 1 x 1 + a x + + a n x n = b, where a 1, a,..., a n, b are given

More information

LEARNING OBJECTIVES FOR THIS CHAPTER

LEARNING OBJECTIVES FOR THIS CHAPTER CHAPTER 2 American mathematician Paul Halmos (1916 2006), who in 1942 published the first modern linear algebra book. The title of Halmos s book was the same as the title of this chapter. Finite-Dimensional

More information

How To Find An Optimal Search Protocol For An Oblivious Cell

How To Find An Optimal Search Protocol For An Oblivious Cell The Conference Call Search Problem in Wireless Networks Leah Epstein 1, and Asaf Levin 2 1 Department of Mathematics, University of Haifa, 31905 Haifa, Israel. lea@math.haifa.ac.il 2 Department of Statistics,

More information

Secure Network Coding on a Wiretap Network

Secure Network Coding on a Wiretap Network IEEE TRANSACTIONS ON INFORMATION THEORY 1 Secure Network Coding on a Wiretap Network Ning Cai, Senior Member, IEEE, and Raymond W. Yeung, Fellow, IEEE Abstract In the paradigm of network coding, the nodes

More information

Solution to Homework 2

Solution to Homework 2 Solution to Homework 2 Olena Bormashenko September 23, 2011 Section 1.4: 1(a)(b)(i)(k), 4, 5, 14; Section 1.5: 1(a)(b)(c)(d)(e)(n), 2(a)(c), 13, 16, 17, 18, 27 Section 1.4 1. Compute the following, if

More information

An example of a computable

An example of a computable An example of a computable absolutely normal number Verónica Becher Santiago Figueira Abstract The first example of an absolutely normal number was given by Sierpinski in 96, twenty years before the concept

More information

CONTINUED FRACTIONS AND FACTORING. Niels Lauritzen

CONTINUED FRACTIONS AND FACTORING. Niels Lauritzen CONTINUED FRACTIONS AND FACTORING Niels Lauritzen ii NIELS LAURITZEN DEPARTMENT OF MATHEMATICAL SCIENCES UNIVERSITY OF AARHUS, DENMARK EMAIL: niels@imf.au.dk URL: http://home.imf.au.dk/niels/ Contents

More information

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model 1 September 004 A. Introduction and assumptions The classical normal linear regression model can be written

More information

Optimal Index Codes for a Class of Multicast Networks with Receiver Side Information

Optimal Index Codes for a Class of Multicast Networks with Receiver Side Information Optimal Index Codes for a Class of Multicast Networks with Receiver Side Information Lawrence Ong School of Electrical Engineering and Computer Science, The University of Newcastle, Australia Email: lawrence.ong@cantab.net

More information

C H A P T E R Regular Expressions regular expression

C H A P T E R Regular Expressions regular expression 7 CHAPTER Regular Expressions Most programmers and other power-users of computer systems have used tools that match text patterns. You may have used a Web search engine with a pattern like travel cancun

More information

= 2 + 1 2 2 = 3 4, Now assume that P (k) is true for some fixed k 2. This means that

= 2 + 1 2 2 = 3 4, Now assume that P (k) is true for some fixed k 2. This means that Instructions. Answer each of the questions on your own paper, and be sure to show your work so that partial credit can be adequately assessed. Credit will not be given for answers (even correct ones) without

More information

Integer Programming Formulation

Integer Programming Formulation Integer Programming Formulation 1 Integer Programming Introduction When we introduced linear programs in Chapter 1, we mentioned divisibility as one of the LP assumptions. Divisibility allowed us to consider

More information

What is Linear Programming?

What is Linear Programming? Chapter 1 What is Linear Programming? An optimization problem usually has three essential ingredients: a variable vector x consisting of a set of unknowns to be determined, an objective function of x to

More information

CHAPTER 3. Methods of Proofs. 1. Logical Arguments and Formal Proofs

CHAPTER 3. Methods of Proofs. 1. Logical Arguments and Formal Proofs CHAPTER 3 Methods of Proofs 1. Logical Arguments and Formal Proofs 1.1. Basic Terminology. An axiom is a statement that is given to be true. A rule of inference is a logical rule that is used to deduce

More information

Approximation Algorithms

Approximation Algorithms Approximation Algorithms or: How I Learned to Stop Worrying and Deal with NP-Completeness Ong Jit Sheng, Jonathan (A0073924B) March, 2012 Overview Key Results (I) General techniques: Greedy algorithms

More information

The Goldberg Rao Algorithm for the Maximum Flow Problem

The Goldberg Rao Algorithm for the Maximum Flow Problem The Goldberg Rao Algorithm for the Maximum Flow Problem COS 528 class notes October 18, 2006 Scribe: Dávid Papp Main idea: use of the blocking flow paradigm to achieve essentially O(min{m 2/3, n 1/2 }

More information

Integer Factorization using the Quadratic Sieve

Integer Factorization using the Quadratic Sieve Integer Factorization using the Quadratic Sieve Chad Seibert* Division of Science and Mathematics University of Minnesota, Morris Morris, MN 56567 seib0060@morris.umn.edu March 16, 2011 Abstract We give

More information

An Adaptive Decoding Algorithm of LDPC Codes over the Binary Erasure Channel. Gou HOSOYA, Hideki YAGI, Toshiyasu MATSUSHIMA, and Shigeichi HIRASAWA

An Adaptive Decoding Algorithm of LDPC Codes over the Binary Erasure Channel. Gou HOSOYA, Hideki YAGI, Toshiyasu MATSUSHIMA, and Shigeichi HIRASAWA 2007 Hawaii and SITA Joint Conference on Information Theory, HISC2007 Hawaii, USA, May 29 31, 2007 An Adaptive Decoding Algorithm of LDPC Codes over the Binary Erasure Channel Gou HOSOYA, Hideki YAGI,

More information

LDPC Codes: An Introduction

LDPC Codes: An Introduction LDPC Codes: An Introduction Amin Shokrollahi Digital Fountain, Inc. 39141 Civic Center Drive, Fremont, CA 94538 amin@digitalfountain.com April 2, 2003 Abstract LDPC codes are one of the hottest topics

More information

T ( a i x i ) = a i T (x i ).

T ( a i x i ) = a i T (x i ). Chapter 2 Defn 1. (p. 65) Let V and W be vector spaces (over F ). We call a function T : V W a linear transformation form V to W if, for all x, y V and c F, we have (a) T (x + y) = T (x) + T (y) and (b)

More information

arxiv:1112.0829v1 [math.pr] 5 Dec 2011

arxiv:1112.0829v1 [math.pr] 5 Dec 2011 How Not to Win a Million Dollars: A Counterexample to a Conjecture of L. Breiman Thomas P. Hayes arxiv:1112.0829v1 [math.pr] 5 Dec 2011 Abstract Consider a gambling game in which we are allowed to repeatedly

More information

15.062 Data Mining: Algorithms and Applications Matrix Math Review

15.062 Data Mining: Algorithms and Applications Matrix Math Review .6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop

More information

Discrete Mathematics and Probability Theory Fall 2009 Satish Rao, David Tse Note 10

Discrete Mathematics and Probability Theory Fall 2009 Satish Rao, David Tse Note 10 CS 70 Discrete Mathematics and Probability Theory Fall 2009 Satish Rao, David Tse Note 10 Introduction to Discrete Probability Probability theory has its origins in gambling analyzing card games, dice,

More information

8 Square matrices continued: Determinants

8 Square matrices continued: Determinants 8 Square matrices continued: Determinants 8. Introduction Determinants give us important information about square matrices, and, as we ll soon see, are essential for the computation of eigenvalues. You

More information

SYSTEMS OF EQUATIONS AND MATRICES WITH THE TI-89. by Joseph Collison

SYSTEMS OF EQUATIONS AND MATRICES WITH THE TI-89. by Joseph Collison SYSTEMS OF EQUATIONS AND MATRICES WITH THE TI-89 by Joseph Collison Copyright 2000 by Joseph Collison All rights reserved Reproduction or translation of any part of this work beyond that permitted by Sections

More information

Tiers, Preference Similarity, and the Limits on Stable Partners

Tiers, Preference Similarity, and the Limits on Stable Partners Tiers, Preference Similarity, and the Limits on Stable Partners KANDORI, Michihiro, KOJIMA, Fuhito, and YASUDA, Yosuke February 7, 2010 Preliminary and incomplete. Do not circulate. Abstract We consider

More information

Adaptive Linear Programming Decoding

Adaptive Linear Programming Decoding Adaptive Linear Programming Decoding Mohammad H. Taghavi and Paul H. Siegel ECE Department, University of California, San Diego Email: (mtaghavi, psiegel)@ucsd.edu ISIT 2006, Seattle, USA, July 9 14, 2006

More information

Matrix Algebra. Some Basic Matrix Laws. Before reading the text or the following notes glance at the following list of basic matrix algebra laws.

Matrix Algebra. Some Basic Matrix Laws. Before reading the text or the following notes glance at the following list of basic matrix algebra laws. Matrix Algebra A. Doerr Before reading the text or the following notes glance at the following list of basic matrix algebra laws. Some Basic Matrix Laws Assume the orders of the matrices are such that

More information

Matrix Representations of Linear Transformations and Changes of Coordinates

Matrix Representations of Linear Transformations and Changes of Coordinates Matrix Representations of Linear Transformations and Changes of Coordinates 01 Subspaces and Bases 011 Definitions A subspace V of R n is a subset of R n that contains the zero element and is closed under

More information

Reading 13 : Finite State Automata and Regular Expressions

Reading 13 : Finite State Automata and Regular Expressions CS/Math 24: Introduction to Discrete Mathematics Fall 25 Reading 3 : Finite State Automata and Regular Expressions Instructors: Beck Hasti, Gautam Prakriya In this reading we study a mathematical model

More information

Mathematics Review for MS Finance Students

Mathematics Review for MS Finance Students Mathematics Review for MS Finance Students Anthony M. Marino Department of Finance and Business Economics Marshall School of Business Lecture 1: Introductory Material Sets The Real Number System Functions,

More information

1 Sets and Set Notation.

1 Sets and Set Notation. LINEAR ALGEBRA MATH 27.6 SPRING 23 (COHEN) LECTURE NOTES Sets and Set Notation. Definition (Naive Definition of a Set). A set is any collection of objects, called the elements of that set. We will most

More information

1 Error in Euler s Method

1 Error in Euler s Method 1 Error in Euler s Method Experience with Euler s 1 method raises some interesting questions about numerical approximations for the solutions of differential equations. 1. What determines the amount of

More information

ECE 842 Report Implementation of Elliptic Curve Cryptography

ECE 842 Report Implementation of Elliptic Curve Cryptography ECE 842 Report Implementation of Elliptic Curve Cryptography Wei-Yang Lin December 15, 2004 Abstract The aim of this report is to illustrate the issues in implementing a practical elliptic curve cryptographic

More information

Achievable Strategies for General Secure Network Coding

Achievable Strategies for General Secure Network Coding Achievable Strategies for General Secure Network Coding Tao Cui and Tracey Ho Department of Electrical Engineering California Institute of Technology Pasadena, CA 91125, USA Email: {taocui, tho}@caltech.edu

More information