Maximum Entropy Functions: Approximate Gács-Körner for Distributed Compression
|
|
|
- Marvin Hancock
- 9 years ago
- Views:
Transcription
1 Maximum Entropy Functions: Approximate Gács-Körner for Distributed Compression Salman Salamatian MIT Asaf Cohen Ben-Gurion University of the Negev Muriel Médard MIT ariv: v [cs.it] Apr 06 Abstract Consider two correlated sources and Y generated from a joint distribution p,y. Their Gács-Körner Common Information, a measure of common information that exploits the combinatorial structure of the distribution p,y, leads to a source decomposition that exhibits the latent common parts in and Y. Using this source decomposition we construct an efficient distributed compression scheme, which can be efficiently used in the network setting as well. Then, we relax the combinatorial conditions on the source distribution, which results in an efficient scheme with a helper node, which can be thought of as a front-end cache. This relaxation leads to an inherent trade-off between the rate of the helper and the rate reduction at the sources, which we capture by a notion of optimal decomposition. We formulate this as an approximate Gács-Körner optimization. We then discuss properties of this optimization, and provide connections with the maximal correlation coefficient, as well as an efficient algorithm, both through the application of spectral graph theory to the induced bipartite graph of p,y. I. INTRODUCTION Consider two distributed sources and Y generated from a joint distribution p,y. The fundamental compression limits of these distributed sources is given by the Slepian-Wolf rate region []. While the fundamental limits are well known, and some low-complexity codes exist, implementations of Slepian-Wolf codes in practical scenarios are seldom seen. Indeed, most present-day systems use naive solutions to avoid redundancy such as deduplication. This is partly explained by more realistic problem instances, for example, one in which the sources are distributed in a network. In such a setting, distributed compression cannot be separated from the network operations, and should be done hand in hand to achieve the optimal rates []. This brings complexity issues, as well as the implementation burden, which makes the optimal Slepian-Wolf codes not well suited for the task. In this paper, we propose to look at a sub-optimal, but very efficient distributed coding technique that uses the combinatorial structure of the correlation for compression. Namely, we try to find a common part in the sources and Y, and send this common part only once to the terminal. We then express and Y given that common part, which result in an overall reduction of the necessary rates. It turns out that the right common part to analyze is the Gács-Körner common information, introduced in [3]. This measure of commoninformation has a strong relationship with the combinatorial structure of the joint distribution p,y, and in particular to its bipartite representation. However, the Gács-Körner common information does not capture, in general, the complete correlation between the sources, an restrict the joint distributions for which such a scheme is possible. Instead, we relax this combinatorial condition, which leads to a scheme in which a helper is required. The role of the helper is to complement the distributed encoders if the common part is not exactly the same at both sources. By increasing the rate of the helper to the terminal, we can achieve better and better rate reduction. This type of approach is particularly well suited to the network setting. As we are decomposing our sources into common and non-common parts, and transmit these latent parts through the network, we can use efficient network coding techniques without affecting the source code the decomposed sources are independent and the rate reduction comes from solely the decomposition. Moreover, very efficient single-source lossless coding can be used. The helper node can be thought of as a front-end cache, which is well connected to the sources and is being used to significantly reduce the rates required in the last-mile to the terminals. Thus, when laying out a network, our result is useful in determining how to position the helper so it acts as a valuable interface between the sources and the terminals, which function of the data to store on it, and what are the required rates. Main Contributions: Our contributions are three-fold. First, we introduce a distributed coding technique based on the structure of the correlation between the two sources and Y, that we describe in terms of functions φ and φ Y. We start with the case in which a helper is not necessary, i.e., the functions φ and φ Y always agree. The common part is then given by the Gács-Körner Common information. Later, we introduce a helper node, which allows us to relax this agreement condition. For each case, we describe a simple and efficient coding method, and compute its achievable region. These regions reveal an inherent trade-off between the rate of the helper and the amount of rate reduction we can obtain at the sources. This trade-off is captured by the notion of optimal decomposition. We describe optimal decomposition as an optimization problem over functions, give it an approximate Gács-Körner formulation and discuss some of its properties in the context of coding and optimization. Finally, as the problem of finding an optimal decomposition is in general hard, we provide an approximate algorithm based on spectral graph theory, and draw some connections with the maximal correlation coefficient. A. Notation II. BACKGROUND Throughout, we define finite random variables with capital letters, e.g., taking values in the finite set [n ] :=
2 {,..., n }, where n is the size of the alphabet. We will be interested in correlated random variables and Y, taking value from a joint distribution p,y. We let P be the n n Y joint distribution matrix, that is {P} i,j = p,y (i, j). We denote by the vector of all ones whose size will be clear from the context. We also denote by I{.} the indicator function. We will represent networks in the following way: let G = (V, E, C) be an acyclic directed graph with edge set E, vertices V and an integer valued function on each edge, C : E N. The value C(e) for e E represents the capacity of the communication link e in bits per unit of time. In addition, let s, s Y V be two sender nodes and let t be a terminal node. We will denote values of min-cuts in the network by ρ( ; ). For example, ρ(s ; t ) represents the value of the min-cut from s to t, and ρ(s, s ; t ) the value of the min-cut from both s and s to t. B. Disjoint Components and Common Information A useful point of view for our problem will be to look at the bipartite representation of the joint distribution probability p,y, which is defined below: Definition. Let p,y be the joint probability distribution of, Y. We denote by the bipartite representation of p,y the bipartite graph with n + n Y nodes indexed by v a,i where a {, Y } and i [n a ]. We let an edge connect node v,i to v Y,j only if p,y (i, j) > 0, and let p,y (i, j) be the weight of that edge. This bipartite representation allows for a simple characterization of the Gács-Körner Common Information. Specifically, we have the following definition of common information: Definition ( []). A set of nodes C such that there are no outgoing edges from C is called a connected component of the bipartite graph. We associate with each connected component C a weight p(c) = n x,n y C P ( = x, Y = y). We call the common information decomposition of p,y, the decomposition of the bipartite graph into a maximal number of connected components C,..., C k. Moreover, we denote by K,Y the common information random variable representing the index of the connected component, with the natural distribution (p(c ),..., p(c k )). The entropy of this random variable is H(K,Y ) = H(C) = ( ) k i= p(c i) log. p(c i) Definition is equivalent to the usual definition of common information in the random variables setting given in [3]: K,Y = argmax H(U) () H(U )=H(U Y )=0 It is well known that the above optimization problem achieves its maximum point by letting U be the index of the connected component, we can rewrite the optimization above by restrciting the space to these functions. We now present an alternative formulation of the above optimization. While being equivalent in the most basic setting, this formulation will allow us to extend the notion of common information to cases where the K,Y is zero or two small to be practical, and be the basis for future arguments in this paper. Precisely, consider functions φ and φ Y mapping [n ] and [n Y ] respectively, into a finite set. It is then easy to see that the common information then becomes the solution of the following optimization problem: max φ,φ Y H(φ ()), s.t. P(φ () φ Y (Y )) = 0. () where the optimization is over finite range functions φ, φ Y. Example: Let P, be defined as in joint probability table below with its corresponding -partite graph representation. Consider the function f(i) that takes value f(i) = if i {, }, and f(i) = if i {3, }. Then the maximizing functions in () are φ = φ Y = f. 3 Y 3 P = 8 C. Maximal Correlation Coefficient We now define an alternative measure of information between random variables and Y namely, the maximal correlation, first defined in [5]. For a given joint distribution p,y, we let ρ m (; Y ) := sup φ,φ Y E[φ ()φ Y (Y )], where the maximization is taken over all mean zero real-valued functions f and g on [n x ] and [n y ], such that E[φ () ] = E[φ Y (Y ) ] =. For random variables taking values in a finite set, it has been shown that the maximal correlation coefficient is related to the spectrum of the matrix P through the following identity proved in []: Theorem. Denote by D the diagonal matrix with diagonal elements (D ) i,i = p (i), and define D Y similarly. We let Q R nx ny be the matrix defined as Q := D / PD / Y. Its singular value decomposition is given by Q = UΣV T, and we denote by λ... λ d the ordered singular values (we refer the readers to [] or [6] to see why the highest singular value always takes value ). In particular, λ = ρ m(; Y ) is the squared maximal correlation coefficient of p,y. III. AN EFFICIENT DISTRIBUTED CODING SCHEME A. Using G-K common information Consider the classical Slepian-Wolf setting, despicted in Fig. In that setting, two separate encoders E and E Y encode at rates R and R Y, respectively. The goal is to recover, in a lossless way, the pair of sources (, Y ) at a destination node D,Y. We will be interested in efficient encoding and decoding for this canonical problem. In particular, when there is non-zero common information, i.e. H(K) > 0, we can use a simple source decomposition of the sources and Y. Namely, let (, K), where K is the index of the connected component, and is the position of inside that component. Note that this is a bijective function of. Note that at the source Y we can also do a similar decomposition Y (Y, K). Noticing that K is common in both decompositions yield a very simple distributed coding scheme, in which and Y are sent as if independent, but K is sent only once to the receiver. All of these operations can
3 Y R H R Y H E E H D,Y ( ˆ, Ŷ ) E Y R R H R Y Fig. : Slepian-Wolf with Helper. In the classical setting, R H = R H = R Y H = 0 be done efficiently, and the resulting coding is a zero-error distributed code. Theorem (Coding with G-K Common Information [7]). Let K be the G-K Common information between and Y. Then, there exist an efficient zero error encoding and decoding of and Y that operates at rates: R H( K), R Y H(Y K), R + R Y H( K)+H(Y K) + H(K). (3) Note that the rates in this result depend on the value of the common information. Indeed, the higher H(K), the greater is the reduction in rates. However, as we mentioned before,the common-information imposes a strong combinatorial condition on the joint distribution matrix. The next section aims at relaxing this strong condition by considering a helper node. Remark. The proposed coding scheme is different than the coding that exist in zero-error distributed coding, see e.g. [8], [9]. In general, the zero-error distributed coding sum-rate is better than the sum rate in the coding scheme above. However, this former type of methods do not generalize very well to any point on the achievable region, as there is generally very little flexibility in the achievable rates. This make these codes less suited for network extensions where the choice of the operating rates is given by the network topology. B. General distributions with a helper We now aim at moving away from the combinatorial condition imposed on the joint distribution p,y. In particular, say the bipartite graph is made of two almost disjoint components, only connected by an edge of small weight ɛ. In that case, the common information evaluates to 0. However, it would seem like being optimistic and disregarding this edge yield two disjoint components, and therefore a non-zero common information. To correct the problematic cases, we can consider an omniscient helping node, which simply sends a bit in case of an error. This insight is developed in the theorem below, where we consider arbitrary binary functions of and Y, and decompose the sources based on these functions. Theorem 3 (Binary functions). Consider φ and φ Y taking values in {0, } be functions of and Y, respectively. Let P r(φ φ Y ) = ɛ. Then, there exist an efficient zero-error encoding and decoding of and Y with a helper of rate R h = h(ɛ), that operates at rates, for some 0 α : R H( φ ()), R Y H(Y φ Y (Y )), R + R Y H( φ ())+H(Y φ Y (Y ))+ αh(φ ()) + ( α)h(φ Y (Y )). () Proof: We prove the corner points (H( φ ()), H(Y )). The rest of the region follows by symmetry and time-sharing. Consider n iid samples (x n, y n ), and the corresponding (φ (x n ), φ Y (y n )). Denote by e n the sequence such that e i = if φ (x i ) φ Y (y i ) and 0 otherwise. Then, we can encode x n given φ Y (x n ) using approximately nh( φ ()) bits. In contrast, we encode y n fully using nh(y ) bits, and let the helper encode e n using nh(ɛ) bits. At the receiver, we can decode x n, and e n. From x n we can obtain φ (x n ) which along with e n determines uniquely φ Y (y n ). Therefore we can obtain y n as well. It is interesting to note that in comparison with the usual Slepian-Wolf region, the rates in () do not describe a symmetric achievable region, that is the dominant face is not necessarily a 5 degrees line. The above theorem can be extended to the setting in which the functions φ and φ Y can take value in larger sets. The rate of the helper then depends on the position on the dominant face, as we need to encode which of the values φ and φ Y to take if φ φ Y. Theorem (Arbitrary functions). Let φ : {,..., n x } and φ Y : Y {,..., n y }. Let the rate of the helper be H(φ () φ(y )). The following corner point is achievable: R H( φ ()), R Y H(Y ), R h H(φ () φ Y (Y )). (5) This theorem can be extended to obtain a complete rate region, by first considering the symmetric corner point and then time-sharing. Note that in this context, the rate of the helper changes depending on the position in the dominant face. The proof is omitted. Example: Let us revisit our previous example. This time δ however, we let an edge of weight 8 join the two components. The common information of the resulting p,y is zero. However, we can find a simple decomposition of the sources by considering the same φ and φ Y as before, and letting the helper encode the errors. More precisely, let E i = { i = x and Y i = y 3 }. Then, the helper can simply encode E n, using approximately nh(e) = nh( δ 8 ) bits. Y 0 0 P = δ δ It is interesting to note that in the previous example, the helper node does not need to know the value of and Y if they are not part of the cut. Precisely, the source only need to send whether the value is =. Similarly, at Y, we only need to describe whether Y = 3. Indeed, this information is enough for the helper to compute whether there was an error. In fact, this observation is not specific to this example, and the helper need not to be fully omniscient. Precisely, let S = {i [n ] φ (i) φ Y (j) and P (i, j) > 0 for some j} be the subset of [n ] for which may cause confusion. We then define the random variable cut as the random variable restricted to that set S, that is cut = I{ S }. Similarly, we can define the set S Y and the random variable Y cut.
4 Theorem 5. Let φ and φ Y be functions on [n ] and [n Y ] respectively. Then, the helper performs as good as an omniscient helper as long as: R H H( cut ), R Y H H(Y cut ), (6) where R H and R Y H are the rates from sources and Y respectively to the helper. Proof: Let the sources encode the sources cut and Y cut, and send them to the helper. Suppose / S, then by construction it must be that φ () = φ Y (Y ), and therefore the helper can operate. If S, it must be that Y S Y too, and therefore the helper can detect the errors. Example: Consider as an example the probability distribution represented below on the right. We let φ (i) = if i =,, and φ (3) =. On the other hand, we let φ Y (j) = j. The resulting rate region R φ derived from () for this choice of functions is represented on the left, along with the Slepian-Wolf rate region R SW, and the Gács-Körner region R GK. In this case, the rate region R φ is larger than the Slepian-Wolf region, but it has to be emphasized that the helper has non-zero rate here! In fact, the rate of the helper is precisely h( ɛ). Also note that the region R φ does not have a 5 degree slope dominant face. R Y R GK R SW R 3 h( ) h( ) H() Remark. Denote by E the output of the helper node. Then, (I(Y ; E ), I(; E Y ), I(; Y E)) I(; Y ), where I(; Y ) is the region of tension defined in [0]. This region of tension is related to a notion of assisted common information, which is somewhat of a dual problem. Indeed, in the assisted common information problem, a helper node communicates with the sources directly, and helps them in agreeing on a common function. In our setting, the helper sends information to the terminal, and not to the sources. C. Network Setting R ( ɛ) ɛ ɛ Y ( ɛ) This type of approach to distributed coding generalizes nicely to the network setting, in particular correlated sources multicast. In this setting, the helper node is a front-end node who has large rate from the original sources, but limited rate towards the terminals. More precisely, consider a network G with two sources s and s Y, and a terminal t T. In addition, we consider a helper node s H, which we assume is different than s and s Y to avoid trivial cases. To efficiently transmit and Y to the terminals we follow these steps. First, we transmit and Y to the front-end helper s H. Then, we transmit and Y to the terminal using source decomposition, with the help of s H. Note that we exploit the fact that φ φ Y to send the function once, and express and Y with respect to the φ and φ Y respectively. The resulting sufficient conditions are expressed below, where we only consider a corner point coding for notation purposes. Theorem 6. Let φ and φ Y be functions. Sources and Y generated at nodes s and s Y can be reliably and efficiently communicated through the network, if the following min-cut relations are satisfied: ρ(s ; s H ) H(), ρ(s Y ; s H ) H(Y ) ρ(s, s Y ; s H ) H() + H(Y ) ρ(s ; t) H( φ ()), ρ(s Y ; t) H(Y ) ρ(s H ; t) H(φ () φ Y (Y )) ρ(s, s Y, s H ; t) H( φ ()) + H(Y φ Y (Y ))+ H(φ (), φ Y (Y )) (7) Proof: The complete proof is omitted, but follows from the min-cut conditions for independent source coding. Note that the first three conditions characterize sufficient mincut conditions from the source to the helper, while the last conditions characterize sufficient conditions from the sources and helper to the terminals. We refer the readers to [7] in which similar results are established. Remark 3. When the rate of the helper is 0, this reduces to the results of [7], in which the common information K between and Y is sent only once to the terminal, and and Y are described given that common information. Remark. Note that the rates to the helper H() and H(Y ) can be reduced by using Theorem 5. IV. OPTIMAL DECOMPOSITIONS So far, we have considered arbitrary decompositions of the the bipartite graph of p,y. In this section we aim at developing a notion of optimal decomposition. Note that there is an inherent trade-off in the construction of decomposition, as on one hand we wish to increase the entropy of the function φ (), and on the other hand we want to make it agree with φ Y (Y ) as often as possible. This trade-off is characterized by the reduction in rates from the sources, and the necessary rates from the helper, and is specified by the definition below: Definition 3 (Maximum Entropy Function). We say a decomposition (, φ ()) and Y (Y, φ Y (Y )) is optimal for helper rate ɛ if it is the solution of the optimization: maximize H(φ ()) subject to H(φ () φ Y (Y )) ɛ (8) where the optimization is taken over all functions φ and φ Y taking values in a finite set. Note that the previous definition indeed captures the rates in Theorem, as maximizing H(φ ()) equivalently minimizes H( φ ()) by observing that H() = H(φ ())+ H( φ ()). Remark 5. In general, the problem in (8) is not equivalent to the symmetric problem of maximizing H(φ Y (Y )) with constraint H(φ Y (Y ) φ ()) ɛ. However, for symmetric matrices P, i.e. P = P T, the two problems are equivalent. It then follows that the rate region under optimal decomposition () will be symmetric for symmetric joint distributions.
5 An alternative formulation of the optimization in (8) is given by the Lagrangian relaxation: Definition (Lagrangian Formulation). The lagrangian formulation of (8) is, for some λ > 0: maximize φ,φ Y H(φ ()) λh(φ () φ Y (Y )) (9) The formulation above is useful as it is an unconstrained optimization problem. Using this formulation, we first show that for large enough λ, the solution for this optimization problem is the G-K common information, and the helper is not required. We then focus on binary functions, and show a useful matrix representation as well. This allows us to harness spectral graph theory and suggest an approximation to the problem. Lemma. There exist a λ max such that for any λ > λ max, the solution of (9) is the solution of (), i.e. the optimal functions are the index of the disjoint components. This Lagrangian relaxation lends itself to a simple matrix notation when φ and φ Y take values in {, }. Recall that in that case, the necessary rate of the helper only encodes the error events, so it suffice to characterize the probability of error. We have: P(φ () φ Y (Y )) = i,j p,y (i, j) ( φ (i)φ Y (j)) = ( φt P φ Y ) (0) where φ and φ Y are column vectors of length n and n Y respectively, such that (φ a ) i = φ a (i), for a {, Y } and i [n a ]. Similarly, the probability P(φ () = ) can be written as ( + φt P). Corollary (Matrix notation). Let φ and φ Y take values in {, }. Then, we can write (9) as : ( ) ( ) max h φ x,φ y ( + φt P) λh ( φt Pφ Y ) () where the maximization is over vectors φ and φ Y of size n x and n y respectively, where each coordinate is in {, }. This matrix notation allows for some interesting insights. First, suppose there is a solution that has probability of error P e := ( φt Pφ Y ). Then, by considering φ Y = φ Y, we obtain a solution that has probability of error P e. As the binary entropy function is symmetric, these two probability of errors lead to an identical rate from the helper, and therefore, there is no loss in only considering solutions such that the probability of error is less than. On the other hand, observe that h ( ( + φt P)) takes its maximum when φ T P is close to 0. Therefore, there is once again no loss in considering φ such that P(φ () = ), as φ () = φ () achieves P( φ () = ) = P(φ () = ). These two observations along with the monotonicity of h( ) allow to simplify the search space, and lead to the following result: Corollary. There exist a λ > 0 such that the solution of () is also the solution of: max φ,φ Y P(φ () = ) P(φ () φ Y (Y )), s.t. P(φ () = ) Proof: The complete proof is omitted but follows essentially from a linearization, and a Dinkelbach decomposition [], which allows to transform the optimization problem () into an equivalent fractional optimization (for some λ). The optimization in Corollary is similar to a graph conductance formulation. Indeed, the term P(φ () φ Y (Y )) can be interpreted as the value of a cut on the bipartite graph for the partition S = {(, Y ) φ = φ Y = }, while the value P(φ () = ) is related to the size of the partition S. This relationship can be made precise in the case of symmetric matrices P, and is the object of some future work. A. Spectral algorithm for finding optimal decomposition In this section, we present an approximation algorithm to optimal decomposition. Indeed, in general the problem is combinatorial and unless λ is large enough, in which case it reduces to finding disjoint components in the bipartite graph, it is a hard problem. In particular, Corollary suggests that this problem is related to the graph conductance problem in some contexts, which is a hard problem to solve exactly. However, a reasonable approach is to look at a spectral clustering approach, in which the goal is to find a well-balanced (in terms of size) partition of nodes such that they are well separated (the cut between the partition is small). This is motivated by results in spectral graph theory, namely Cheeger s Inequality [], which relate the value of the conductance of the graph to the second eigenvalue of the so-called normalize Laplacian matrix. However, because of the special structure of our graph, namely a bipartite graph describing a joint distribution, we will exhibit a relationship between the normalized Laplacian matrix, and the spectral decomposition in Theorem, and show that the second eigenvalue of the normalized laplacian is in fact the maximal correlation coefficient. Before we proceed, let us recall the definition of the normalized Laplacian matrix. Let G = (V, E, w) be an undirected graph, with V = n nodes, and weights on the edges w i,j. We let A be the n n adjacency matrix defined as (A) i,j = w i,j. Let D be the diagonal matrix with diagonal entries (D) i,i = i,j w i,j. Then, the normalized Laplacian matrix N is defined as N := I D / AD /. The following theorem links the second smallest eigenvalue of N for the bipartite representation of p,y, with ρ m(; Y ). Theorem 7. Denote by ν the second smallest eigenvalue of N. Then, ν = ρ m(; Y ). Proof: Noticing that for the bipartite graph, the adjacency matrix A and the corresponding D can be written as: [ ] [ ] 0 P D 0 A = P T, D =. 0 0 D Y [ ] D / AD / 0 D / = PD / Y D / Y P T D / 0 Therefore, the eigenvalues of D / AD / are the plus or minus singular values of Q = D / PD / Y, and the second largest singular value of Q must be equal to ν. This theorem relates the maximal correlation coefficient to ν, and therefore indirectly to the conductance of the bipartite
6 graph. In fact, the following corralary is immediate, and shows that when there exist disjoint components, the maximal correlation is precisely. Corollary 3. Let ρ m (; Y ) be the maximal correlation of p,y, and denote by K the G-K common information. Then, H(K) 0 if and only if ρ m (; Y ) =. In addition, call m the multiplicity of the singular value in the singular value decomposition of Q. Then H(K) log(m). [9] H. S. Witsenhausen, The zero-error side information problem and chromatic numbers (corresp.), IEEE Trans. on Info. Theory, 976. [0] V. M. Prabhakaran and M. M. Prabhakaran, Assisted common information with an application to secure two-party sampling, IEEE Trans. on Info. Theory, 0. [] W. Dinkelbach, On nonlinear fractional programming, Management Science, vol. 3, no. 7, pp. 9 98, 967. [] D. A. Spielman, Spectral graph theory and its applications, 007. In view of these results, we propose a simple and efficient algorithm for finding binary functions φ and φ Y. We compute the left and right singular vectors u and v, respectively, corresponding to the second highest singular value of Q. Then, we chose a threshold and let φ (i) = if u i > t and φ (i) = otherwise, and similarly for φ Y (j) and v j. Note that for an appropriate choice of t, this algorithm finds disjoint components in the bipartite graph, if they exist. Remark 6. The above algorithm can be iteratively applied to each partition in order to find functions φ and φ Y that have non-binary range, or equivalently to find more clusters. V. CONCLUDING REMARKS The optimization problem in (9) leads to a couple of interesting observations. First, we have only considered the case where the function φ is a function of one single symbol. A natural extension would be to consider the product distribution p n,y n and functions that take as input n symbols. In contrast with the case of Gács-Körner common information, in which taking the product distribution could not help as the number of disjoint components would not increase, the optimization in (9) will likely improve by taking product distributions as it increases the options available for edge removal. Another interesting open question is related to the relationship between the maximal correlation and the optimal source decomposition. In particular, we have shown that the maximal correlation coefficient can be used as an approximation to optimal decomposition, but only for some specific λ. A question of interest would be to determine whether a similar spectral method can be used to approximate the trade-off fully. REFERENCES [] D. Slepian and J. K. Wolf, Noiseless coding of correlated information sources, IEEE Trans. Info. Theory, vol. 9, no., pp. 7 80, 973. [] A. Ramamoorthy, K. Jain, P. Chou, and M. Effros, Separating distributed source coding from network coding, IEEE Trans. on Info. Theory, 006. [3] P. Gács and J. Körner, Common information is far less than mutual information, Problems of Control and Info. Theory, 973. [] H. S. Witsenhausen, On sequences of pairs of dependent random variables, SIAM Journal on Applied Mathematics, 975. [5] A. Rényi, On measures of dependence, Acta Mathematica Academiae Scientiarum Hungarica, vol. 0, no. 3, pp. 5, 959. [6] F. P. Calmon, M. Varia, and M. Médard, An exploration of the role of principal inertia components in information theory, in Information Theory Workshop (ITW), 0 IEEE, Nov 0, pp [7] S. Salamatian, A. Cohen, and M. Médard, Efficient coding for multisource networks using Gács-Körner common information, Arxiv, 06. [8] P. Koulgi, E. Tuncel, S. L. Regunathan, and K. Rose, On zero-error source coding with decoder side information, IEEE Trans. on Info. Theory, 003.
COMBINATORIAL PROPERTIES OF THE HIGMAN-SIMS GRAPH. 1. Introduction
COMBINATORIAL PROPERTIES OF THE HIGMAN-SIMS GRAPH ZACHARY ABEL 1. Introduction In this survey we discuss properties of the Higman-Sims graph, which has 100 vertices, 1100 edges, and is 22 regular. In fact
On the Multiple Unicast Network Coding Conjecture
On the Multiple Unicast Networ Coding Conjecture Michael Langberg Computer Science Division Open University of Israel Raanana 43107, Israel [email protected] Muriel Médard Research Laboratory of Electronics
SHARP BOUNDS FOR THE SUM OF THE SQUARES OF THE DEGREES OF A GRAPH
31 Kragujevac J. Math. 25 (2003) 31 49. SHARP BOUNDS FOR THE SUM OF THE SQUARES OF THE DEGREES OF A GRAPH Kinkar Ch. Das Department of Mathematics, Indian Institute of Technology, Kharagpur 721302, W.B.,
A Network Flow Approach in Cloud Computing
1 A Network Flow Approach in Cloud Computing Soheil Feizi, Amy Zhang, Muriel Médard RLE at MIT Abstract In this paper, by using network flow principles, we propose algorithms to address various challenges
Weakly Secure Network Coding
Weakly Secure Network Coding Kapil Bhattad, Student Member, IEEE and Krishna R. Narayanan, Member, IEEE Department of Electrical Engineering, Texas A&M University, College Station, USA Abstract In this
Fairness in Routing and Load Balancing
Fairness in Routing and Load Balancing Jon Kleinberg Yuval Rabani Éva Tardos Abstract We consider the issue of network routing subject to explicit fairness conditions. The optimization of fairness criteria
Codes for Network Switches
Codes for Network Switches Zhiying Wang, Omer Shaked, Yuval Cassuto, and Jehoshua Bruck Electrical Engineering Department, California Institute of Technology, Pasadena, CA 91125, USA Electrical Engineering
Part 2: Community Detection
Chapter 8: Graph Data Part 2: Community Detection Based on Leskovec, Rajaraman, Ullman 2014: Mining of Massive Datasets Big Data Management and Analytics Outline Community Detection - Social networks -
Conductance, the Normalized Laplacian, and Cheeger s Inequality
Spectral Graph Theory Lecture 6 Conductance, the Normalized Laplacian, and Cheeger s Inequality Daniel A. Spielman September 21, 2015 Disclaimer These notes are not necessarily an accurate representation
Linear Codes. Chapter 3. 3.1 Basics
Chapter 3 Linear Codes In order to define codes that we can encode and decode efficiently, we add more structure to the codespace. We shall be mainly interested in linear codes. A linear code of length
DATA ANALYSIS II. Matrix Algorithms
DATA ANALYSIS II Matrix Algorithms Similarity Matrix Given a dataset D = {x i }, i=1,..,n consisting of n points in R d, let A denote the n n symmetric similarity matrix between the points, given as where
I. GROUPS: BASIC DEFINITIONS AND EXAMPLES
I GROUPS: BASIC DEFINITIONS AND EXAMPLES Definition 1: An operation on a set G is a function : G G G Definition 2: A group is a set G which is equipped with an operation and a special element e G, called
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS 1. SYSTEMS OF EQUATIONS AND MATRICES 1.1. Representation of a linear system. The general system of m equations in n unknowns can be written a 11 x 1 + a 12 x 2 +
CHAPTER 9. Integer Programming
CHAPTER 9 Integer Programming An integer linear program (ILP) is, by definition, a linear program with the additional constraint that all variables take integer values: (9.1) max c T x s t Ax b and x integral
1 Solving LPs: The Simplex Algorithm of George Dantzig
Solving LPs: The Simplex Algorithm of George Dantzig. Simplex Pivoting: Dictionary Format We illustrate a general solution procedure, called the simplex algorithm, by implementing it on a very simple example.
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS Systems of Equations and Matrices Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a
Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay
Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay Lecture - 17 Shannon-Fano-Elias Coding and Introduction to Arithmetic Coding
Secure Network Coding on a Wiretap Network
IEEE TRANSACTIONS ON INFORMATION THEORY 1 Secure Network Coding on a Wiretap Network Ning Cai, Senior Member, IEEE, and Raymond W. Yeung, Fellow, IEEE Abstract In the paradigm of network coding, the nodes
5.1 Bipartite Matching
CS787: Advanced Algorithms Lecture 5: Applications of Network Flow In the last lecture, we looked at the problem of finding the maximum flow in a graph, and how it can be efficiently solved using the Ford-Fulkerson
Lecture 3: Finding integer solutions to systems of linear equations
Lecture 3: Finding integer solutions to systems of linear equations Algorithmic Number Theory (Fall 2014) Rutgers University Swastik Kopparty Scribe: Abhishek Bhrushundi 1 Overview The goal of this lecture
Introduction to Matrix Algebra
Psychology 7291: Multivariate Statistics (Carey) 8/27/98 Matrix Algebra - 1 Introduction to Matrix Algebra Definitions: A matrix is a collection of numbers ordered by rows and columns. It is customary
Practical Guide to the Simplex Method of Linear Programming
Practical Guide to the Simplex Method of Linear Programming Marcel Oliver Revised: April, 0 The basic steps of the simplex algorithm Step : Write the linear programming problem in standard form Linear
A Practical Scheme for Wireless Network Operation
A Practical Scheme for Wireless Network Operation Radhika Gowaikar, Amir F. Dana, Babak Hassibi, Michelle Effros June 21, 2004 Abstract In many problems in wireline networks, it is known that achieving
α = u v. In other words, Orthogonal Projection
Orthogonal Projection Given any nonzero vector v, it is possible to decompose an arbitrary vector u into a component that points in the direction of v and one that points in a direction orthogonal to v
Continued Fractions and the Euclidean Algorithm
Continued Fractions and the Euclidean Algorithm Lecture notes prepared for MATH 326, Spring 997 Department of Mathematics and Statistics University at Albany William F Hammond Table of Contents Introduction
Notes on Determinant
ENGG2012B Advanced Engineering Mathematics Notes on Determinant Lecturer: Kenneth Shum Lecture 9-18/02/2013 The determinant of a system of linear equations determines whether the solution is unique, without
Triangle deletion. Ernie Croot. February 3, 2010
Triangle deletion Ernie Croot February 3, 2010 1 Introduction The purpose of this note is to give an intuitive outline of the triangle deletion theorem of Ruzsa and Szemerédi, which says that if G = (V,
Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh
Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh Peter Richtárik Week 3 Randomized Coordinate Descent With Arbitrary Sampling January 27, 2016 1 / 30 The Problem
INDISTINGUISHABILITY OF ABSOLUTELY CONTINUOUS AND SINGULAR DISTRIBUTIONS
INDISTINGUISHABILITY OF ABSOLUTELY CONTINUOUS AND SINGULAR DISTRIBUTIONS STEVEN P. LALLEY AND ANDREW NOBEL Abstract. It is shown that there are no consistent decision rules for the hypothesis testing problem
UNCOUPLING THE PERRON EIGENVECTOR PROBLEM
UNCOUPLING THE PERRON EIGENVECTOR PROBLEM Carl D Meyer INTRODUCTION Foranonnegative irreducible matrix m m with spectral radius ρ,afundamental problem concerns the determination of the unique normalized
Au = = = 3u. Aw = = = 2w. so the action of A on u and w is very easy to picture: it simply amounts to a stretching by 3 and 2, respectively.
Chapter 7 Eigenvalues and Eigenvectors In this last chapter of our exploration of Linear Algebra we will revisit eigenvalues and eigenvectors of matrices, concepts that were already introduced in Geometry
Online Adwords Allocation
Online Adwords Allocation Shoshana Neuburger May 6, 2009 1 Overview Many search engines auction the advertising space alongside search results. When Google interviewed Amin Saberi in 2004, their advertisement
The Goldberg Rao Algorithm for the Maximum Flow Problem
The Goldberg Rao Algorithm for the Maximum Flow Problem COS 528 class notes October 18, 2006 Scribe: Dávid Papp Main idea: use of the blocking flow paradigm to achieve essentially O(min{m 2/3, n 1/2 }
Definition 11.1. Given a graph G on n vertices, we define the following quantities:
Lecture 11 The Lovász ϑ Function 11.1 Perfect graphs We begin with some background on perfect graphs. graphs. First, we define some quantities on Definition 11.1. Given a graph G on n vertices, we define
Inner Product Spaces
Math 571 Inner Product Spaces 1. Preliminaries An inner product space is a vector space V along with a function, called an inner product which associates each pair of vectors u, v with a scalar u, v, and
This article has been accepted for inclusion in a future issue of this journal. Content is final as presented, with the exception of pagination.
IEEE/ACM TRANSACTIONS ON NETWORKING 1 A Greedy Link Scheduler for Wireless Networks With Gaussian Multiple-Access and Broadcast Channels Arun Sridharan, Student Member, IEEE, C Emre Koksal, Member, IEEE,
MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.
MATH10212 Linear Algebra Textbook: D. Poole, Linear Algebra: A Modern Introduction. Thompson, 2006. ISBN 0-534-40596-7. Systems of Linear Equations Definition. An n-dimensional vector is a row or a column
Efficient Recovery of Secrets
Efficient Recovery of Secrets Marcel Fernandez Miguel Soriano, IEEE Senior Member Department of Telematics Engineering. Universitat Politècnica de Catalunya. C/ Jordi Girona 1 i 3. Campus Nord, Mod C3,
7 Gaussian Elimination and LU Factorization
7 Gaussian Elimination and LU Factorization In this final section on matrix factorization methods for solving Ax = b we want to take a closer look at Gaussian elimination (probably the best known method
Mining Social-Network Graphs
342 Chapter 10 Mining Social-Network Graphs There is much information to be gained by analyzing the large-scale data that is derived from social networks. The best-known example of a social network is
Mathematical Induction
Mathematical Induction (Handout March 8, 01) The Principle of Mathematical Induction provides a means to prove infinitely many statements all at once The principle is logical rather than strictly mathematical,
I. INTRODUCTION. of the biometric measurements is stored in the database
122 IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, VOL 6, NO 1, MARCH 2011 Privacy Security Trade-Offs in Biometric Security Systems Part I: Single Use Case Lifeng Lai, Member, IEEE, Siu-Wai
Math 4310 Handout - Quotient Vector Spaces
Math 4310 Handout - Quotient Vector Spaces Dan Collins The textbook defines a subspace of a vector space in Chapter 4, but it avoids ever discussing the notion of a quotient space. This is understandable
CS 598CSC: Combinatorial Optimization Lecture date: 2/4/2010
CS 598CSC: Combinatorial Optimization Lecture date: /4/010 Instructor: Chandra Chekuri Scribe: David Morrison Gomory-Hu Trees (The work in this section closely follows [3]) Let G = (V, E) be an undirected
DETERMINANTS IN THE KRONECKER PRODUCT OF MATRICES: THE INCIDENCE MATRIX OF A COMPLETE GRAPH
DETERMINANTS IN THE KRONECKER PRODUCT OF MATRICES: THE INCIDENCE MATRIX OF A COMPLETE GRAPH CHRISTOPHER RH HANUSA AND THOMAS ZASLAVSKY Abstract We investigate the least common multiple of all subdeterminants,
Lecture 15 An Arithmetic Circuit Lowerbound and Flows in Graphs
CSE599s: Extremal Combinatorics November 21, 2011 Lecture 15 An Arithmetic Circuit Lowerbound and Flows in Graphs Lecturer: Anup Rao 1 An Arithmetic Circuit Lower Bound An arithmetic circuit is just like
THE DIMENSION OF A VECTOR SPACE
THE DIMENSION OF A VECTOR SPACE KEITH CONRAD This handout is a supplementary discussion leading up to the definition of dimension and some of its basic properties. Let V be a vector space over a field
BANACH AND HILBERT SPACE REVIEW
BANACH AND HILBET SPACE EVIEW CHISTOPHE HEIL These notes will briefly review some basic concepts related to the theory of Banach and Hilbert spaces. We are not trying to give a complete development, but
Minimally Infeasible Set Partitioning Problems with Balanced Constraints
Minimally Infeasible Set Partitioning Problems with alanced Constraints Michele Conforti, Marco Di Summa, Giacomo Zambelli January, 2005 Revised February, 2006 Abstract We study properties of systems of
Matrix Representations of Linear Transformations and Changes of Coordinates
Matrix Representations of Linear Transformations and Changes of Coordinates 01 Subspaces and Bases 011 Definitions A subspace V of R n is a subset of R n that contains the zero element and is closed under
Mathematical finance and linear programming (optimization)
Mathematical finance and linear programming (optimization) Geir Dahl September 15, 2009 1 Introduction The purpose of this short note is to explain how linear programming (LP) (=linear optimization) may
Notes on Symmetric Matrices
CPSC 536N: Randomized Algorithms 2011-12 Term 2 Notes on Symmetric Matrices Prof. Nick Harvey University of British Columbia 1 Symmetric Matrices We review some basic results concerning symmetric matrices.
Quadratic forms Cochran s theorem, degrees of freedom, and all that
Quadratic forms Cochran s theorem, degrees of freedom, and all that Dr. Frank Wood Frank Wood, [email protected] Linear Regression Models Lecture 1, Slide 1 Why We Care Cochran s theorem tells us
3. Mathematical Induction
3. MATHEMATICAL INDUCTION 83 3. Mathematical Induction 3.1. First Principle of Mathematical Induction. Let P (n) be a predicate with domain of discourse (over) the natural numbers N = {0, 1,,...}. If (1)
Section 6.1 - Inner Products and Norms
Section 6.1 - Inner Products and Norms Definition. Let V be a vector space over F {R, C}. An inner product on V is a function that assigns, to every ordered pair of vectors x and y in V, a scalar in F,
Similarity and Diagonalization. Similar Matrices
MATH022 Linear Algebra Brief lecture notes 48 Similarity and Diagonalization Similar Matrices Let A and B be n n matrices. We say that A is similar to B if there is an invertible n n matrix P such that
Linear Programming for Optimization. Mark A. Schulze, Ph.D. Perceptive Scientific Instruments, Inc.
1. Introduction Linear Programming for Optimization Mark A. Schulze, Ph.D. Perceptive Scientific Instruments, Inc. 1.1 Definition Linear programming is the name of a branch of applied mathematics that
Split Nonthreshold Laplacian Integral Graphs
Split Nonthreshold Laplacian Integral Graphs Stephen Kirkland University of Regina, Canada [email protected] Maria Aguieiras Alvarez de Freitas Federal University of Rio de Janeiro, Brazil [email protected]
Polarization codes and the rate of polarization
Polarization codes and the rate of polarization Erdal Arıkan, Emre Telatar Bilkent U., EPFL Sept 10, 2008 Channel Polarization Given a binary input DMC W, i.i.d. uniformly distributed inputs (X 1,...,
OPTIMAL DESIGN OF DISTRIBUTED SENSOR NETWORKS FOR FIELD RECONSTRUCTION
OPTIMAL DESIGN OF DISTRIBUTED SENSOR NETWORKS FOR FIELD RECONSTRUCTION Sérgio Pequito, Stephen Kruzick, Soummya Kar, José M. F. Moura, A. Pedro Aguiar Department of Electrical and Computer Engineering
The Characteristic Polynomial
Physics 116A Winter 2011 The Characteristic Polynomial 1 Coefficients of the characteristic polynomial Consider the eigenvalue problem for an n n matrix A, A v = λ v, v 0 (1) The solution to this problem
Linear Programming. March 14, 2014
Linear Programming March 1, 01 Parts of this introduction to linear programming were adapted from Chapter 9 of Introduction to Algorithms, Second Edition, by Cormen, Leiserson, Rivest and Stein [1]. 1
4.5 Linear Dependence and Linear Independence
4.5 Linear Dependence and Linear Independence 267 32. {v 1, v 2 }, where v 1, v 2 are collinear vectors in R 3. 33. Prove that if S and S are subsets of a vector space V such that S is a subset of S, then
Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model
Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model 1 September 004 A. Introduction and assumptions The classical normal linear regression model can be written
a 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2.
Chapter 1 LINEAR EQUATIONS 1.1 Introduction to linear equations A linear equation in n unknowns x 1, x,, x n is an equation of the form a 1 x 1 + a x + + a n x n = b, where a 1, a,..., a n, b are given
Discrete Mathematics & Mathematical Reasoning Chapter 10: Graphs
Discrete Mathematics & Mathematical Reasoning Chapter 10: Graphs Kousha Etessami U. of Edinburgh, UK Kousha Etessami (U. of Edinburgh, UK) Discrete Mathematics (Chapter 6) 1 / 13 Overview Graphs and Graph
Let H and J be as in the above lemma. The result of the lemma shows that the integral
Let and be as in the above lemma. The result of the lemma shows that the integral ( f(x, y)dy) dx is well defined; we denote it by f(x, y)dydx. By symmetry, also the integral ( f(x, y)dx) dy is well defined;
Secure Network Coding for Wiretap Networks of Type II
1 Secure Network Coding for Wiretap Networks of Type II Salim El Rouayheb, Emina Soljanin, Alex Sprintson Abstract We consider the problem of securing a multicast network against a wiretapper that can
Markov random fields and Gibbs measures
Chapter Markov random fields and Gibbs measures 1. Conditional independence Suppose X i is a random element of (X i, B i ), for i = 1, 2, 3, with all X i defined on the same probability space (.F, P).
Functional-Repair-by-Transfer Regenerating Codes
Functional-Repair-by-Transfer Regenerating Codes Kenneth W Shum and Yuchong Hu Abstract In a distributed storage system a data file is distributed to several storage nodes such that the original file can
Discrete Mathematics and Probability Theory Fall 2009 Satish Rao, David Tse Note 2
CS 70 Discrete Mathematics and Probability Theory Fall 2009 Satish Rao, David Tse Note 2 Proofs Intuitively, the concept of proof should already be familiar We all like to assert things, and few of us
Adaptive Online Gradient Descent
Adaptive Online Gradient Descent Peter L Bartlett Division of Computer Science Department of Statistics UC Berkeley Berkeley, CA 94709 bartlett@csberkeleyedu Elad Hazan IBM Almaden Research Center 650
A linear combination is a sum of scalars times quantities. Such expressions arise quite frequently and have the form
Section 1.3 Matrix Products A linear combination is a sum of scalars times quantities. Such expressions arise quite frequently and have the form (scalar #1)(quantity #1) + (scalar #2)(quantity #2) +...
15.062 Data Mining: Algorithms and Applications Matrix Math Review
.6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop
Vector and Matrix Norms
Chapter 1 Vector and Matrix Norms 11 Vector Spaces Let F be a field (such as the real numbers, R, or complex numbers, C) with elements called scalars A Vector Space, V, over the field F is a non-empty
Bindel, Spring 2012 Intro to Scientific Computing (CS 3220) Week 3: Wednesday, Feb 8
Spaces and bases Week 3: Wednesday, Feb 8 I have two favorite vector spaces 1 : R n and the space P d of polynomials of degree at most d. For R n, we have a canonical basis: R n = span{e 1, e 2,..., e
Solving Systems of Linear Equations
LECTURE 5 Solving Systems of Linear Equations Recall that we introduced the notion of matrices as a way of standardizing the expression of systems of linear equations In today s lecture I shall show how
Notes on Complexity Theory Last updated: August, 2011. Lecture 1
Notes on Complexity Theory Last updated: August, 2011 Jonathan Katz Lecture 1 1 Turing Machines I assume that most students have encountered Turing machines before. (Students who have not may want to look
This asserts two sets are equal iff they have the same elements, that is, a set is determined by its elements.
3. Axioms of Set theory Before presenting the axioms of set theory, we first make a few basic comments about the relevant first order logic. We will give a somewhat more detailed discussion later, but
THE NUMBER OF GRAPHS AND A RANDOM GRAPH WITH A GIVEN DEGREE SEQUENCE. Alexander Barvinok
THE NUMBER OF GRAPHS AND A RANDOM GRAPH WITH A GIVEN DEGREE SEQUENCE Alexer Barvinok Papers are available at http://www.math.lsa.umich.edu/ barvinok/papers.html This is a joint work with J.A. Hartigan
Lecture 1: Course overview, circuits, and formulas
Lecture 1: Course overview, circuits, and formulas Topics in Complexity Theory and Pseudorandomness (Spring 2013) Rutgers University Swastik Kopparty Scribes: John Kim, Ben Lund 1 Course Information Swastik
MOST error-correcting codes are designed for the equal
IEEE TRANSACTIONS ON COMMUNICATIONS, VOL. 55, NO. 3, MARCH 2007 387 Transactions Letters Unequal Error Protection Using Partially Regular LDPC Codes Nazanin Rahnavard, Member, IEEE, Hossein Pishro-Nik,
NOTES ON LINEAR TRANSFORMATIONS
NOTES ON LINEAR TRANSFORMATIONS Definition 1. Let V and W be vector spaces. A function T : V W is a linear transformation from V to W if the following two properties hold. i T v + v = T v + T v for all
IN THIS PAPER, we study the delay and capacity trade-offs
IEEE/ACM TRANSACTIONS ON NETWORKING, VOL. 15, NO. 5, OCTOBER 2007 981 Delay and Capacity Trade-Offs in Mobile Ad Hoc Networks: A Global Perspective Gaurav Sharma, Ravi Mazumdar, Fellow, IEEE, and Ness
The chromatic spectrum of mixed hypergraphs
The chromatic spectrum of mixed hypergraphs Tao Jiang, Dhruv Mubayi, Zsolt Tuza, Vitaly Voloshin, Douglas B. West March 30, 2003 Abstract A mixed hypergraph is a triple H = (X, C, D), where X is the vertex
ON THE DEGREES OF FREEDOM OF SIGNALS ON GRAPHS. Mikhail Tsitsvero and Sergio Barbarossa
ON THE DEGREES OF FREEDOM OF SIGNALS ON GRAPHS Mikhail Tsitsvero and Sergio Barbarossa Sapienza Univ. of Rome, DIET Dept., Via Eudossiana 18, 00184 Rome, Italy E-mail: [email protected], [email protected]
Chapter 3. Distribution Problems. 3.1 The idea of a distribution. 3.1.1 The twenty-fold way
Chapter 3 Distribution Problems 3.1 The idea of a distribution Many of the problems we solved in Chapter 1 may be thought of as problems of distributing objects (such as pieces of fruit or ping-pong balls)
MATH 4330/5330, Fourier Analysis Section 11, The Discrete Fourier Transform
MATH 433/533, Fourier Analysis Section 11, The Discrete Fourier Transform Now, instead of considering functions defined on a continuous domain, like the interval [, 1) or the whole real line R, we wish
COMP 250 Fall 2012 lecture 2 binary representations Sept. 11, 2012
Binary numbers The reason humans represent numbers using decimal (the ten digits from 0,1,... 9) is that we have ten fingers. There is no other reason than that. There is nothing special otherwise about
1 The Line vs Point Test
6.875 PCP and Hardness of Approximation MIT, Fall 2010 Lecture 5: Low Degree Testing Lecturer: Dana Moshkovitz Scribe: Gregory Minton and Dana Moshkovitz Having seen a probabilistic verifier for linearity
Approximation Algorithms
Approximation Algorithms or: How I Learned to Stop Worrying and Deal with NP-Completeness Ong Jit Sheng, Jonathan (A0073924B) March, 2012 Overview Key Results (I) General techniques: Greedy algorithms
Tree-representation of set families and applications to combinatorial decompositions
Tree-representation of set families and applications to combinatorial decompositions Binh-Minh Bui-Xuan a, Michel Habib b Michaël Rao c a Department of Informatics, University of Bergen, Norway. [email protected]
Chapter 17. Orthogonal Matrices and Symmetries of Space
Chapter 17. Orthogonal Matrices and Symmetries of Space Take a random matrix, say 1 3 A = 4 5 6, 7 8 9 and compare the lengths of e 1 and Ae 1. The vector e 1 has length 1, while Ae 1 = (1, 4, 7) has length
FINITE DIMENSIONAL ORDERED VECTOR SPACES WITH RIESZ INTERPOLATION AND EFFROS-SHEN S UNIMODULARITY CONJECTURE AARON TIKUISIS
FINITE DIMENSIONAL ORDERED VECTOR SPACES WITH RIESZ INTERPOLATION AND EFFROS-SHEN S UNIMODULARITY CONJECTURE AARON TIKUISIS Abstract. It is shown that, for any field F R, any ordered vector space structure
RN-coding of Numbers: New Insights and Some Applications
RN-coding of Numbers: New Insights and Some Applications Peter Kornerup Dept. of Mathematics and Computer Science SDU, Odense, Denmark & Jean-Michel Muller LIP/Arénaire (CRNS-ENS Lyon-INRIA-UCBL) Lyon,
Lecture 16 : Relations and Functions DRAFT
CS/Math 240: Introduction to Discrete Mathematics 3/29/2011 Lecture 16 : Relations and Functions Instructor: Dieter van Melkebeek Scribe: Dalibor Zelený DRAFT In Lecture 3, we described a correspondence
6.2 Permutations continued
6.2 Permutations continued Theorem A permutation on a finite set A is either a cycle or can be expressed as a product (composition of disjoint cycles. Proof is by (strong induction on the number, r, of
