Compressed Sensing with Non-Gaussian Noise and Partial Support Information



Similar documents
When is missing data recoverable?

Clarify Some Issues on the Sparse Bayesian Learning for Sparse Signal Recovery

SECRET sharing schemes were introduced by Blakley [5]

A Negative Result Concerning Explicit Matrices With The Restricted Isometry Property

Subspace Pursuit for Compressive Sensing: Closing the Gap Between Performance and Complexity

Stable Signal Recovery from Incomplete and Inaccurate Measurements

Sparse LMS via Online Linearized Bregman Iteration

Part II Redundant Dictionaries and Pursuit Algorithms

THE Walsh Hadamard transform (WHT) and discrete

Stable Signal Recovery from Incomplete and Inaccurate Measurements

MOST error-correcting codes are designed for the equal

Functional-Repair-by-Transfer Regenerating Codes

The p-norm generalization of the LMS algorithm for adaptive filtering

An Improved Reconstruction methods of Compressive Sensing Data Recovery in Wireless Sensor Networks

THE problem of computing sparse solutions (i.e., solutions

ON THE DEGREES OF FREEDOM OF SIGNALS ON GRAPHS. Mikhail Tsitsvero and Sergio Barbarossa

By choosing to view this document, you agree to all provisions of the copyright laws protecting it.

Sparse recovery and compressed sensing in inverse problems

Continued Fractions and the Euclidean Algorithm

Sparsity-promoting recovery from simultaneous data: a compressive sensing approach

Linear Codes. Chapter Basics

Parallel Data Selection Based on Neurodynamic Optimization in the Era of Big Data

Noisy and Missing Data Regression: Distribution-Oblivious Support Recovery

Lecture Topic: Low-Rank Approximations

JPEG compression of monochrome 2D-barcode images using DCT coefficient distributions

Optimal Index Codes for a Class of Multicast Networks with Receiver Side Information

Efficient Data Recovery scheme in PTS-Based OFDM systems with MATRIX Formulation

On the representability of the bi-uniform matroid

Maximum Likelihood Estimation of ADC Parameters from Sine Wave Test Data. László Balogh, Balázs Fodor, Attila Sárhegyi, and István Kollár

Adaptive Linear Programming Decoding

Analysis of Mean-Square Error and Transient Speed of the LMS Adaptive Algorithm

I. INTRODUCTION. of the biometric measurements is stored in the database

Enhancing the SNR of the Fiber Optic Rotation Sensor using the LMS Algorithm

Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches

Variational approach to restore point-like and curve-like singularities in imaging

Practical Guide to the Simplex Method of Linear Programming

AT&T Global Network Client for Windows Product Support Matrix January 29, 2015

Sublinear Algorithms for Big Data. Part 4: Random Topics

NMR Measurement of T1-T2 Spectra with Partial Measurements using Compressive Sensing

Similarity and Diagonalization. Similar Matrices

A Network Flow Approach in Cloud Computing

Evolutionary denoising based on an estimation of Hölder exponents with oscillations.

RECENTLY, there has been a great deal of interest in

Inner Product Spaces

Aalborg Universitet. Published in: I E E E Signal Processing Letters. DOI (link to publication from Publisher): /LSP.2014.

An Adaptive Decoding Algorithm of LDPC Codes over the Binary Erasure Channel. Gou HOSOYA, Hideki YAGI, Toshiyasu MATSUSHIMA, and Shigeichi HIRASAWA

On an algorithm for classification of binary self-dual codes with minimum distance four

Clustering & Visualization

Combating Anti-forensics of Jpeg Compression

Duality of linear conic problems

Research Article A Method of Data Recovery Based on Compressive Sensing in Wireless Structural Health Monitoring

IN THIS paper, we address the classic image denoising

Adaptive Online Gradient Descent

How To Understand The Problem Of Decoding By Linear Programming

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION

Lecture 15 An Arithmetic Circuit Lowerbound and Flows in Graphs

1 if 1 x 0 1 if 0 x 1

Orthogonal Projections and Orthonormal Bases

IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 51, NO. 7, JULY G. George Yin, Fellow, IEEE, and Vikram Krishnamurthy, Fellow, IEEE

THE problems of characterizing the fundamental limits

Least Squares Estimation

ADVANCED APPLICATIONS OF ELECTRICAL ENGINEERING

Component Ordering in Independent Component Analysis Based on Data Power

Wavelet analysis. Wavelet requirements. Example signals. Stationary signal 2 Hz + 10 Hz + 20Hz. Zero mean, oscillatory (wave) Fast decay (let)

Nonlinear Iterative Partial Least Squares Method

Solving polynomial least squares problems via semidefinite programming relaxations

The Advantages and Disadvantages of Online Linear Optimization

The Dantzig selector: statistical estimation when p is much larger than n

ADAPTIVE ALGORITHMS FOR ACOUSTIC ECHO CANCELLATION IN SPEECH PROCESSING

FUZZY CLUSTERING ANALYSIS OF DATA MINING: APPLICATION TO AN ACCIDENT MINING SYSTEM

Analysis of Load Frequency Control Performance Assessment Criteria

Communication on the Grassmann Manifold: A Geometric Approach to the Noncoherent Multiple-Antenna Channel

How To Fix Out Of Focus And Blur Images With A Dynamic Template Matching Algorithm

Two classes of ternary codes and their weight distributions

Applications to Data Smoothing and Image Processing I

The Steepest Descent Algorithm for Unconstrained Optimization and a Bisection Line-search Method

MixedÀ¾ нOptimization Problem via Lagrange Multiplier Theory

Stability of the LMS Adaptive Filter by Means of a State Equation

CS3220 Lecture Notes: QR factorization and orthogonal transformations

CONSTANT-SIGN SOLUTIONS FOR A NONLINEAR NEUMANN PROBLEM INVOLVING THE DISCRETE p-laplacian. Pasquale Candito and Giuseppina D Aguí

Victoria Kostina Curriculum Vitae - September 6, 2015 Page 1 of 5. Victoria Kostina

Adaptive Equalization of binary encoded signals Using LMS Algorithm

IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 52, NO. 4, APRIL Compressed Sensing. David L. Donoho, Member, IEEE

(2) (3) (4) (5) 3 J. M. Whittaker, Interpolatory Function Theory, Cambridge Tracts

The Image Deblurring Problem

Bag of Pursuits and Neural Gas for Improved Sparse Coding

JPEG Image Compression by Using DCT

1 VECTOR SPACES AND SUBSPACES

Introduction to General and Generalized Linear Models

A Numerical Study on the Wiretap Network with a Simple Network Topology

Lattice-Based Threshold-Changeability for Standard Shamir Secret-Sharing Schemes

UNIVERSAL SPEECH MODELS FOR SPEAKER INDEPENDENT SINGLE CHANNEL SOURCE SEPARATION

LECTURE 5: DUALITY AND SENSITIVITY ANALYSIS. 1. Dual linear program 2. Duality theory 3. Sensitivity analysis 4. Dual simplex method

DYNAMIC RANGE IMPROVEMENT THROUGH MULTIPLE EXPOSURES. Mark A. Robertson, Sean Borman, and Robert L. Stevenson

IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, ACCEPTED FOR PUBLICATION 1

Statistical machine learning, high dimension and big data

Solution of Linear Systems

THE great leap forward in digital processing over the

Notes on Factoring. MA 206 Kurt Bryan

Binary Image Reconstruction

Transcription:

IEEE SIGNAL PROCESSING LETTERS, VOL. 22, NO. 10, OCTOBER 2015 1703 Compressed Sensing with Non-Gaussian Noise Partial Support Information Ahmad Abou Saleh, Fady Alajaji, Wai-Yip Chan Abstract We study the problem of recovering sparse compressible signals using a weighted minimization with from noisy compressed sensing measurements when part of the support is known a priori. To better model different types of non- Gaussian (bounded) noise, the minimization program is subject to a data-fidelity constraint expressed as the norm of the residual error. We show theoretically that the reconstruction error of this optimization is bounded (stable) if the sensing matrix satisfies an extended restricted isometry property. Numerical results show that the proposed method, which extends the range of comparing with previous works, outperforms other noise-aware basis pursuit programs. For,sincetheoptimization is not convex, we use a variant of an iterative reweighted algorithm for computing a local minimum. Index Terms Compressed sensing, denoising, nonconvex optimization, sparsity, weighted minimization. I. INTRODUCTION T HE theory of compressed sensing (CS) is to reliably recover high dimensional sparse or compressible signals from a significantly fewer number of linear measurements than the signal space dimension [1], [2]. Consider a signal of dimension to be sparse in some orthonormal basis ; in other words,, is a sparse vector, i.e., there are only nonzero elements in.in CS, the encoder records linear measurements given by, is a measurement matrix that satisfies the restricted isometry property (RIP); satisfies the RIP of order if there exists such that, for any sparse,, is the RIP constant of order denotes the norm. Roughly speaking, the RIP condition states that any (sufficient) small subset of the columns of the sensing matrix nearly forms an orthogonal basis. The condition is satisfied when the elements of the matrix are samples from an independent identically distributed (i.i.d.) Gaussian process is of the order of. In this case, recovery of Manuscript received January 22, 2015; accepted April 20, 2015. Date of publication April 27, 2015; date of current version May 01, 2015. This work was supported in part by the NSERC of Canada. The associate editor coordinating the review of this manuscript approving it for publication was Prof. Alex Dimakis. A. Abou Saleh W-Y. Chan are with the Department of Electrical Computer Engineering, Queen s University, Kingston, ON K7L 3N6 Canada (e-mail: ahmad.abou.saleh@queensu.ca; chan@queensu.ca). F. Alajaji is with the Department of Mathematics Statistics the Department of Electrical Computer Engineering, Queen s University, Kingston, ON K7L 3N6 Canada (e-mail: fady@mast.queensu.ca). Color versions of one or more of the figures in this paper are available online at http://ieeexplore.ieee.org. Digital Object Identifier 10.1109/LSP.2015.2426654 from works by solving the convex basis pursuit (BP) program:, denotes the norm ( ) is the transpose operator [3]. In practice, the collected measurements are disturbed by noise (i.e., ). Using a noise-aware version of BP, the signal is decoded by solving the convex basis pursuit denoising (BPDN) program:, bounds the total amount of noise in the measurements. In many applications, it is possible to obtain some estimate about the support of the signal. To incorporate such prior information, several modified versions of the BPDN were studied in [4] [6] based on a weighted minimization approach with zero weights on the known support. In [7], the author extends the work of [5] to the case of compressible signals noisy measurements. In [8], an RIP-based condition for recovering compressively sampled signals using partial support information is derived. On the other h, several works have attempted to close the gap in performance from using instead of (combinatorial problem) minimization by solving a nonconvex ( ) minimization. In [9] [11], the authors provide sufficient conditions for recoverability of sparse/compressible signals from noisy measurements via such minimization. In [12], the authors study weighted minimization with zero weights on the known support. In most previous works, the CS recovery program uses an norm fidelity constraint. From a Bayesian stpoint, the norm used in the fidelity term (of the BPDN) corresponds to the conditional log-likelihood associated with additive white Gaussian noise. However, the noise in the disturbed measurement might not be Gaussian. As an example, quantization distortion is highly non-gaussian is essentially bounded by the quantization bin widths. This motivates the need for a new CS decoder that better models different types of noise. The authors in [13] adapt BPDN to integrate the quantization distortion constraint. In [14], the authors study a new class of optimization problems coined basis pursuit dequantizing (BPDQ) that models more faithfully different types of noise than BPDN; this new class of optimization proceeds by minimizing the sparsity of the signal subject to a data-fidelity constraint expressed in the norm of the residual error instead of the norm as in BPDN; the case of isstudiedin[15].the BPDQ of moment,,is bounds the measurement noise. The authors in [14] study the problem in (1) show that this approach gives some gain over BPDN. However, no prior information on the unknown signal is assumed in [14]. Motivated by (1) 1070-9908 2015 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_stards/publications/rights/index.html for more information.

1704 IEEE SIGNAL PROCESSING LETTERS, VOL. 22, NO. 10, OCTOBER 2015 the fact that the CS algorithm may have prior knowledge about the signal that disturbance in the measurements might come from quantization distortion, we study weighted minimization subject to an norm fidelity constraint for compressively sampled signals with (not fully accurate) partial support information; we develop RIP-based guarantee conditions for stable recovery. This generalizes the problems in [8], [14] [16] to a wider range of.wealso discuss the implementation aspect for the algorithm for. In this case, the problem is no longer convex; we resort to the use of an iterative reweighted convex problem to solve the original nonconvex optimization. II. WEIGHTED BASIS PURSUIT DEQUANTIZING In this section, we present a CS recovery method that incorporates partial knowledge about the sparsity support models more faithfully different types of measurement noise. Next we assume a canonical basis ( is the identity matrix). Definition 1: satisfies the extended of order if,for all sparse signals,a an constant [14]. Note that is equal to 1 for. Consider a compressible signal.let be sparse the nonzero coefficients of are equal to the largest-in-magnitude coefficients of ; we refer to as the best term approximation of.let be the support of (the locations of the nonzero coefficients),. Let the support estimate be a subset of with cardinality, for some, be a sub-vector of with coefficient. The generalized recovery process, coined the weighted ( ), is given by,,,,, bounds the total amount of noise in the measurements. Next, we show that is stable if satisfies an. Theorem 1: Define so that,. Suppose that there exists an, that satisfies with with given by of. Then the solution to (2) satisfies (2) (3) (4) (5).Thefirst term in the right h side of (5) depends on the measurement matrix the measurement noise; the second term, which also depends on the measurement matrix, comes from the compressibility of the signal. The proof of Theorem 1 is presented in the Appendix; the proof uses a combination of techniques used to prove [2, Theorem 2], [14, Theorem 2] [8, Theorem 3]. Remark 1: For,, Theorem 1 reduces to Theorem 2 in [14]. For,,the sufficient condition the bound on the reconstruction error given by Theorem 1 is identical to that of Theorem 2 in [14]; this means that choosing in (2) gives the best performance for. Next, we discuss some implementation aspects of the optimization problem given by (2). For the case of,the program is convex can be solved using the toolboxin[17].for, however, the optimization is no longer convex. Algorithms that have been used to solve such optimization include the projected gradient method [9], [16], the iterative reweighed method [18] the iterative reweighted least squares method [19], [20]. We herein use a variant of the iteratively reweighted algorithm to compute a local minimum. The approach is to replace the weighted objective function in (2) by a weighted norm as follows ; the weight is updated at each iteration is a regularization factor. This process is repeated with decreasing value of until convergence. Since,allowing to be zero, the weights in (7) are undefined whenever the value of the signal is zero ( ). To overcome this, we use the approach of regularizing the optimization problem by incorporating a small. Algorithm 1 is stated as follows 1) Set, start with an initial solution. 2) Set ;solvefor using (7). 3) Decrease as follow:. 4) Evaluate the convergence function.if is less than some positive threshold or stop iterating. Else go to step 2. In our simulation, we solve (7) using the toolbox in [17]; we set,, to be the solution of the problem for. Remark 2: We also implemented a variant of the iterative reweighted method in [18] to solve the nonconvex problem; we replace the objective function in (7) in Algorithm 1 with a weighted function,. Simulation results show that this method gives similar performance as the reweighted norm method. III. NUMERICAL RESULTS We consider to be sparse in the canonical basis the nonzero elements in are chosen to be i.i.d. zero-mean (6) (7)

ABOU SALEH et al.: COMPRESSED SENSING WITH NON-GAUSSIAN NOISE AND PARTIAL SUPPORT INFORMATION 1705 Fig. 1. Performance of with uniform noise for ; the support knowledge has with 80% accuracy rate. The noise is uniform on. Gaussian with unit variance the signal support is uniformly distributed. The support knowledge is uniformly chosen with accuracy rate.fig.1showsthe performance versus the number of measurements for an accuracy rate of 80% (i.e., ). We assume that the noise is uniform on the interval choose a weight parameter of in (2). We notice that the ( ), solved using Algorithm 1, gives the best performance. Comparing to (i.e., WBPDN), we notice that the SDR gap increases with.this is from the fact that when is low, we operate in the region the constraint on the constant (3) in Theorem 1 is barely met hence the benefit of using the norm constraint over the is not as high as larger. Although Fig. 1 aims to show the performance of different recovery algorithms over, it is worth mentioning that the value of used is not the optimal one. Fig. 2 shows the SDR performance of ( )versus for different accuracy rates uniform noise on. We remark that for low accuracy,itis beneficial to increase ; for high accuracy, however, using a lower gives a better performance. Comparing the gap between over several accuracy rates, we notice that this gap gets wider for higher accuracy. This can be explained from the fact that for higher accuracy, the cardinality of the unknown sparsity pattern is lower hence less measurements are needed to satisfy the constraint of Theorem 1; as a result, the benefit of using norm is more visible. Note that for the extreme case of, the SDR gap between is small; this is from the fact that no side information is used at the decoder since is relatively small compared to. Fig. 3 shows the SDR performance as a function of for different values of accuracy weight. We can notice that using a larger support estimate favours better performance for most cases. The only case increasing decreases the SDR performance is when choosing (very small) for low accuracy rate. This can be explained from the fact that for low accuracy, increasing leads to a relatively more increase in the amount of wrong indices than true ones; as a result, choosing (i.e., favouring the indices in ) leads to a decrease in the performance. Moreover, the drastic increase Fig. 2. Performance of the (for )versus for,. The support knowledge has with 80% accuracy rate the noise is uniform on. Fig. 3. SDR Performance versus for uniform noise distribution. The solid curves correspond to, while the dashed ones are for WBPDN. The signal has. The number of measurements is ; Fig. 3 (a) is for while (b) for. in the performance for in Fig. 3(a) is from the fact that in this case contains all indices of the true support. Fig. 4 shows the SDR performance of ( ) as a function of for different measurement noise models; our aim is to show the importance of with respect to the noise model. As expected, for Gaussian noise, the best performance is shown for. For uniform noise, we notice that the best is 3. We also study the performance of for the Beta distribution (which subsumes the uniform distribution as a special case), with domain [0,1] a probability distribution function (pdf) given by, is a normalization constant, denote the distribution parameters. We experiment with different parameters for the Beta distribution, we notice that choosing higher gives better performance than choosing for several cases. More precisely, when the pdf is symmetric behaves more like a uniform distribution for most of its range ( ), the best is greater than 2. For, the distribution has a similar shape as a limited one sided-gaussian; this makes to be the best choice.

1706 IEEE SIGNAL PROCESSING LETTERS, VOL. 22, NO. 10, OCTOBER 2015 Lemma 1: [14] Let, be sparse sparse, respectively, with disjoint supports. If is of order with constant of orders with constants, respectively, then (11) is as defined in (4). From the definition of, similar to [14], we can write Fig. 4. Performance of the versus for zero-mean Beta uniform noise distribution. The signal has ; the support knowledge has with 80% accuracy rate. Note that every curve corresponds to a different noise level. Moreover, choosing leads to similar conclusions; reducing, in principle, gives a better performance. It is worth mentioning that in Fig. 4, the curves do not have the same noise levels; changing the noise level does not affect the curve trend (the curve simply shifts vertically). For future work, we will investigate different penalty functions to increase the recovery tolerance against large (unbounded) noise [21], [22]. Taking the th power using the fact that for any,,weobtain (12) (13) Using the fact that are feasible solutions, we have ; applying Holder s inequality with parameters,weget APPENDIX Proof of Theorem 1: Let be a minimizer of the weighted minimization problem subject to an norm as given in (2). As a result, we have.bythe choice of weights in (2) using the forward the reverse triangle inequalities on, we obtain after some manipulations (similar to [8]) Similar to [2], we next sort the coefficients of partition into finite disjoint subsets, each of size,.thismeansthat indexes the largest in magnitude coefficients of, indexes the second largest in magnitude coefficients of so on. As a result. Moreover, by denoting we can write (8) (9) (14) the last inequality follows from the definition of bynotingthat is an sparse vector. Using Lemma 1 the fact that is sparse is sparse for, we can write the following (15). Combining (13), (14) (15), can be upper bounded. By lower bounding using applying it to the upper bound found on, we obtain after some manipulations (16) Using (10) in (16), the fact that,,,wederive the following after some manipulations Combining (8) (9), we obtain the following (10) (17) Using (17) (10), we bound ; this completes the proof provided that the denominator in (17) is positive which reduces to (3).

ABOU SALEH et al.: COMPRESSED SENSING WITH NON-GAUSSIAN NOISE AND PARTIAL SUPPORT INFORMATION 1707 REFERENCES [1] D. Donoho, Compressed sensing, IEEE Trans. Inf. Theory, vol.52, no. 4, pp. 1289 1306, Apr. 2006. [2]E.Ces,J.Romberg,T.Tao, Stablesignalrecoveryfrom incomplete inaccurate measurements, Commun. Pure Appl. Math., vol. 59, no. 8, pp. 1207 1223, 2006. [3] S. S. Chen, D. L. Donoho, M. A. Saunders, Atomic decomposition by basis pursuit, SIAM J. Sci. Comput., vol. 20, no. 1, pp. 31 61, 1998. [4] R.V.Borries,C.Miosso,C.Potes, Compressed sensing using prior information, in IEEE Int. Workshop on Computational Advances in Multi-Sensor Adaptive Process, St. Thomas,, Virgin Isls, Dec. 2007, pp. 121 124. [5] N. Vaswani W. Lu, Modified-CS: Modifying compressive sensing for problems with partially known support, IEEE Trans. Signal Process., vol. 58, no. 9, pp. 4595 4607, Sep. 2010. [6] M. A. Khajehnejad, W. Xu, A. S. Avestimehr, B. Hassibi, Weighted l1 minimization for sparse recovery with prior information, in IEEE Int. Symp. Inform. Theory, Seoul, Korea, Jun. 2009. [7] L. Jacques, A short note on compressed sensing with partially known signal support, Signal Process., vol. 90, no. 12, pp. 3308 3312, Dec. 2010. [8] M. P. Friedler, H. Mansour, R. Saab, O. Yilmaz, Recovering compressively sampled signals using partial support information, IEEE Trans. Inf. Theory, vol. 58, no. 2, pp. 1122 1134, Feb. 2012. [9] R. Chartr, Exact reconstruction of sparse signals via nonconvex minimization, IEEE Signal Process. Lett., vol. 14, no. 10, pp. 707 710, Oct. 2007. [10] R. Chartr V. Staneva, Restricted isometry properties nonconvex compressive sensing, Inv. Probl., vol. 24, no. 035020, pp. 1 14, May 2008. [11] R. Saab, R. Chartr, O. Yilmaz, Stable sparse approximations via nonconvex optimization, in IEEE Int. Conf. on Acoustics, Speech Signal Process, Las Vegas, NV, Apr. 2008. [12] T. Ince, A. Nacaroglu, N. Watsuji, Nonconvex compressed sensing with partially known signal support, Signal Process., vol. 93, no. 1, pp. 338 344, Jan. 2013. [13] W. Dai O. Milenkovic, Information theoretical algorithmic approaches to quantized compressive sensing, IEEE Trans. Commun., vol. 59, no. 7, pp. 1857 1866, Jul. 2011. [14] L. Jacques, D. K. Hammond, J. M. Fadili, Dequantizing compressed sensing: When oversampling non-gaussian constraints combine, IEEE Trans. Inf. Theory, vol. 57, no. 1, pp. 559 571, Jan. 2011. [15] J. Fuchs, Fast implementation of a regularized sparse representations algorithm, in IEEE Int. Conf. on Acoustics, Speech Signal Process, Taipei, Taiwan, Apr. 2009. [16] N. Ghadermarzy, H. Mansour, O. Yilmaz, Non-convex compressed sensing using partial support information, Sampling Theory Signal Image Process., vol. 13, no. 3, pp. 251 272, 2014. [17] M. Grant S. Boyd, CVX: Matlab software for disciplined convex programming, version 2.1, Mar. 2014 [Online]. Available: http://cvxr. com/cvx [18] X. Chen W. Zhou, Convergence of reweighted minimization algorithms unique solution of truncated minimization, Dept. Appl. Math., Hong Kong Polytechnic Univ., 2010. [19] B. D. Rao K. Kreutz-Delgado, An affine scaling methodology for best basis selection, IEEE Trans. Signal Process., vol. 47, no. 1, pp. 187 200, Jan. 1999. [20] R. Chartr W. Yin, Iteratively reweighted algorithms for compressive sensing, in IEEE Int. Conf. on Acoustics, Speech Signal Process, Las Vegas, Nevada, Apr. 2008. [21] G. R. Arce, D. Otero, A. B. Ramirez, J.-L. Paredes, Reconstruction of sparse signal from dimensionality-reduced Cauchy rom projections, in IEEE Int. Conf. on Acoustics, Speech Signal Process, Dallas, Texas, Mar. 2010. [22] R. E. Carrillo, K. E. Barner, T. C. Aysal, Robust sampling reconstruction methods for sparse signals in the presence of impulsive noise, IEEE J. Sel. Topics Signal Process., vol. 4, no. 2, pp. 392 408, Apr. 2010.