HowHow To Choose A Good Stock Broker



Similar documents
Communication on the Grassmann Manifold: A Geometric Approach to the Noncoherent Multiple-Antenna Channel

Statistical Machine Learning

ADVANCED APPLICATIONS OF ELECTRICAL ENGINEERING

Capacity Limits of MIMO Channels

A Model of Optimum Tariff in Vehicle Fleet Insurance

Introduction to General and Generalized Linear Models

Component Ordering in Independent Component Analysis Based on Data Power

Joint Optimal Pilot Placement and Space Frequency (SF) Code Design for MIMO-OFDM Systems

Kristine L. Bell and Harry L. Van Trees. Center of Excellence in C 3 I George Mason University Fairfax, VA , USA kbell@gmu.edu, hlv@gmu.

MIMO CHANNEL CAPACITY

Eigenvalues, Eigenvectors, Matrix Factoring, and Principal Components

Signal Detection. Outline. Detection Theory. Example Applications of Detection Theory

By choosing to view this document, you agree to all provisions of the copyright laws protecting it.

8 MIMO II: capacity and multiplexing

Enhancing the SNR of the Fiber Optic Rotation Sensor using the LMS Algorithm

TCOM 370 NOTES 99-4 BANDWIDTH, FREQUENCY RESPONSE, AND CAPACITY OF COMMUNICATION LINKS

Least Squares Estimation

IN current film media, the increase in areal density has

Logistic Regression. Jia Li. Department of Statistics The Pennsylvania State University. Logistic Regression

Multivariate Normal Distribution

Lecture 8: Signal Detection and Noise Assumption

Lecture 3: Linear methods for classification

Implementation of Digital Signal Processing: Some Background on GFSK Modulation

Kalman Filter Applied to a Active Queue Management Problem

CONTROL SYSTEMS, ROBOTICS, AND AUTOMATION - Vol. V - Relations Between Time Domain and Frequency Domain Prediction Error Methods - Tomas McKelvey

Least-Squares Intersection of Lines

Nonparametric adaptive age replacement with a one-cycle criterion

Data Mining: Algorithms and Applications Matrix Math Review

Multivariate Analysis of Variance (MANOVA): I. Theory

October 3rd, Linear Algebra & Properties of the Covariance Matrix

Revenue Management with Correlated Demand Forecasting

Probability Hypothesis Density filter versus Multiple Hypothesis Tracking

NOTES ON LINEAR TRANSFORMATIONS

Simple Linear Regression Inference

Exact Inference for Gaussian Process Regression in case of Big Data with the Cartesian Product Structure

1 Portfolio mean and variance

IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 55, NO. 1, JANUARY

Chapter 1 Introduction

Introduction to Matrix Algebra

Data Mining and Data Warehousing. Henryk Maciejewski. Data Mining Predictive modelling: regression

Chapter 4: Vector Autoregressive Models

Fitting Subject-specific Curves to Grouped Longitudinal Data

Diversity and Multiplexing: A Fundamental Tradeoff in Multiple-Antenna Channels

Linear Codes. Chapter Basics

Statistical Machine Learning from Data

Please follow the directions once you locate the Stata software in your computer. Room 114 (Business Lab) has computers with Stata software

Mathematics Course 111: Algebra I Part IV: Vector Spaces

PHASE ESTIMATION ALGORITHM FOR FREQUENCY HOPPED BINARY PSK AND DPSK WAVEFORMS WITH SMALL NUMBER OF REFERENCE SYMBOLS

Coding and decoding with convolutional codes. The Viterbi Algor

5 Signal Design for Bandlimited Channels

Time Series Analysis

A Cold Startup Strategy for Blind Adaptive M-PPM Equalization

Advanced Signal Processing and Digital Noise Reduction

How Far from Kronecker can a MIMO Channel be? Does it Matter?

Cyber-Security Analysis of State Estimators in Power Systems

Interference Alignment and the Degrees of Freedom of Wireless X Networks

24. The Branch and Bound Method

Efficient Data Recovery scheme in PTS-Based OFDM systems with MATRIX Formulation

5 Capacity of wireless channels

MixedÀ¾ нOptimization Problem via Lagrange Multiplier Theory

MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.

Example: Credit card default, we may be more interested in predicting the probabilty of a default than classifying individuals as default or not.

Alternative proof for claim in [1]

Probability and Random Variables. Generation of random variables (r.v.)

Energy Efficiency of Cooperative Jamming Strategies in Secure Wireless Networks

Degrees of Freedom and Model Search

Stability of the LMS Adaptive Filter by Means of a State Equation

Time and Frequency Domain Equalization

QUICKEST MULTIDECISION ABRUPT CHANGE DETECTION WITH SOME APPLICATIONS TO NETWORK MONITORING

1.2 Solving a System of Linear Equations

Gaussian Conjugate Prior Cheat Sheet

Lecture 5: Variants of the LMS algorithm

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model

Functional-Repair-by-Transfer Regenerating Codes

Date: April 12, Contents

A Piggybacking Design Framework for Read-and Download-efficient Distributed Storage Codes

NCSS Statistical Software Principal Components Regression. In ordinary least squares, the regression coefficients are estimated using the formula ( )

SYSTEMS OF REGRESSION EQUATIONS

Log-Likelihood Ratio-based Relay Selection Algorithm in Wireless Network

Duobinary Modulation For Optical Systems

2WB05 Simulation Lecture 8: Generating random variables

Reject Inference in Credit Scoring. Jie-Men Mok

How To Understand And Solve A Linear Programming Problem

The Image Deblurring Problem

is in plane V. However, it may be more convenient to introduce a plane coordinate system in V.

Adaptive Solution for Blind Identification/Equalization Using Deterministic Maximum Likelihood

On the Traffic Capacity of Cellular Data Networks. 1 Introduction. T. Bonald 1,2, A. Proutière 1,2

Adaptive Equalization of binary encoded signals Using LMS Algorithm

The Effect of Network Cabling on Bit Error Rate Performance. By Paul Kish NORDX/CDT

Web Hosting Service Level Agreements

a 11 x 1 + a 12 x a 1n x n = b 1 a 21 x 1 + a 22 x a 2n x n = b 2.

A SURVEY ON CONTINUOUS ELLIPTICAL VECTOR DISTRIBUTIONS

Mehtap Ergüven Abstract of Ph.D. Dissertation for the degree of PhD of Engineering in Informatics

Discussion on the paper Hypotheses testing by convex optimization by A. Goldenschluger, A. Juditsky and A. Nemirovski.

SAS Software to Fit the Generalized Linear Model

SF2940: Probability theory Lecture 8: Multivariate Normal Distribution

IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 54, NO. 8, AUGUST If the capacity can be expressed as C(SNR) =d log(snr)+o(log(snr))

ADVANCED ALGORITHMS FOR EQUALIZATION ON ADSL CHANNEL

Continued Fractions and the Euclidean Algorithm

An introduction to OBJECTIVE ASSESSMENT OF IMAGE QUALITY. Harrison H. Barrett University of Arizona Tucson, AZ

Transcription:

AN EMPIRICAL LIELIHOOD METHOD FOR DATA AIDED CHANNEL IDENTIFICATION IN UNNOWN NOISE FIELD Frédéric Pascal 1, Jean-Pierre Barbot 1, Hugo Harari-ermadec, Ricardo Suyama 3, and Pascal Larzabal 1 1 SATIE, ENS Cachan, CNRS, UniverSud, 61, av President Wilson, F-9430 Cachan, France CREST-LS and University Paris-Dauphine, France, 3 DSPCOM, DECOM,FEEC, State University of Campinas, CP 6101, Campinas, SP, Cep 13083-970, Brazil email: pascal@satieens-cachanfr, barbot@satieens-cachanfr, harari@ensaefr, rsuyama@decomfeeunicampbr ABSTRACT In this wor we propose a different approach to the problem of estimating a transfer matrix with data aided, which can be applied to SISO and SIMO channels, either baud-rate or fractionally sampled signals The approach is based on the Empirical Lielihood method 1, ], a flexible semi-parametric estimation method, which can easily integrate in the estimation procedure some prior informations on the structure of the parameter of interest Moreover, this improved estimation method does not assume any model for the data distribution The contributions of this paper is twofold: first, we introduce the Empirical Lielihood method in a general context, ie without any prior informations and then, we derive closed-form expressions of estimators for the transfer matrix All results are presented under Gaussian assumptions and under a mixture of Gaussian and Student-t distributions This allows to show the robustness of the proposed method 1 INTRODUCTION The main objective of a communication system is to allow information to be properly exchanged between a transmitter and a receiver that are interconnected by a channel As a rule, this communication channel will introduce distortions intersymbol interference, noise, etc that are usually compensated for using an equalizer, whose design often needs an accurate estimator of the channel 3] Traditionally, a training sequence is used to aid the channel estimation tas This emitted nown sequence maes the identification of the channel feasible since both the input and the output signals are nown during the transmission of the training sequence Note that most of the propagation channel estimators are derived under the hypothesis of an additive Gaussian noise model With the resultant propagation channel coefficients, several linear and nonlinear methods can be used to estimate the emitted symbols 3] In this wor we propose a different approach to the problem, which can be applied to SISO and SIMO channels, either baud-rate or fractionally sampled signals The approach is based on the emprirical lielihood 1, ], a flexible semiparametric estimation method, which can easily incorporate a priori nowledge about the problem at hand In order to expose the proposed technique, this paper is organized as follows Section presents the problem of interest In Section 3 some bacground on the EL procedure is provided The section 4 details the introduction of prior on the transfer matrix Section 5 presents the simulation results, followed by the concluding remars in Section 6 PROBLEM FORMULATION 1 Mathematical notation In the following, boldface letters respectively capital letters denote column vectors resp matrices, H denotes the conjugate transpose operator, T denotes the transpose operator, E ] stands for the expectation of a random variable, E P ] is the expectation under the data probability P C resp R denotes the set of complex resp real numbers, while for any integer p, C p resp R p represents the set of p-vectors with complex resp real elements For z C, we write Rez and I mz its real and imaginary parts Signal Model The proposed signal model is the same as in 4] Consider a SIMO channel, modeled by a set of L finite response filters FIR of length M + 1, each one composed of a set of taps h i, = 0,,M, i = 0,,L 1 and a transmitted signal composed of a finite sequence of symbols s The received signal x i from the i-th subchannel is then given by x i M = l=0 h i s l + n i, 1 where n i denotes an additive white gaussian noise of unnown variance The transmitted sequence s is assumed to be iid, digitally modulated signal, and the channel impulse response is supposed time-invariant during the observation record The data aided channel estimation tas consists of determining the set of coefficients h i based on a set of observed samples x i and with the nowledge of the emitted signal Let x i = x i,xi 1,,xi N+1 ]T represent a vector containing N samples of the i-th subchannel output Denoting by s = s,s 1,,s M N ] T the vector containing M + N transmitted symbols, it is possible to express x i as follows: x i = H i s + n i, where n represents the vector of additive gaussian noise, and h 0 h 1 h M 0 0 H i 0 h 0 h 1 h M 0 = 0 0 h 0 h 1 h M 3

denotes the convolution matrix related to the i-th subchannel Hence, stacing the received vectors x i into a single vector x, yields the following received signal model: x 0 ḳ H 0 n 0 = ḳ s + 4 x L 1 x H L 1 H n L 1 n The following section is devoted to the estimation of the transfer matrix H First, we introduce an improved estimation procedure, the Empirical Lielihood and then, we estimate H with prior informations on its structure since H is a Sylvester matrix 3 EMPIRICAL LIELIHOOD The Empirical Lielihood method is a recent semiparametric method 5] designed for estimation problems in which the parameter of interest H is defined as the solution of an estimating equation The covariance matrix estimation problem, that we consider in this paper, can be formulated in such a way: xs H H ] = 0 5 where 0 denotes the null vector with appropriate dimension n = NL N + M 1 here since the signal s is a random vector containing 1 or 1 and such that Ess H ] = I Let us recall that NL is the dimension of the observations x and N + M 1 the dimension of the signal s Notice that the expectation is taen under the true Probability Density Function P 0 of x which is assumed to be unnown This is an important feature of the EL procedure which does requires any assumptions on the data distribution In most of the time, the additive noise is assumed to be Gaussian In this paper, we only use the EL method to estimate the transfer matrix H Therefore, we will restrict ourselves to built the corresponding Empirical Lielihood and to derive its Maximum Empirical Lielihood MEL estimator 31 Lielihood In this section, we introduce EL methodology from the classical lielihood context For that purpose, we consider the family of multinomials G charging the data set as if it was a parametric model for the data Notice that this parametric model assumption will never be made in our method, we just made it to interpret EL as a classical lielihood All the details and technical arguments are given in Owen s boo 5] We define, for G and H verifying the moment condition E G xs H H ] = 0, the distribution or equivalently, q if, x = x Gx = 0 otherwise, Gx = =1 6 q δx x, 7 where δx stands for the Dirac measure at element x, 0 < q < 1 and =1 q = 1 The corresponding lielihood, as if the data were distributed according to G, is called Empirical Lielihood: ELH = ELx 1,,x,s 1,,s,H 8 } = sup q H x s q =1q H = 0, q = 1 =1 =1 The main technical difficulty, evaluating the empirical lielihood ELH at any given H, is resolved by a Lagrangian method and ELH can be written in an easier form: logelh = inf λ =1 log 1 + λ H x s H }, 9 because the optimal weights write q = 1 1 + λ H x s H 1, 10 where λ is the optimal Lagrange multiplier and depends on H 3 Maximum Empirical Lielihood without prior information Now, Ĥ EL is define as the arginf of ELH: Ĥ EL = arg inf H inf λ =1 log 1 + λ H x s H } 11 If no restriction is assumed on the structure of H, the Maximal Empirical Lielihood MEL estimator is given by the following expression Ĥ EL1 = xs H = 1 =1 x s H 1 where the notation a is used for the empirical mean of the vector a: 1 =1 a Notice that Ĥ EL1 is equal to the Maximum Lielihood ML estimator under Gaussian assumption This is the purpose of the following proposition: Proposition 31 Let y 1,,y an iid data set in R p, with common distribution P 0 and expectation θ 0 R p and finite variancecovariance matrix Then the Maximal Empirical Lielihood estimator of the expectation is given by ˆθ MEL = x 13 where x denotes the empirical mean, x = 1 =1 x

Remar 31 This result is interesting since it provides an estimator of the mean without any assumptions on the data distribution and it corresponds to the Maximum Lielihood estimator under Gaussian assumptions and many others classical distributions Another important feature of the EL method is that prior informations can be taen into account into the estimation procedure This is one of the two contribution of this paper 33 Additional prior information This explicit expression given by equation 1 for the estimator can be obtain even when some restrictions are made on the covariance matrix structure The particular case of Sylvester matrix will be considered in section 4 The estimation scheme can tae into account some priors information by introducing an additional equation to equation 5 as follows: vech xs H c prior ] = 0p 0 s 14 where c prior is a vector resulting in a data transformation which reflects the priors information and where the operator vec reshapes a m n matrix elements into a mn column vector By rewriting equation 14 as v w ] = vech, 15 0 where v = vecxs H, w = c prior and by setting, for the covariance matrix of the vector v w v V11 V Var = 1, 16 w V 1 V one obtains the following closed-form expression for the MEL estimator: ˆθ MEL = v V 1 V 1 w 17 This closed-form can only be obtained when priors information can be written in terms of equation 14 This result has been proved in 5], page 5 In practice, the empirical versions of V 1 and V allow to obtain an approximation of the MEL, without significant computational cost Notice that when there is no prior information, ie w = 0, one obtains Ĥ EL1 This is also the case when w is uncorrelated with v because V 1 is the null matrix 4 APPLICATION TO SYLVESTER MATRIX ESTIMATION In this section, we use prior information on the transfer matrix structure, defined in equation 4 This section is divided in three steps: It is first assumed that H is a real matrix Then, the full Sylvester structure of H is contained in c prior Notice that there already exist methods for structured covariance matrix estimation in which the Toeplitz case is treated, see eg 6, 7], but these methods are based on a parametric model, usually the Gaussian one 41 First prior information: H is real A first step is to assume that H has real valued elements Therefore, the estimating equation is modified to tae the structure into account For that purpose, we define the function m as x j1 s H j h j 1 j1 NL,1 j M+L 1 m x,s,h = I m x j1 s H j h j1 j 1 j1 NL,1 j M+L 1 18 This leads to a new estimator Ĥ EL which integrates the constraint on the real transfer matrix H This writes Ĥ EL = arg inf log 1 + λ N,λ m x,s,h } 19 =1 4 Second prior information: H shares a Sylvester structure with L M non null elements In this subsection, the unnown parameters of H are only L M real scalars: for i = 1,,L, h i 1,,hi M To estimate H by taing into account the prior informations, we will focus on each Sylvester matrix H i m 3 x i,s,h i = where x i j1 sh j1+ j hi j 1 j1 N, 0 j M 0 This leads to the last estimator Ĥ EL3 of H, defined by Ĥ EL3 = Ĥ 0 Ĥ L 1 1 Ĥ i = arg inf H,λ } log 1 + λ m 3 x i i,s,h i =1 We can rewrite the constraint in terms of expectations in order to obtain an explicit form of the estimator by means of equation 17 For simplicity purpose, we give the constraints for M = and N = Let v = Re x 1 s H 1,x 1s H w = I m x 1 s H 1,x 1 s H,x1 s H 1 x s H, x 1 s H x s H 3,x 1 s H 3,x s H 1 3 Let us explain constraints containing in the vector w:

The first elements allows to specify that h 1 and h are real scalars, thus their imaginary part is null The two subtractions means that the two diagonals of H contains the same element, and thus the difference is null Finally, last elements are the null elements of the transfer matrix Equation 17 gives estimators for the first line of N: ĥ1,ĥ = v V 1 V 1 w 4 The estimator Ĥ i writes then ] Ĥ i ĥ1 ĥ = 0 0 ĥ 1 ĥ Remar 41 One can give a general expression for all these estimators: Ĥ EL j = =1 q jx s H, 5 where the q j depend on the constraints For example, in the case of no constraint ie Ĥ EL1, the q 1 are all equal to 1 This corresponds to the ML estimator under Gaussian assumptions These theoretical estimators of H will be studied in the section 5 thans to simulations on their Mean Square Error MSE under Gaussian and non-gaussian assumptions 5 SIMULATIONS In order to enlighten results provided in sections 3 and 4, some simulation results are presented We focus on the problem of transfer matrix estimation in the case of an additive Gaussian or non-gaussian noise In order to compare all estimators, we will plot the Mean Square Error MSE in different realistic situations The MSE used in this section is the following criterion: ] Ĥ H MSEĤ,H = E, H where stands for the Frobenius norm and Ĥ denotes the studied estimator of H The transfer matrix H which has to be estimated shares a Sylvester structure and is defined as follows: H 0 H =, 6 H L 1 where h 0 h 1 h M 0 0 H i 0 h 0 h 1 h M 0 = 7 0 0 h 0 h 1 h M In the simulations, the number L of response filters is L = 4, the length M + 1 of each response filter is M + 1 = 5 and the number of samples is N = 10 Thus, the matrix H has N L lines and M +N 1 columns Moreover, the number of x and s is = 00 The transfer matrix H to be estimated is defined thans to the following response filter: h 1 = 0049;048; 0556;1; 0171], h = 0443;1;091;0189; 0087], h 3 = 011; 0199;1; 084;0136], h 4 = 0417;1;0873;085; 0049] To evaluate performance of our method, we will compare the three following estimators of H: The well-nown Sample Covariance Matrix which corresponds to the ML estimator under Gaussian assumptions and defined as follows Ĥ ML = 1 =1 x s H Ĥ ML is used as a benchmar but it is not appropriate to our problem since it does not tae into account the structure of the real covariance matrix Notice that this ML estimator is the same as Ĥ EL1 which is the MEL estimator without any prior informations Then, we plot Ĥ EL which the MEL estimator in the case of a real transfer matrix H Ĥ EL is defined by equation 19 Finally, we plot Ĥ EL3, the MEL estimator which uses all prior information on the structure Ĥ EL is defined by equation 5 Concerning the EL method, notations of section 4 are still valid: Ĥ EL1, Ĥ EL and Ĥ EL3 We first studied the behavior of the proposed method in a Gaussian context and then in a non-gaussian context 51 Gaussian distribution Figure 1 shows the MSE of each estimator versus the Signalto-Noise Ratio under Gaussian assumptions, ie the additive noise b is Gaussian distributed 5 Mixture of Student-t and Gaussian distributions For = 1,,, b 1 αc T + α C N 0,I 1 α + α where C T is a complex Student-t distribution with degree of freedom 1, which has been centered and normalized while C N the classical complex Gaussian distribution The set of parameters in this subsection is p = 3, = 100 and ρ = 05 53 Simulations comments A first comment is that, as expected, the MSE decreases as the estimators mae use of more prior informations On each graphic, the estimator EL1 that does not use any prior on the structure has MSE higher than the MSE of the

MSE 18 16 14 1 1 08 Comparison of MSE for each estimate vs the SNR, under Gaussian assumptions Classical Lielihood on H = EL1 EL EL3: with all constraints 6 CONCLUSION This contribution presents an Empirical Lielihood based approach for the identification of multichannel FIR filters in data aided context The Empirical Lielihood a flexible semi-parametric estimation method, which can easily integrate in the estimation procedure some prior informations on the structure of the parameter of interest Moreover, this improved estimation method does not assume any model for the data distribution The numerical simulation evidence that the proposed method yield reasonable estimates of the channel coefficients even in non-gaussian additive noise 06 04 0 0 10 0 10 0 30 SNR Figure 1: MSE against the Signal-to-Noise Ratio, under Gaussian assumptions, for L = 4, M = 5 and N = 10 estimators that mae use of the structure informations EL and EL3 These comments are still valid under Gaussian and non-gaussian assumptions MSE 0 18 16 14 1 10 8 6 Comparison of MSE for each estimate vs the coefficient of mixture Classical Lielihood on H = EL1 EL EL3: with all constraints 7 ACNOWLEDGEMENT This wor was supported by the French Aerospace Lab through the PEA DTC, the European networ of excellence NEWCOM++ and the CNRS/FAPESP program REFERENCES 1] F Pascal, H Harari-ermadec, and P Larzabal, The empirical lielihood: an alternative for signal processing estimation, IEEE Trans-SP submitted, 008 ] H Harari-ermadec and F Pascal, On the use of empirical lielihood for non-gaussian clutter covariance matrix estimation, in Accepted to the IEEE-RADAR 008, Roma, June 008 3] J G Proais, Digital Communicarions McGraw-Hill, Third Ed, New Yor, 1995 4] E Moulines, P Duhamel, J F Cardoso, and S Mayrargue, Subspace methods for the blind identification of multichannel FIR filters, IEEE Trans on Signal Processing, vol 43, no, pp 516 55, Feb 1995 5] A B Owen, Empirical Lielihood Chapman & Hall/CRC, Boca Raton, 001 6] J P Burg, D G Luenberger, and D L Wenger, Estimation of structured covariance matrices, Proc IEEE, vol 70, no 9, pp 963 974, September 198 7] D R Fuhrmann, Application of toeplitz covariance estimation to adaptive beamforming and detection, IEEE Trans-SP, vol 39, no 10, pp 194 198, October 1991 4 0 01 0 03 04 05 06 07 08 09 Coefficient α of mixture Figure : MSE against α, for a mixture of Student-t and Gaussian distributions, for L = 4, M = 5 and N = 10 Then, we consider the figure 1 which deals with Gaussian distribution One can see that all estimators have a MSE which decreases as the SNR increases It is also interesting to notice that the difference between all curves is the same for all SNR On figure, one can notice the improvement introduced by the proposed method in comparison with the classical ML method which degrades for small values of α, ie for highly non-gaussian noise