Lars Nielsen. Abstract. In this paper, a general technique for evaluation of measurements by the method of

Size: px
Start display at page:

Download "Lars Nielsen. Abstract. In this paper, a general technique for evaluation of measurements by the method of"

Transcription

1 Evaluation of measurements by the method of least squares Lars Nielsen Danish Institute of Fundamental Metrology (DFM), Lyngby, DK LNdfmdtudk Abstract In this paper, a general technique for evaluation of measurements by the method of Least Squares is presented The input to the method consist of estimates and associated uncertainties of the values of measured quantities together with specied constraints between the measured quantities and any additional quantities for which no information about their values are known a priori The output of the method consist of estimates of both groups of quantities that satisfy the imposed constraints and the uncertainties of these estimates Techniques for testing the consistency between the estimates obtained by measurement and the imposed constraints are presented It is shown that linear regression is just a special case of the method It is also demonstrated that the procedure for evaluation of measurement uncertainty that is currently agreed within the metrology community can be considered as another special case in which no redundant information is available The practical applicability of the method is demonstrated by two examples Introduction In 787, the French mathematician and physicist Laplace (749{827) used the method of Least Squares to estimate 8 unknown orbital parameters from 75 discrepant observations of the position of Jupiter and Saturn taken over the period 582{745 Since then, the method of Least Squares has been used extensively in data analysis Like Laplace, most people use a special case of the method, known as unweighted linear regression The calculation of the average and the standard deviation of a repeated set of observations is the most simple example of that The unweighted regression analysis is based on the assumptions that the observations are independent and have the same (unknown) variance In addition, the linear regression is based on the assumption that the observations can be modelled by a function that is linear in the unknown quantities to be determined by the regression analysis For most measurements carried out in practice, none of these assumptions can be justied In order to evaluate the result of a general measurement, in which some redundant information has been obtained, one therefore has to apply the method of Least Squares in its general form This paper describes how measurements can be evaluated by the method of Least Squares in general The paper is based on an earlier work of the author [2] but includes Address: uilding 37, Matematiktorvet, DK-28 Lyngby, Denmark

2 2 Lars Nielsen several new features not published before as well as practical examples from the daily work at DFM An alternative approach is described in [6] 2 Measurement model In a general measurement, a number m>ofquantities is either measured directly using measuring instruments or known a priori, for example from tables of physical constants etc The (exact) values of these m quantities are denoted =( ::: m ) T : Due to measurement uncertainty, the values z obtained by the measurement (or from tables etc) z =(z ::: z m ) T are only estimates of the values The standard uncertainties of the estimates z i, u(z i ) i = ::: m are determined in accordance with the GUM [] and depend on the accuracy of the instruments and the reliability ofany tabulated value used In general, some of the estimates z i may be correlated If r(z i z j ) is the correlation coecient between the estimates z i and z j then the covariance u(z i z j )between these two estimates is given by u(z i z j )=u(z i )r(z i z j )u(z j ): ecause of the uncertainty, the estimates z can be considered as an outcome of a m- dimensional random variable Z with expectation (the exact values of the quantities) and covariance matrix = u(z z T )= u 2 (z ) u(z z 2 ) u(z z m ) u(z 2 z ) u 2 (z 2 ) u(z 2 z m ) A : u(z m z ) u(z m z 2 ) u 2 (z m ) In addition to the m quantities for which prior information is available either from direct measurement or from other sources, a general measurement may involve a number k of quantities for which no prior information is available The values of these quantities are denoted by =( ::: k ) T : In general, the values and are constrained by anumber n of physical or empirical laws These constraints may be written in terms of an n-dimensional function f( ) = f ( ) f 2 ( ) f n ( ) A = A k n<m+ k: (2) It is assumed that f i :! R i=:::n, are dierentiable functions (with con-

3 Evaluation of Measurements 3 tinuous derivatives) dened in a region R k+m around ( ) As indicated in (2), the number n has to be larger than or equal to the number k of quantities for which no prior information is available otherwise some of the values cannot be determined In addition, the number n of constraints has to be smaller than the total number k + m of quantities involved otherwise the values of and would be uniquely determined by the constraints and no measurements would be needed The estimates z, the covariance matrix and the n-dimensional function f( ) are the input to the general Least Squares method It should be stressed that no probability distribution has to be assigned to the input estimates z On the contrary, if a probability distribution has been assigned to an estimate, it should be used to calculate the mean value and the variance of the estimate which should then serve as input to the Least Squares method Like any other covariance matrix, the covariance matrix u(z z T )= is positive semi-denite Otherwise, at least one linear combination x T z of the estimates z would have negative variance u(x T z z T x)=x T x In the following it is assumed that is positive denite and therefore non-singular 3 Normal equations Least Squares estimates ^ and ^ of the values and are found by minimizing the chi-square function 2 ( z) =(z ; ) T ; (z ; ) subject to the constraints f( ) =: It is convenient to solve this minimization problem by using Lagrange multipliers [5]: If a solution ( ^ ^) to the minimization problem exists, the solution satises the equation where r ( ) ( ^ ^ z) = ( z) =(z ; ) T ; (z ; )+2 T f( ) for a particular set of Lagrange multipliers =( ::: n ) T y taking the gradient of the function, the following n + m + k equations in ( ^ ^ ) evolve: where r f = f f n r f( ^ ^) T = ; ; (z ; ^)+r f( ^ ^) T = f( ^ ^) = f k f n k A and r f = f f n f m f n m A : The equations (3) are called the normal equations of the Least Squares problem (3)

4 4 Lars Nielsen 4 Solving the normal equations If ( l l l )isanapproximate solution to the normal equations, a rened solution ( l+ l+ l+ ) can be found by the iteration l+ l+ l+ A = l l The step ( l l l+ ) is given by where D( l l )= D( l l ) l l l+ A + l l l+ A = A l = ::: : (k ) ; (z ; l ) ;f( l l ) (k k) (k m) r f( l l ) T (m k) ; r f( l l ) T r f( l l ) r f( l l ) (n n) A (4) A (42) is a symmetric matrix This iteration procedure is similar to Newton iteration except that the second order partial derivatives of the functions f i have been neglected as it is practice to do in non-linear Least Squares estimation [4] In order to reduce the eects of numerical rounding errors, it is recommended to calculate the step ( l l l+ )by solving the linear equations (4) by Gauss-Jordan elimination with full pivoting [4] This algorithm also provides the inverse matrix D( l l ) ; which is needed at the nal stage for estimating the covariance matrix of the solution as shown in Section 5 If proper starting values ( ) are selected, the iteration is expected to converge towards the solution (^ ^) ^ ^ A =lim l! l l l Since the solutions ^ are expected to be close to the estimates z of available a priori, the estimates z are obviously the proper starting values to be selected for the iteration The selection of proper starting values is more dicult in general If, however, f( ) are linear functions in the variables, the iteration process will converge after a few iterations, independent ofthechoice of Most dierentiable functions f( ) can be handled by the described method In order to get reliable standard uncertainties, it is required, however, that the function can be approximated by a rst order Taylor expansion, ie A : f( ) = f( ^ ^)+r f( ^ ^)( ; ^)+r f( ^ ^)( ; ^) when the values and are varied around the solution ^ and ^ on a scale comparable to the standard uncertainties of the solution If this vaguely formulated criterion is met, the function f( ) issaidtobelinearizable Note that almost any dierentiable function

5 Evaluation of Measurements 5 will be linearizable if the standard uncertainties are suciently small On the other hand, if the uncertainties are suciently high, all non-linear functions will no longer be linearizable The requirement that f( ) is linearizable is considered to be the only major limitation of the method of Least Squares! It should be mentioned that the minimization using Lagrange multipliers will fail in case the gradients r f i and r f i of one of the constraint functions f i are both equal to zero at the point of the solution (^ ^) This gives some restrictions on how a constraint maybeformulated A function f i dening a constraint may, for example, be replaced by the square of that function, fi 2 ut since f i(^ ^) =, the gradient offi 2 will be zero at the point of the solution (^ ^) although the gradient off i is not 5 Properties of the solution Since the solution ( ^ ^ ) depends on the estimates z, which are considered as the realization of the multivariate random variable Z, the solution (^ ^ ) canalsobe regarded as a multivariate random variable If the functions f i (^ ^) are linearizable, the estimates (^ ^) are linear in Z ^ ^ A = A + D( ) ; In that case, the expectation of the solution is E ^ ^ A = (k ) ; (Z ; ) (n ) A A : (5) which means that (^ ^) are central estimators of the values ( ) Under the same ; assumption, the covariances of the solution are given by the symmetric matrix D( ^ ^) provided by the Gauss-Jordan elimination algorithm 2 u( ^ ^ T ) u( ^ ^ T ) () (k n) u(^ ^ T ) u(^ ^ T ) () (m n) () (n k) () (n m) ;u( T ) A = D( ) ; = D( ^ ^) ; : (52) This relation can be derived as follows Partition the symmetric matrix D ; into nine sub-matrices similar to the left hand side of (52) or similar to the partitioning of D according to the denition (42) Express the covariance matrix of the solution (5) in terms of the covariance matrix of the random variable Z and the matrix D ; Insert the partitioned D ; into the resulting matrix double product and express the covariances of the solution in terms of ; and the sub-matrices of D ; Reduce these expressions to the nal result by multiple use of the nine relations between the sub-matrices of D ; and D derived from the identity DD ; = I 2 The empty brackets in the left hand matrix indicates the parts of D ; that do not contain information about covariances

6 6 Lars Nielsen From equation (5) and (52) the covariance matrices between ( ^ ^) and the estimates z are found to be u(z ^ T )=u(^ ^ T ) u(z ^ T )=u(^ ^ T ): From the last of these two relations, a relation of particular interest is derived, u(z ; ^ z T ; ^ T )=u(z z T ) ; u(^ ^ T ): For the diagonal elements, this relation reads u 2 (z i ; ^ i )=u 2 (z i ) ; u 2 (^ i ) i = ::: m: That is, the variance of the dierence between the initial estimate z i of i and the rened estimate ^ i is equal to the variance of z i minus the variance of ^ i This relation is useful when testing if the dierence z i ; i is signicantly dierent from its zero expectation 6 2 test for consistency When the estimates (^ ^) have been found, the minimum 2 value 2 (^ z) =(z ; ^) T ; (z ; ^) can be used to test if the measured values z are consistent with the measurement model (2) within the uncertainties dened by thecovariance matrix If the model is linearizable, the expectation of the random variable 2 (^ Z) is equal to the number m of measured quantities, minus the number m + k of adjusted quantities, plus the number n of constraints, that is h i E 2 (^ Z) = m ; (m + k)+n = n ; k = : If, in addition, the random variables Z are assumed to follow amultivariate normal distribution with mean values and covariance matrix, the random variable 2 (^ Z) will follow a 2 () distribution with = n ; k degrees of freedom In that case, the probability p of nding a 2 value larger than the value 2 (^ z) actually observed can be calculated from the 2 () distribution n o p = P 2 () > 2 (^ z) =; P f 2 () 2 (^ z)g: If this probability p is smaller than a certain value, the hypothesis that the measured values are consistent with the measurement model has to be rejected at a level of signicance equal to As the result of measurements are normally quoted at a 95% level of condence, an = 5% level of signicance is a reasonable choice for the consistency test Although the assumption of a normal distribution of Z may not be fullled, it is suggested to carry out the test of consistency as described above anyway This is justied by the fact that a value of 2 (^ z) signicantly higher than the expectation indicates inconsistency no matter what the distribution of Z might be The calculated probability

7 Evaluation of Measurements 7 p simply describes how unlikely the observed 2 assigned to Z value is if a normal distribution is 7 Normalized deviations If the test described in the previous section leads to a rejection of the measurements, a tool for identifying the outlying measurements is desirable A measured value z i is dened as an outlier if the dierence z i ; ^ i is signicantly dierent from zero taking into account the standard uncertainty u(z i ; ^ i ) of that dierence This leads to the introduction of the normalized deviation d i dened by 3 d i = z i ; ^ i u(z i ; ^ i ) = z i ; q ^ i u 2 (z i ) ; u 2 (^i ) i = ::: m: The normalized deviation d i has zero expectation and variance A normalized deviation with jd i j larger than 2 or 3 is therefore rather unlikely no matter what the distribution of the random variable d i might be Ifamultivariate normal distribution is assigned to Z and the model function f( ) is linearizable, the normalized deviation d i is normally distributed, In that case d i 2 N( ) i = ::: m: P fjd i j > 2g =5% and a measurement with jd i j > 2 is therefore identied as an outlier at a 5% level of signicance It is suggested to use the criteria jd i j > 2 to identify potential outliers even if the distribution assigned to Z is not normal 8 Adjustment ofavariance 2 If some values z i have a common but unknown variance u 2 (z i )= 2,thisvariance can be estimated by adjusting 2 by an iterative procedure until the \observed" 2 value becomes equal to its expectation value 2 (^ z) =(z ; ^) T ; (z ; ^) = where the covariance matrix is a function of the unknown variance 2 As the estimates ^ depends on the value assigned to 2, these estimates have to be updated together with the estimates ^ each time the value of 2 is changed during the iteration This way of estimating the unknown variance 2 leads to the well-known expression for the standard deviation in the case of a repeated measurement of a single quantity as shown in Section 3 3 If u(z i ; ^i ) =, the dierence z i ; ^i will be zero as well and d i may be set equal to zero This situation occurs whenever there is no redundant information available regarding the value of the quantity i

8 8 Lars Nielsen 9 Example: alibration of an analytical balance An analytical balance with capacity Max=22 g, resolution d= mg, and built-in adjustment weight was calibrated by DFM in October 999 during an inter-laboratory measurement comparison piloted by DFMTwo mass standards were used as reference standards One of them was a traditional 2 g weight (named R2g) of known conventional mass value 4 m R and density R The other reference standard was a specially designed 2 g stack of weights consisting of four discs (named g, 5g, 25g and 25g*) machined from the same metal bar of known density The conventional mass values m, m 2, m 3, m 4 respectively of these four discs were not known a priori only the conventional mass value m S = m + m 2 + m 3 + m 4 of the stack was known The calibration was performed by placing a weight combination at the weighing pan of the balance and by recording the corresponding average indication I in the display A total of 8 weight combinations were used Each weight combination was weighed 3 times from which the average indication was calculated The calibration was repeated 4 times during a period of days in which the inter-laboratory comparison took place From these four calibrations, a grand average indication I i i= ::: 8 was calculated for each of the 8 weight combinations specied in Table The standard uncertainty of the grand average was estimated from the observed variation in indication over the four calibrations I I 2 I 3 I 4 I 5 I 6 I 7 I 8 I 9 g g g g g g g g g 5g 5g 5g 5g 5g 25g 25g* 25g 25g 25g 25g 25g* 25g* 25g* 25g* I I I 2 I 3 I 4 I 5 I 6 I 7 I 8 5g 5g 5g 5g 25g 25g 25g* R2g R2g 25g 25g 25g* 25g* 25g* Tab The weight combinations corresponding to the 8 balance indications I i Due to the eect of air buoyancy, the balance indication depends not only on the mass of the weighed body, but also on the density ofthebodyaswell as the density of the air When calibrated in air of known density a, the reference indication I R of the balance corresponding to a load generated by aweight with conventional mass value m and density is given by I R = m ; (a ; a ) ; where a =2 kg/m 3 and =8 kg/m 3 are the reference densities of the air and the weight respectively to which the conventional value of mass refers As a model for the 4 The conventional massvalue of a body is dened as the mass of a hypothetical weight of density 8 kg/m 3 that balances the body when weighed in air of density 2 kg/m 3 and temperature 2

9 Evaluation of Measurements 9 m S m R R a I [g] [g] [kg/m 3 ] [kg/m 3 ] [kg/m 3 ] [div] z u(z) ^ u(^) d I 2 I 3 I 4 I 5 I 6 I 7 [div] [div] [div] [div] [div] [div] z u(z) ^ u(^) d I 8 I 9 I I I 2 I 3 [div] [div] [div] [div] [div] [div] z u(z) ^ u(^) d I 4 I 5 I 6 I 7 I 8 [div] [div] [div] [div] [div] z u(z) ^ u(^) 3 d f A m m 2 m 3 m 4 [g/div] [/div] [g] [g] [g] [g] ^ 86-44E u( ^) 9 E-9 Tab 2 Measured and estimated values and associated standard uncertainties calibration curve of the balance, a second order polynomial through zero is assumed I R = f ; I + AI 2 where f and A are unknown quantities to be determined from the calibration data In this example, there are m = 23 quantities for which prior information is available from the measurements performed: =(m S m R R a I ::: I 8 ) T whereas there are k = 6 quantities for which no prior information is available: =(f A m m 2 m 3 m 4 ) T :

10 Lars Nielsen f( ) = f A m m 2 m 3 m 4 f A m m m m Tab 3 orrelation coecients of the estimated ^ values etween these quantities, there are n = 9 constraints: (m + m 2 + m 3 + m 4 ) ; (a ; a ) m R ; (a ; a ) ; ; f ; I + AI 2 R ; ; f ; I 8 + AI 2 8 m S ; (m + m 3 + m 3 + m 4 ) A = : The measured values z and associated standard uncertainties are given in Table 2 under the row headings z and u(z) All measured values are assumed to be uncorrelated y solving the normal equations, the estimates ^ and ^ and associated standard uncertainties given in Table 2 under the row headings ^, u(^), ^ and u( ^) are obtained Selected correlation coecients derived from D( ^ ^) ; are given in Table 3 The observed minimum 2 value is (^ z) =8:6 which should be compared to the expectation value = n ; k =9; 6 = 3 Since P 2 (3) > 8:6 =8:3%, it is concluded that the measured values are consistent with the specied constraints taking into account the measurement uncertainties This conclusion is conrmed by the calculated normalized deviations given in Table 2 under the row heading d all normalized deviations satisfy the criterion jdj < 2 From the estimates of the quantities f and A and the associated covariance matrix, the error of indication E, dened as E = I ; I R = I ; f ; I + AI 2 and the associated standard uncertainty u(e) can be calculated as a function of the indication I The result is shown in Figure as the full lines representing E ; u(e), E, and E +u(e) The measured points E i i= ::: 8 shown in the gure are the observed average balance indications I i minus the corresponding reference values I R The error bars of the measured points indicate the standard uncertainties u(e i )thathave been calculated taking into account the covariance between I i and I R Example: Evaluation of calibration history Aweight (named Rmg) of nominal mass mg has been calibrated 39 times in the period For calibration number i, themass m i of the weight at the time t i and the associated standard uncertainties u(m i )andu(t i ) are given The calibration history of the weight isshown in Figure 2 as dots with error bars indicating the standard

11 Evaluation of Measurements alance P22S Error of indication E /mg Fit Fit+u(Fit) Fit-u(Fit) Measured alance indication I /g Fig Error of indication of the calibrated balance uncertainties the scale mark 992- on the time axis indicates the position of the date January 992 etc Due to wear and changes in the amount of dirt adsorbed to the surface, the mass of the weight is expected to change in time A reasonable model of the change in mass as a function of time is a superposition of a deterministic linear drift and a random variation m i = a + a 2 t i + m i i = ::: 39 where m i isarandomvariable with zero expectation and variance 2 The drift parameters a, a 2 and the associated covariance matrix as well as the variance 2 are unknown a priori and are to be estimated from the calibration history available Once the estimates ^a and ^a 2 have been found, it is possible to predict a value ^m of the mass of the weight as a function of time t ^m =^a +^a 2 t + ^ m where m ^ = with standard uncertainty u( m) ^ = The standard uncertainty ofthe predicted mass value is given by u 2 (^m) =u 2 (^a )+t 2 u 2 (^a 2 )+2tu(^a ^a 2 )+ 2 : The measurement model used for evaluating the calibration history is =(m ::: m 39 t ::: t 39 m ::: m 39 ) T =(a a 2 ) T

12 2 Lars Nielsen Rmg (efore adjustment of σ ) 75 Mass Value [mg] ν = 37, χ 695 obs = 66 6, σ = µ g P{ χ ( 37) > 66 6} = 2% Time of alibration Result Ignore Fit + u Fit Fit - u Normalized Deviation Time of alibration Deviation Fig 2 Evaluation of the calibration history of a mg weight assuming that = Rmg (After adjustment of σ) 75 Mass Value [mg] ν = 37, χobs = 37, σ = 92 µ g P{ χ ( 37) > 37 } = 46 7% Time of alibration Result Ignore Fit + u Fit Fit - u Normalized Deviation Time of alibration Deviation Fig 3 Evaluation of the calibration history of the mg weight with adjusted to 92 g f( ) = m ; (a + a 2 t + m ) m 39 ; (a + a 2 t 39 + m 39 ) A = : The measured values z are given by the calibration history, except for the values of

13 Evaluation of Measurements 3 m i i= ::: 39 which are set equal to the expectation value zero The associated covariance matrix u(z z T )= is built up from the uncertainties u(m i )andu(t i )available from the calibration history and a negligible but nite 5 initial value of the unknown variance 2 Since the standard uncertainties u(m i ) are of the order g, the value =E-7 g is considered negligible and is selected as a starting point y solving the normal equations, estimates ^a and ^a 2 of the drift parameters and the associated covariance matrix are found after a few iterations The predicted value ^m of the mass of the weight and the associated standard uncertainty u(^m) as a function of time are shown in Figure 2 as solid lines The normalized deviations d associated with the mass values m i are shown in Figure 2 as well 6 The observed minimum chi-square value is 2 =66:6 which is large compared to the expectation value =39; 2=37 Since P f 2 (37) > 66:6g =:2%, the hypothesis =,ornorandomvariation in the mass, is rejected at a 2% level of signicance The value of is therefore increased as described in Section 8 until the calculated minimum 2 value becomes equal to its expectation value = 37 In this way the standard uncertainty reecting the random variation of the mass of the weight is found to be =92 g The result of the evaluation of the calibration history after adjustment of is shown in Figure 3 Note the signicant increase in the standard uncertainty of the predicted value of the mass of the weight and the decrease in the absolute value of the normalized deviations d The calibration history can also be evaluated by an iterativetechnique based on linear regression [3] The results obtained are identical to the results presented in this section ase I: Univariate output quantity, Y = h(x ::: X N ) In this section it is shown that the evaluation of measurements by the method of least squares is consistent with the generally accepted principles for evaluating measurement uncertainty as described in the GUM [] Using the nomenclature of the GUM, a univariate output quantity Y is assumed to be related to N input quantities X ::: X N through a specied function h, Y = h(x ::: X N ): The values assigned to the input and output quantities are denoted x ::: x N and by y respectively In the nomenclature of this paper, the measurement model is =(X ::: X N ) T =(Y ) The measured values are f( ) =(Y ; h(x ::: X N )) = : z =(x ::: x N ) T 5 If the variance 2 is assumed to be exactly zero, the quantities m i have toberemoved from the model Otherwise the covariance matrix will be singular 6 The absolute value of normalized deviations of t i and m i is equal to the absolute value of the normalized deviation of m i

14 4 Lars Nielsen with the known covariance matrix = u(z z T )= u 2 (x ) u(x x N ) u(x N x ) u 2 (x N ) The coecient matrix D of the normal equations is where D( ^ ^) = ( N) (N ) ; ;r x h(x) T ;r x h(x) r x h = h h ::: : X X N In the present case, the solution to the normal equations is found after one iteration, A : A y = ^ = h(x ::: x N ) ^ =(x ::: x N ) T =: The associated covariances are given by u2 (y) u(y ^ T ) () ( ) u(^ y) u(^ ^ T ) () (N ) () ( ) () ( N) ;u 2 () In other words, ; A = D( ^ ^) = u 2 (y) =r x h(x) r x h(x) T = r xh(x) r x h(x) T r x h(x) r x h(x) T NX NX i= j= c i u(x i x j )c j c i h X i (x i ) which is identical to the linear variance propagation formula given in the GUM 2 ase II: Linear regression, Y = Xa Linear regression is applied when there is a linear relationship Y = Xa between some observed quantities Y and some unknown quantities a The design matrix X is made up of known elements that maybegiven as specied functions of one or several independent variables In the notation of this paper, the measurement model for the linear regression problem is = Y =(Y ::: Y n ) T = a =(a ::: a k ) T f( ) =Y ; Xa = where X (n k) is the known design matrix The measured values are z = y =(y ::: y n ) T A :

15 Evaluation of Measurements 5 with known covariance matrix = u(z z T )= u 2 (y ) u(y y n ) u(y n y ) u 2 (y n ) The coecient matrix D of the normal equations is D( ^ ^) = (k k) (k n) ;X T (n k) ; I (n n) ;X I (n n) (n n) A : A : Again, the solution to the normal equations is found after one iteration, ^a = ^ = X T ; y ^Y = ^ = X^a = ; ; ( ^Y ; y) where (X T ; X) ; The associated covariances are given by u(^a ^at ) u(^a ^Y T ) () (k n) u( ^Y ^a T ) u( ^Y ^Y T ) () (n n) () (n k) () (n n) ;u( T ) that is, = ; A = D( ^ ^) X T ;X T ; X XX T I ; XX T ; ; ; X I ; ; XX T ; XX T ; ; ; ^a = X T ; y u(^a ^a T )= =(X T ; X) ; as is known from the theory of linear regression 3 ase III: Repeated observations of a single quantity Assume that a quantity X is measured n times with the same uncertainty Sucha measurement can be modelled by n quantities X ::: X n having a common value The measured values are = X =(X ::: X n ) T =() f( ) = X ; X n ; A = : z = x =(x ::: x n ) T and under the assumption that the measurement results are mutually independent, the associated covariance matrix is given by = u(z z T )= 2 2 A : A

16 6 Lars Nielsen The coecient matrix D of the normal equations is D( ^ ^) = ( n) ; ( n) (n ) ;2 I (n n) I (n n) ; (n ) I (n n) (n n) ( ) where denotes a matrix with all elements equal to The solution of the normal equations is found after one iteration, ^ = ^ = n nx i= The associated covariances are given by u2 (^) u(^ ^X T ) () ( n) u( ^X ^) u( ^X ^X T ) () (n n) () (n ) () (n n) ;u( T ) As expected, = A x i ^X = ^ =^ (n ) = ; ;2 ( ^X ; x): ; A = D( ^ ^) 2 n ; 2 n ; ( n) n ; ( n) 2 n ; (n ) 2 n ; (n n) I (n n) ; n ; (n n) n ; (n ) I (n n) ; n ; (n n) ;2 (n ; (n n) ; I (n n) ) ^ = n nx i= x i u 2 (^) = 2 n: If 2 is not known a priori, it can be estimated by solving the equation ie, 2 (^ z) = nx i= 2 = n ; (x i ; ^) 2 2 nx i= (x i ; ^) 2 = n ; which is the well known expression for the experimental standard deviation s 4 onclusion A general technique for evaluation of measurements by the method of Least Squares has been presented The applicability of the method has been demonstrated by two examples It has been shown that the method is fully compatible with the generally accepted principles for evaluation of measurement uncertainty laid down in the GUM and that ordinary linear regression is just a special case of the method The input to the method consists of An estimate of the value of each measured quantity, including anyrelevant inuence quantity The covariance matrix of these estimates formed by the standard uncertainties of the estimates and the correlation coecients between the estimates A :

17 Evaluation of Measurements 7 A measurement model describing all the known relations between the measured quantities and some additional quantities (if needed) for which no prior information is available The output of the method consists of An adjusted estimate of the value of each measured quantity and an estimate of each additional quantity introduced in the measurement model The covariance matrix of all these estimates from which the standard uncertainties and correlation coecients can be calculated Achi-square value which is a measure of the degree of consistency between the measurement model, the input estimates, and the covariances of the input quantities The adjusted estimate of the value of a measured quantity diers from the input estimate only if the measurement model imposes additional information regarding the value of that particular quantity In that case the standard uncertainty of the adjusted estimate will be smaller than the standard uncertainty of the input estimate For a good measurement, the dierence between the adjusted estimate and the input estimate of a measured quantity should not be large compared to the standard uncertainty of that dierence It has therefore been suggested that the ratio d of the dierence to its standard uncertainty is calculated and assessed against a selected criterion, eg jdj < 2 y plotting the d values of the adjusted estimates it is possible to assess whether a too high chi-square value is caused by a few poor input estimates or is due to a poor model ibliography IPM, IE, IF, ISO, IUPA, IUPAP, OIML, Guide to the expression of uncertainty in measurement, ISO, L Nielsen, Least-squares estimation using Lagrange multipliers, Metrologia 35 (998), 5{8 Erratum, Metrologia 37 (2), 83 3 L Nielsen, Evaluation of the calibration history of a measurement standard, DFM report DFM--R25, 2, {6 4 W H Press, S A Teukolsky, W T Vetterling and P Flannery, Numerical Recipies in, 2nd ed, ambridge, ambridge University Press, 992, 36{4 and 68{688 5 T L Saaty and J ram, Nonlinear Mathematics, Dover Publications, New York, 98, 93{95 6 K Weise and W Woger, Uncertainty and Measurement Data Evaluation, Wiley- VH, 999, 83{224 [in German]

Introduction to Logistic Regression

Introduction to Logistic Regression OpenStax-CNX module: m42090 1 Introduction to Logistic Regression Dan Calderon This work is produced by OpenStax-CNX and licensed under the Creative Commons Attribution License 3.0 Abstract Gives introduction

More information

Factor Analysis. Chapter 420. Introduction

Factor Analysis. Chapter 420. Introduction Chapter 420 Introduction (FA) is an exploratory technique applied to a set of observed variables that seeks to find underlying factors (subsets of variables) from which the observed variables were generated.

More information

Introduction to General and Generalized Linear Models

Introduction to General and Generalized Linear Models Introduction to General and Generalized Linear Models General Linear Models - part I Henrik Madsen Poul Thyregod Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby

More information

Linear Threshold Units

Linear Threshold Units Linear Threshold Units w x hx (... w n x n w We assume that each feature x j and each weight w j is a real number (we will relax this later) We will study three different algorithms for learning linear

More information

Least Squares Estimation

Least Squares Estimation Least Squares Estimation SARA A VAN DE GEER Volume 2, pp 1041 1045 in Encyclopedia of Statistics in Behavioral Science ISBN-13: 978-0-470-86080-9 ISBN-10: 0-470-86080-4 Editors Brian S Everitt & David

More information

Part 2: Analysis of Relationship Between Two Variables

Part 2: Analysis of Relationship Between Two Variables Part 2: Analysis of Relationship Between Two Variables Linear Regression Linear correlation Significance Tests Multiple regression Linear Regression Y = a X + b Dependent Variable Independent Variable

More information

NCSS Statistical Software Principal Components Regression. In ordinary least squares, the regression coefficients are estimated using the formula ( )

NCSS Statistical Software Principal Components Regression. In ordinary least squares, the regression coefficients are estimated using the formula ( ) Chapter 340 Principal Components Regression Introduction is a technique for analyzing multiple regression data that suffer from multicollinearity. When multicollinearity occurs, least squares estimates

More information

99.37, 99.38, 99.38, 99.39, 99.39, 99.39, 99.39, 99.40, 99.41, 99.42 cm

99.37, 99.38, 99.38, 99.39, 99.39, 99.39, 99.39, 99.40, 99.41, 99.42 cm Error Analysis and the Gaussian Distribution In experimental science theory lives or dies based on the results of experimental evidence and thus the analysis of this evidence is a critical part of the

More information

Simple Linear Regression Inference

Simple Linear Regression Inference Simple Linear Regression Inference 1 Inference requirements The Normality assumption of the stochastic term e is needed for inference even if it is not a OLS requirement. Therefore we have: Interpretation

More information

On computer algebra-aided stability analysis of dierence schemes generated by means of Gr obner bases

On computer algebra-aided stability analysis of dierence schemes generated by means of Gr obner bases On computer algebra-aided stability analysis of dierence schemes generated by means of Gr obner bases Vladimir Gerdt 1 Yuri Blinkov 2 1 Laboratory of Information Technologies Joint Institute for Nuclear

More information

constraint. Let us penalize ourselves for making the constraint too big. We end up with a

constraint. Let us penalize ourselves for making the constraint too big. We end up with a Chapter 4 Constrained Optimization 4.1 Equality Constraints (Lagrangians) Suppose we have a problem: Maximize 5, (x 1, 2) 2, 2(x 2, 1) 2 subject to x 1 +4x 2 =3 If we ignore the constraint, we get the

More information

Revenue Management with Correlated Demand Forecasting

Revenue Management with Correlated Demand Forecasting Revenue Management with Correlated Demand Forecasting Catalina Stefanescu Victor DeMiguel Kristin Fridgeirsdottir Stefanos Zenios 1 Introduction Many airlines are struggling to survive in today's economy.

More information

Introduction to Matrix Algebra

Introduction to Matrix Algebra Psychology 7291: Multivariate Statistics (Carey) 8/27/98 Matrix Algebra - 1 Introduction to Matrix Algebra Definitions: A matrix is a collection of numbers ordered by rows and columns. It is customary

More information

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS Systems of Equations and Matrices Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a

More information

International College of Economics and Finance Syllabus Probability Theory and Introductory Statistics

International College of Economics and Finance Syllabus Probability Theory and Introductory Statistics International College of Economics and Finance Syllabus Probability Theory and Introductory Statistics Lecturer: Mikhail Zhitlukhin. 1. Course description Probability Theory and Introductory Statistics

More information

Example: Credit card default, we may be more interested in predicting the probabilty of a default than classifying individuals as default or not.

Example: Credit card default, we may be more interested in predicting the probabilty of a default than classifying individuals as default or not. Statistical Learning: Chapter 4 Classification 4.1 Introduction Supervised learning with a categorical (Qualitative) response Notation: - Feature vector X, - qualitative response Y, taking values in C

More information

Simple Regression Theory II 2010 Samuel L. Baker

Simple Regression Theory II 2010 Samuel L. Baker SIMPLE REGRESSION THEORY II 1 Simple Regression Theory II 2010 Samuel L. Baker Assessing how good the regression equation is likely to be Assignment 1A gets into drawing inferences about how close the

More information

SOLVING LINEAR SYSTEMS

SOLVING LINEAR SYSTEMS SOLVING LINEAR SYSTEMS Linear systems Ax = b occur widely in applied mathematics They occur as direct formulations of real world problems; but more often, they occur as a part of the numerical analysis

More information

Multivariate Analysis of Variance (MANOVA)

Multivariate Analysis of Variance (MANOVA) Chapter 415 Multivariate Analysis of Variance (MANOVA) Introduction Multivariate analysis of variance (MANOVA) is an extension of common analysis of variance (ANOVA). In ANOVA, differences among various

More information

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS 1. SYSTEMS OF EQUATIONS AND MATRICES 1.1. Representation of a linear system. The general system of m equations in n unknowns can be written a 11 x 1 + a 12 x 2 +

More information

Data analysis in supersaturated designs

Data analysis in supersaturated designs Statistics & Probability Letters 59 (2002) 35 44 Data analysis in supersaturated designs Runze Li a;b;, Dennis K.J. Lin a;b a Department of Statistics, The Pennsylvania State University, University Park,

More information

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model 1 September 004 A. Introduction and assumptions The classical normal linear regression model can be written

More information

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION Introduction In the previous chapter, we explored a class of regression models having particularly simple analytical

More information

Estimates of Uncertainty of the Calibration of Balances

Estimates of Uncertainty of the Calibration of Balances Estimates of Uncertainty of the Calibration of Balances Chart shows the uncertainty (estimates) of the calibration of representative electronic balances. Chart. Uncertainty of calibration (estimates) Minimum

More information

Statistics for Business Decision Making

Statistics for Business Decision Making Statistics for Business Decision Making Faculty of Economics University of Siena 1 / 62 You should be able to: ˆ Summarize and uncover any patterns in a set of multivariate data using the (FM) ˆ Apply

More information

Two-Sample T-Tests Assuming Equal Variance (Enter Means)

Two-Sample T-Tests Assuming Equal Variance (Enter Means) Chapter 4 Two-Sample T-Tests Assuming Equal Variance (Enter Means) Introduction This procedure provides sample size and power calculations for one- or two-sided two-sample t-tests when the variances of

More information

Estimating Weighing Uncertainty From Balance Data Sheet Specifications

Estimating Weighing Uncertainty From Balance Data Sheet Specifications Estimating Weighing Uncertainty From Balance Data Sheet Specifications Sources Of Measurement Deviations And Uncertainties Determination Of The Combined Measurement Bias Estimation Of The Combined Measurement

More information

Mehtap Ergüven Abstract of Ph.D. Dissertation for the degree of PhD of Engineering in Informatics

Mehtap Ergüven Abstract of Ph.D. Dissertation for the degree of PhD of Engineering in Informatics INTERNATIONAL BLACK SEA UNIVERSITY COMPUTER TECHNOLOGIES AND ENGINEERING FACULTY ELABORATION OF AN ALGORITHM OF DETECTING TESTS DIMENSIONALITY Mehtap Ergüven Abstract of Ph.D. Dissertation for the degree

More information

Lean Six Sigma Analyze Phase Introduction. TECH 50800 QUALITY and PRODUCTIVITY in INDUSTRY and TECHNOLOGY

Lean Six Sigma Analyze Phase Introduction. TECH 50800 QUALITY and PRODUCTIVITY in INDUSTRY and TECHNOLOGY TECH 50800 QUALITY and PRODUCTIVITY in INDUSTRY and TECHNOLOGY Before we begin: Turn on the sound on your computer. There is audio to accompany this presentation. Audio will accompany most of the online

More information

STATISTICA Formula Guide: Logistic Regression. Table of Contents

STATISTICA Formula Guide: Logistic Regression. Table of Contents : Table of Contents... 1 Overview of Model... 1 Dispersion... 2 Parameterization... 3 Sigma-Restricted Model... 3 Overparameterized Model... 4 Reference Coding... 4 Model Summary (Summary Tab)... 5 Summary

More information

Understanding and Applying Kalman Filtering

Understanding and Applying Kalman Filtering Understanding and Applying Kalman Filtering Lindsay Kleeman Department of Electrical and Computer Systems Engineering Monash University, Clayton 1 Introduction Objectives: 1. Provide a basic understanding

More information

4. Continuous Random Variables, the Pareto and Normal Distributions

4. Continuous Random Variables, the Pareto and Normal Distributions 4. Continuous Random Variables, the Pareto and Normal Distributions A continuous random variable X can take any value in a given range (e.g. height, weight, age). The distribution of a continuous random

More information

Section 14 Simple Linear Regression: Introduction to Least Squares Regression

Section 14 Simple Linear Regression: Introduction to Least Squares Regression Slide 1 Section 14 Simple Linear Regression: Introduction to Least Squares Regression There are several different measures of statistical association used for understanding the quantitative relationship

More information

Gravimetric determination of pipette errors

Gravimetric determination of pipette errors Gravimetric determination of pipette errors In chemical measurements (for instance in titrimetric analysis) it is very important to precisely measure amount of liquid, the measurement is performed with

More information

AP Physics 1 and 2 Lab Investigations

AP Physics 1 and 2 Lab Investigations AP Physics 1 and 2 Lab Investigations Student Guide to Data Analysis New York, NY. College Board, Advanced Placement, Advanced Placement Program, AP, AP Central, and the acorn logo are registered trademarks

More information

11 Linear and Quadratic Discriminant Analysis, Logistic Regression, and Partial Least Squares Regression

11 Linear and Quadratic Discriminant Analysis, Logistic Regression, and Partial Least Squares Regression Frank C Porter and Ilya Narsky: Statistical Analysis Techniques in Particle Physics Chap. c11 2013/9/9 page 221 le-tex 221 11 Linear and Quadratic Discriminant Analysis, Logistic Regression, and Partial

More information

Calculating P-Values. Parkland College. Isela Guerra Parkland College. Recommended Citation

Calculating P-Values. Parkland College. Isela Guerra Parkland College. Recommended Citation Parkland College A with Honors Projects Honors Program 2014 Calculating P-Values Isela Guerra Parkland College Recommended Citation Guerra, Isela, "Calculating P-Values" (2014). A with Honors Projects.

More information

1 Example of Time Series Analysis by SSA 1

1 Example of Time Series Analysis by SSA 1 1 Example of Time Series Analysis by SSA 1 Let us illustrate the 'Caterpillar'-SSA technique [1] by the example of time series analysis. Consider the time series FORT (monthly volumes of fortied wine sales

More information

CURVE FITTING LEAST SQUARES APPROXIMATION

CURVE FITTING LEAST SQUARES APPROXIMATION CURVE FITTING LEAST SQUARES APPROXIMATION Data analysis and curve fitting: Imagine that we are studying a physical system involving two quantities: x and y Also suppose that we expect a linear relationship

More information

Summary of Formulas and Concepts. Descriptive Statistics (Ch. 1-4)

Summary of Formulas and Concepts. Descriptive Statistics (Ch. 1-4) Summary of Formulas and Concepts Descriptive Statistics (Ch. 1-4) Definitions Population: The complete set of numerical information on a particular quantity in which an investigator is interested. We assume

More information

Gaussian Conjugate Prior Cheat Sheet

Gaussian Conjugate Prior Cheat Sheet Gaussian Conjugate Prior Cheat Sheet Tom SF Haines 1 Purpose This document contains notes on how to handle the multivariate Gaussian 1 in a Bayesian setting. It focuses on the conjugate prior, its Bayesian

More information

How To Understand And Solve A Linear Programming Problem

How To Understand And Solve A Linear Programming Problem At the end of the lesson, you should be able to: Chapter 2: Systems of Linear Equations and Matrices: 2.1: Solutions of Linear Systems by the Echelon Method Define linear systems, unique solution, inconsistent,

More information

Least-Squares Intersection of Lines

Least-Squares Intersection of Lines Least-Squares Intersection of Lines Johannes Traa - UIUC 2013 This write-up derives the least-squares solution for the intersection of lines. In the general case, a set of lines will not intersect at a

More information

Finite cloud method: a true meshless technique based on a xed reproducing kernel approximation

Finite cloud method: a true meshless technique based on a xed reproducing kernel approximation INTERNATIONAL JOURNAL FOR NUMERICAL METHODS IN ENGINEERING Int. J. Numer. Meth. Engng 2001; 50:2373 2410 Finite cloud method: a true meshless technique based on a xed reproducing kernel approximation N.

More information

Balance Accuracy. FlexiWeigh

Balance Accuracy. FlexiWeigh FlexiWeigh The balance is an integral part of the FlexiWeigh workstation. The role and limitations of the balance are important to understand in this environment. This technical note describes the different

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.cs.toronto.edu/~rsalakhu/ Lecture 6 Three Approaches to Classification Construct

More information

Two-Sample T-Tests Allowing Unequal Variance (Enter Difference)

Two-Sample T-Tests Allowing Unequal Variance (Enter Difference) Chapter 45 Two-Sample T-Tests Allowing Unequal Variance (Enter Difference) Introduction This procedure provides sample size and power calculations for one- or two-sided two-sample t-tests when no assumption

More information

1.5 Oneway Analysis of Variance

1.5 Oneway Analysis of Variance Statistics: Rosie Cornish. 200. 1.5 Oneway Analysis of Variance 1 Introduction Oneway analysis of variance (ANOVA) is used to compare several means. This method is often used in scientific or medical experiments

More information

Confidence Intervals for the Difference Between Two Means

Confidence Intervals for the Difference Between Two Means Chapter 47 Confidence Intervals for the Difference Between Two Means Introduction This procedure calculates the sample size necessary to achieve a specified distance from the difference in sample means

More information

JCGM 101:2008. First edition 2008

JCGM 101:2008. First edition 2008 Evaluation of measurement data Supplement 1 to the Guide to the expression of uncertainty in measurement Propagation of distributions using a Monte Carlo method Évaluation des données de mesure Supplément

More information

Oscillations of the Sending Window in Compound TCP

Oscillations of the Sending Window in Compound TCP Oscillations of the Sending Window in Compound TCP Alberto Blanc 1, Denis Collange 1, and Konstantin Avrachenkov 2 1 Orange Labs, 905 rue Albert Einstein, 06921 Sophia Antipolis, France 2 I.N.R.I.A. 2004

More information

Since it is necessary to consider the ability of the lter to predict many data over a period of time a more meaningful metric is the expected value of

Since it is necessary to consider the ability of the lter to predict many data over a period of time a more meaningful metric is the expected value of Chapter 11 Tutorial: The Kalman Filter Tony Lacey. 11.1 Introduction The Kalman lter ë1ë has long been regarded as the optimal solution to many tracing and data prediction tass, ë2ë. Its use in the analysis

More information

3.2. Solving quadratic equations. Introduction. Prerequisites. Learning Outcomes. Learning Style

3.2. Solving quadratic equations. Introduction. Prerequisites. Learning Outcomes. Learning Style Solving quadratic equations 3.2 Introduction A quadratic equation is one which can be written in the form ax 2 + bx + c = 0 where a, b and c are numbers and x is the unknown whose value(s) we wish to find.

More information

Some probability and statistics

Some probability and statistics Appendix A Some probability and statistics A Probabilities, random variables and their distribution We summarize a few of the basic concepts of random variables, usually denoted by capital letters, X,Y,

More information

Nonlinear Iterative Partial Least Squares Method

Nonlinear Iterative Partial Least Squares Method Numerical Methods for Determining Principal Component Analysis Abstract Factors Béchu, S., Richard-Plouet, M., Fernandez, V., Walton, J., and Fairley, N. (2016) Developments in numerical treatments for

More information

Module 3: Correlation and Covariance

Module 3: Correlation and Covariance Using Statistical Data to Make Decisions Module 3: Correlation and Covariance Tom Ilvento Dr. Mugdim Pašiƒ University of Delaware Sarajevo Graduate School of Business O ften our interest in data analysis

More information

HOW FAST IS THE STOCHASTIC EVALUATION OT THE EXPECTED PERFORMANCE AND ITS PARAMETRIC SENSITIVITY Michael Shalmon Universite du Quebec INRS-Telecommunications 16 Place du Commerce, Verdun, Quebec, Canada,

More information

individualdifferences

individualdifferences 1 Simple ANalysis Of Variance (ANOVA) Oftentimes we have more than two groups that we want to compare. The purpose of ANOVA is to allow us to compare group means from several independent samples. In general,

More information

The Method of Least Squares

The Method of Least Squares Hervé Abdi 1 1 Introduction The least square methods (LSM) is probably the most popular technique in statistics. This is due to several factors. First, most common estimators can be casted within this

More information

Regression III: Advanced Methods

Regression III: Advanced Methods Lecture 16: Generalized Additive Models Regression III: Advanced Methods Bill Jacoby Michigan State University http://polisci.msu.edu/jacoby/icpsr/regress3 Goals of the Lecture Introduce Additive Models

More information

Biostatistics: DESCRIPTIVE STATISTICS: 2, VARIABILITY

Biostatistics: DESCRIPTIVE STATISTICS: 2, VARIABILITY Biostatistics: DESCRIPTIVE STATISTICS: 2, VARIABILITY 1. Introduction Besides arriving at an appropriate expression of an average or consensus value for observations of a population, it is important to

More information

Exploratory Factor Analysis and Principal Components. Pekka Malo & Anton Frantsev 30E00500 Quantitative Empirical Research Spring 2016

Exploratory Factor Analysis and Principal Components. Pekka Malo & Anton Frantsev 30E00500 Quantitative Empirical Research Spring 2016 and Principal Components Pekka Malo & Anton Frantsev 30E00500 Quantitative Empirical Research Spring 2016 Agenda Brief History and Introductory Example Factor Model Factor Equation Estimation of Loadings

More information

CHI-SQUARE: TESTING FOR GOODNESS OF FIT

CHI-SQUARE: TESTING FOR GOODNESS OF FIT CHI-SQUARE: TESTING FOR GOODNESS OF FIT In the previous chapter we discussed procedures for fitting a hypothesized function to a set of experimental data points. Such procedures involve minimizing a quantity

More information

How to report the percentage of explained common variance in exploratory factor analysis

How to report the percentage of explained common variance in exploratory factor analysis UNIVERSITAT ROVIRA I VIRGILI How to report the percentage of explained common variance in exploratory factor analysis Tarragona 2013 Please reference this document as: Lorenzo-Seva, U. (2013). How to report

More information

Non-Inferiority Tests for Two Means using Differences

Non-Inferiority Tests for Two Means using Differences Chapter 450 on-inferiority Tests for Two Means using Differences Introduction This procedure computes power and sample size for non-inferiority tests in two-sample designs in which the outcome is a continuous

More information

Lecture 3: Linear methods for classification

Lecture 3: Linear methods for classification Lecture 3: Linear methods for classification Rafael A. Irizarry and Hector Corrada Bravo February, 2010 Today we describe four specific algorithms useful for classification problems: linear regression,

More information

1 Portfolio mean and variance

1 Portfolio mean and variance Copyright c 2005 by Karl Sigman Portfolio mean and variance Here we study the performance of a one-period investment X 0 > 0 (dollars) shared among several different assets. Our criterion for measuring

More information

1 Teaching notes on GMM 1.

1 Teaching notes on GMM 1. Bent E. Sørensen January 23, 2007 1 Teaching notes on GMM 1. Generalized Method of Moment (GMM) estimation is one of two developments in econometrics in the 80ies that revolutionized empirical work in

More information

SECOND DERIVATIVE TEST FOR CONSTRAINED EXTREMA

SECOND DERIVATIVE TEST FOR CONSTRAINED EXTREMA SECOND DERIVATIVE TEST FOR CONSTRAINED EXTREMA This handout presents the second derivative test for a local extrema of a Lagrange multiplier problem. The Section 1 presents a geometric motivation for the

More information

Fall 2011. Andrew U. Frank, October 23, 2011

Fall 2011. Andrew U. Frank, October 23, 2011 Fall 2011 Andrew U. Frank, October 23, 2011 TU Wien, Department of Geoinformation and Cartography, Gusshausstrasse 27-29/E127.1 A-1040 Vienna, Austria frank@geoinfo.tuwien.ac.at Part I Introduction 1 The

More information

Exact Confidence Intervals

Exact Confidence Intervals Math 541: Statistical Theory II Instructor: Songfeng Zheng Exact Confidence Intervals Confidence intervals provide an alternative to using an estimator ˆθ when we wish to estimate an unknown parameter

More information

Nonlinear Regression:

Nonlinear Regression: Zurich University of Applied Sciences School of Engineering IDP Institute of Data Analysis and Process Design Nonlinear Regression: A Powerful Tool With Considerable Complexity Half-Day : Improved Inference

More information

Exercise 1.12 (Pg. 22-23)

Exercise 1.12 (Pg. 22-23) Individuals: The objects that are described by a set of data. They may be people, animals, things, etc. (Also referred to as Cases or Records) Variables: The characteristics recorded about each individual.

More information

A Primer on Mathematical Statistics and Univariate Distributions; The Normal Distribution; The GLM with the Normal Distribution

A Primer on Mathematical Statistics and Univariate Distributions; The Normal Distribution; The GLM with the Normal Distribution A Primer on Mathematical Statistics and Univariate Distributions; The Normal Distribution; The GLM with the Normal Distribution PSYC 943 (930): Fundamentals of Multivariate Modeling Lecture 4: September

More information

(Uncertainty) 2. How uncertain is your uncertainty budget?

(Uncertainty) 2. How uncertain is your uncertainty budget? (Uncertainty) 2 How uncertain is your uncertainty budget? Paper Author and Presenter: Dr. Henrik S. Nielsen HN Metrology Consulting, Inc 10219 Coral Reef Way, Indianapolis, IN 46256 Phone: (317) 849 9577,

More information

Preliminary Version { Comments Welcome

Preliminary Version { Comments Welcome Option Pricing under the Mixture of Distributions Hypothesis Marco Neumann Diskussionspapier Nr. 208 First Version: December 14, 1997 Current Version: June 27, 1998 Preliminary Version { Comments Welcome

More information

Copulas in Financial Risk Management Dr Jorn Rank Department for Continuing Education University of Oxford A thesis submitted for the diploma in Mathematical Finance 4 August 2000 Diploma in Mathematical

More information

(Quasi-)Newton methods

(Quasi-)Newton methods (Quasi-)Newton methods 1 Introduction 1.1 Newton method Newton method is a method to find the zeros of a differentiable non-linear function g, x such that g(x) = 0, where g : R n R n. Given a starting

More information

Machine Learning and Data Mining. Regression Problem. (adapted from) Prof. Alexander Ihler

Machine Learning and Data Mining. Regression Problem. (adapted from) Prof. Alexander Ihler Machine Learning and Data Mining Regression Problem (adapted from) Prof. Alexander Ihler Overview Regression Problem Definition and define parameters ϴ. Prediction using ϴ as parameters Measure the error

More information

1 Review of Least Squares Solutions to Overdetermined Systems

1 Review of Least Squares Solutions to Overdetermined Systems cs4: introduction to numerical analysis /9/0 Lecture 7: Rectangular Systems and Numerical Integration Instructor: Professor Amos Ron Scribes: Mark Cowlishaw, Nathanael Fillmore Review of Least Squares

More information

STRUTS: Statistical Rules of Thumb. Seattle, WA

STRUTS: Statistical Rules of Thumb. Seattle, WA STRUTS: Statistical Rules of Thumb Gerald van Belle Departments of Environmental Health and Biostatistics University ofwashington Seattle, WA 98195-4691 Steven P. Millard Probability, Statistics and Information

More information

LOGISTIC REGRESSION. Nitin R Patel. where the dependent variable, y, is binary (for convenience we often code these values as

LOGISTIC REGRESSION. Nitin R Patel. where the dependent variable, y, is binary (for convenience we often code these values as LOGISTIC REGRESSION Nitin R Patel Logistic regression extends the ideas of multiple linear regression to the situation where the dependent variable, y, is binary (for convenience we often code these values

More information

Point Biserial Correlation Tests

Point Biserial Correlation Tests Chapter 807 Point Biserial Correlation Tests Introduction The point biserial correlation coefficient (ρ in this chapter) is the product-moment correlation calculated between a continuous random variable

More information

Elasticity Theory Basics

Elasticity Theory Basics G22.3033-002: Topics in Computer Graphics: Lecture #7 Geometric Modeling New York University Elasticity Theory Basics Lecture #7: 20 October 2003 Lecturer: Denis Zorin Scribe: Adrian Secord, Yotam Gingold

More information

Sales forecasting # 2

Sales forecasting # 2 Sales forecasting # 2 Arthur Charpentier arthur.charpentier@univ-rennes1.fr 1 Agenda Qualitative and quantitative methods, a very general introduction Series decomposition Short versus long term forecasting

More information

Variance Reduction. Pricing American Options. Monte Carlo Option Pricing. Delta and Common Random Numbers

Variance Reduction. Pricing American Options. Monte Carlo Option Pricing. Delta and Common Random Numbers Variance Reduction The statistical efficiency of Monte Carlo simulation can be measured by the variance of its output If this variance can be lowered without changing the expected value, fewer replications

More information

5. Linear Regression

5. Linear Regression 5. Linear Regression Outline.................................................................... 2 Simple linear regression 3 Linear model............................................................. 4

More information

Dimensionality Reduction: Principal Components Analysis

Dimensionality Reduction: Principal Components Analysis Dimensionality Reduction: Principal Components Analysis In data mining one often encounters situations where there are a large number of variables in the database. In such situations it is very likely

More information

Introduction; Descriptive & Univariate Statistics

Introduction; Descriptive & Univariate Statistics Introduction; Descriptive & Univariate Statistics I. KEY COCEPTS A. Population. Definitions:. The entire set of members in a group. EXAMPLES: All U.S. citizens; all otre Dame Students. 2. All values of

More information

Gamma Distribution Fitting

Gamma Distribution Fitting Chapter 552 Gamma Distribution Fitting Introduction This module fits the gamma probability distributions to a complete or censored set of individual or grouped data values. It outputs various statistics

More information

CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES. From Exploratory Factor Analysis Ledyard R Tucker and Robert C.

CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES. From Exploratory Factor Analysis Ledyard R Tucker and Robert C. CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES From Exploratory Factor Analysis Ledyard R Tucker and Robert C MacCallum 1997 180 CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES In

More information

Review Jeopardy. Blue vs. Orange. Review Jeopardy

Review Jeopardy. Blue vs. Orange. Review Jeopardy Review Jeopardy Blue vs. Orange Review Jeopardy Jeopardy Round Lectures 0-3 Jeopardy Round $200 How could I measure how far apart (i.e. how different) two observations, y 1 and y 2, are from each other?

More information

Operation Count; Numerical Linear Algebra

Operation Count; Numerical Linear Algebra 10 Operation Count; Numerical Linear Algebra 10.1 Introduction Many computations are limited simply by the sheer number of required additions, multiplications, or function evaluations. If floating-point

More information

E3: PROBABILITY AND STATISTICS lecture notes

E3: PROBABILITY AND STATISTICS lecture notes E3: PROBABILITY AND STATISTICS lecture notes 2 Contents 1 PROBABILITY THEORY 7 1.1 Experiments and random events............................ 7 1.2 Certain event. Impossible event............................

More information

2 ESTIMATION. Objectives. 2.0 Introduction

2 ESTIMATION. Objectives. 2.0 Introduction 2 ESTIMATION Chapter 2 Estimation Objectives After studying this chapter you should be able to calculate confidence intervals for the mean of a normal distribution with unknown variance; be able to calculate

More information

Liski and Montero As a Successful Solution to the Recession

Liski and Montero As a Successful Solution to the Recession University of Tübingen Working Papers in Economics and Finance No. 75 Forward Trading and Collusion of Firms in Volatile Markets by Markus Aichele Faculty of Economics and Social Sciences www.wiwi.uni-tuebingen.de

More information

DEGREES OF FREEDOM - SIMPLIFIED

DEGREES OF FREEDOM - SIMPLIFIED 1 Aust. J. Geod. Photogram. Surv. Nos 46 & 47 December, 1987. pp 57-68 In 009 I retyped this paper and changed symbols eg ˆo σ to VF for my students benefit DEGREES OF FREEDOM - SIMPLIFIED Bruce R. Harvey

More information

1) Write the following as an algebraic expression using x as the variable: Triple a number subtracted from the number

1) Write the following as an algebraic expression using x as the variable: Triple a number subtracted from the number 1) Write the following as an algebraic expression using x as the variable: Triple a number subtracted from the number A. 3(x - x) B. x 3 x C. 3x - x D. x - 3x 2) Write the following as an algebraic expression

More information

Uncertainty of Force Measurements

Uncertainty of Force Measurements European Association of National Metrology Institutes Uncertainty of Force Measurements EURAMET cg-4 Version.0 (03/011) Previously EA-10/04 Calibration Guide EURAMET cg-4 Version.0 (03/011) UNCERTAINTY

More information

Logistic Regression (1/24/13)

Logistic Regression (1/24/13) STA63/CBB540: Statistical methods in computational biology Logistic Regression (/24/3) Lecturer: Barbara Engelhardt Scribe: Dinesh Manandhar Introduction Logistic regression is model for regression used

More information