Kernel density estimation with adaptive varying window size

Size: px
Start display at page:

Download "Kernel density estimation with adaptive varying window size"

Transcription

1 Pattern Recognition Letters 23 (2002) Kernel density estimation with adaptive varying window size Vladimir Katkovnik a, Ilya Shmulevich b, * a Kwangju Institute of Science and Technology, Kwangju, South Korea b University of Texas M.D. Anderson Cancer Center, Houston, TX 77030, USA Received 2 November 2000; received in revised form 2 August 200 Abstract A new method of kernel density estimation with a varying adaptive window size is proposed. It is based on the socalled intersection of confidence intervals (ICI) rule. Several examples of the proposed method are given for different types of densities and the quality of the adaptive density estimate is assessed by means of numerical simulations. Ó 2002 Elsevier Science B.V. All rights reserved. Keywords: Non-parametric; Kernel; Density estimation; Parzen; ICI rule. Introduction In pattern recognition, optimal algorithms often require the knowledge of underlying densities of signals and/or noise. As these densities are usually unknown, unrealistic assumptions are frequently made, thus compromising the performance of the algorithms in question. A common approach to this problem is to estimate the density from the data. If a particular form of the density is assumed or known, then parametric estimation is used. If nothing is assumed about the density shape, non-parametric estimation is employed. Besides being widely used in the field of pattern recognition and classification (Fukunaga, 990), non-parametric probability density estimation has * Corresponding author. Tel.: ; fax: addresses: vlkatkov@hotmail.com (V. Katkovnik), is@ieee.org (I. Shmulevich). been applied in image processing (Sindoukas et al., 997; Wright et al., 997), communications (Zabin and Wright, 2000), and many other fields. One of the most well-known and popular techniques of non-parametric density estimation is the kernel or Parzen density estimate (Fukunaga, 990; Parzen, 962; Cacoullos, 966). Given N samples X ;...; X N drawn from a population with density function f ðxþ, x 2 R, the Parzen density estimate at x is given by ^f h ðxþ ¼ N X N i¼ h j x X i h ; ðþ where jðþ is a window or kernel function and h is the window width, smoothing parameter, or simply the R kernel size. Traditionally, it is assumed that jðuþdu ¼ and jðþ is symmetric, that is, jðuþ ¼ jð uþ. One p popular choice is the Gaussian kernel jðuþ ¼= ffiffiffiffiffi 2p expð u 2 =2Þ. The kernel size h is the most important characteristic of the Parzen density estimate (Raudys, /02/$ - see front matter Ó 2002 Elsevier Science B.V. All rights reserved. PII: S (02)0027-7

2 642 V. Katkovnik, I. Shmulevich / Pattern Recognition Letters 23 (2002) ; Silverman, 978). One can compute the ideal or optimal value of h by minimizing the meansquare error MSEff ^ h ðxþg ¼ Ef½f ^ h ðxþ f ðxþš 2 g ð2þ between the true and estimated densities, with respect to h. The MSE is a function of x and so the optimal kernel size h is also a function of x. In order to minimize the MSE, a best compromise between variance and bias must be selected. Using Taylor series approximations of the moments of f^ h ðxþ and noting that MSEff ^ h ðxþg ¼ ½Eff ^ h ðxþg f ðxþš 2 þ Varff ^ h ðxþg; ð3þ where Ef ^ f h ðxþg ¼ and Z Varf ^ f h ðxþg ¼ N jðuþf ðx þ huþdu; h Z E 2 f ^ f h ðxþg j 2 ðuþf ðx þ huþ ð4þ : ð5þ We have that for small h and large N (h! 0, N!, and Nh!), Biasff ^ h ðxþg ¼ Eff ^ h ðxþg f ðxþ 8 < h 2 2 f 00 ðxþ R u 2 jðuþdu; if jðuþ ¼jð uþ : hf 0 ðxþ R ujðuþdu; otherwise ð6þ and Varff ^ h ðxþg f ðxþ Z Nh j 2 ðuþdu: ð7þ Then, the optimal value of the kernel size can be shown to be equal to (Parzen, 962) 0 f ðxþ R =5 B j 2 ðuþdu h 0 ðxþ ¼ h N f 00 ðxþ R i 2 A ð8þ u 2 jðuþdu for a symmetric kernel, when h! 0asN! and Nh!, assuring asymptotic unbiasedness and consistency of the estimate. Similarly, 0 B f ðxþ R j 2 ðuþdu h 0 ðxþ ¼ h 2N f 0 ðxþ R i 2 A ujðuþdu =3 ð9þ for a non-symmetric window. As can be seen from Eqs. (8) and (9), the optimal kernel size depends on the value of the density and on its second or first derivative. These equations, in particular the order with respect to N, depend essentially on whether the kernel is symmetric or non-symmetric. It is also possible to obtain an optimal constant kernel size independent of x by minimizing either the integral mean-square error R MSEff ^ h ðxþgdx or the expected mean-square error R MSEff ^ h ðxþgf ðxþdx (Fukunaga, 990; Parzen, 962). Clearly, in practice, one does not have access to the true density function f ðxþ which is proposed to be estimated. Thus, a number of heuristic approaches can be taken for finding the window width. For instance, the optimal constant h can be computed for, say, the Normal distribution (as a function of N) and then used for making the estimate f ^ h ðxþ (Fukunaga, 990). Since the density estimates are often used for classification purposes, another approach is to determine h on the basis of the expected probability of misclassification (Raudys, 99). Although h is usually taken to be a constant, several important approaches have been proposed to vary it. One is the well-known kth nearest neighbor estimate (Loftsgaarden and Quesenberry, 965). Another method is the adaptive kernel estimate proposed in (Breiman et al., 977). A number of other papers considering the problem of kernel size selection exist (e.g. Abramson, 982; Terrell and Scott, 992; Chiu, 992; Sheather and Jones, 99; Hall et al., 99; Taylor, 989). In this paper, we propose and develop a special rule (statistic) for choosing a data-driven kernel size that is selected in a point-wise manner for every argument value of the density. Our main motivation for this complex estimator is to make it adaptive to unknown and varying smoothness of the density to be estimated. This method is described in Section 2. In Section 3, we assess the quality of the density estimate by comparing it to a known density which we are

3 V. Katkovnik, I. Shmulevich / Pattern Recognition Letters 23 (2002) estimating, under the mean-square error criterion. It will be shown that the estimates based on the variable-sized kernels are superior to the estimates based on optimal constant-sized kernels. 2. Adaptive kernel size selection The method of adaptive kernel size selection is based on the ICI rule, proposed in (Goldenshluger and Nemirovsky, 997) for adaptive regression smoothing and later developed for signal filtering in (Katkovnik, 999). One of several attractive properties of the ICI rule is that it is spatially adaptive over a wide range of signal classes in the sense that its quality is close to that which one could achieve if smoothness of the original signal was known in advance (Goldenshluger and Nemirovsky, 997). We briefly reintroduce this method here in the context of density estimation. Consider the ratio of the standard deviation Stdff ^ h ðxþg of the estimate to the absolute value of the bias jbiasff ^ h ðxþgj of the estimate, evaluated at the ideal value of h 0 ðxþ, given in Eqs. (8) and (9). We get Stdf ^ f h0 ðxþg jbiasf ^ f h0 ðxþgj ¼ k; k ¼ 2; p ffiffiffi if jðuþ ¼jð uþ 2 ; otherwise It is useful to note that as the standard deviation and bias are monotonically increasing and decreasing, respectively, as h! 0, we have that k Stdf f ^ h ðxþg P jbiasff ^ h ðxþgj; if h 6 h 0 ; k Stdf f ^ h ðxþg 6 jbiasff ^ h ðxþgj; if h P h 0 : ð0þ Although it is known that in regions where the density function is convex, it is theoretically possible to find bandwidths for which the point-wise bias is equal to zero (Sain and Scott, 996), we assume the monotonicity of the bias with respect to the bandwidth locally, for the theory and initial motivation. Moreover, these zero-bias bandwidths are typically much larger than the asymptotic data-dependent bandwidth given in (8), while Eq. (0) holds for h! 0. For a given kernel size h, the estimation error can be represented as jf ^ h ðxþ f ðxþj ¼ jbiasff ^ h ðxþg þ n h ðxþj 6 jbiasff ^ h ðxþgj þ jn h ðxþj; where n h ðxþ is a random variable with zero mean and standard deviation equal to Stdf ^ f h ðxþg. Thus, jf ^ h ðxþ f ðxþj 6 jbiasff ^ h ðxþgj þ v p Stdff ^ h ðxþg ðþ holds with an arbitrary probability p, for a suitably chosen v p. Using the relationships in (0) together with (), we get that for h 6 h 0, jf ^ h ðxþ f ðxþj 6 þ v k p Stdff ^ h ðxþg ¼ C Stdff ^ h ðxþg; ð2þ where C ¼ þ v k p. Larger values of C correspond to larger values of p. The ICI rule essentially tests the hypothesis h 6 h 0 for various values of h and in this way selects an h close to h 0 as follows. Suppose H ¼fh < h 2 < < h J g is a finite collection of kernel sizes, starting with a small h. Using inequality (2), we determine a sequence of confidence intervals D j ¼½L j ; U j Š; j ¼ ;...; J L j ¼ f ^ hj ðxþ C Stdff ^ hj ðxþg U j ¼ f ^ hj ðxþþc Stdff ^ hj ðxþg; ð3þ each one corresponding to a kernel size in H. The ICI rule, then, can be stated as follows (Katkovnik, 999). ICI rule: Consider the intersection of the intervals D j,6 j 6 i with increasing i and let i þ be the largest of those i for which the intervals D j, 6 j i have a point in common. This i þ defines the adaptive kernel size h þ ðxþ ¼h i þ and consequently, the density estimate f ^ h ðxþðxþ. þ It is important to note that the kernel size selection procedure based on the ICI rule requires only the knowledge of the density estimate and its variance, for which Eq. (7) can be used. However, this variance, in turn, depends on the unknown

4 644 V. Katkovnik, I. Shmulevich / Pattern Recognition Letters 23 (2002) density to be estimated. A pilot estimate of the density can be used in (7) instead of f ðxþ. However, it is emphasized that this pilot estimate should be obtained more or less independently with respect to the final estimate f ^ h ðxþðxþ. In fact, þ this is a general rule of using pilot estimates in statistics (see e.g., Fan and Gijbels, 996). The kernel density estimate with a constant window size h ¼ h is a good choice for the considered problem. In our simulation experiments, we employ the Sheather Jones plug in method (Sheather and Jones, 99) for the estimation of h in the pilot density estimate. This method is known to have excellent performance as compared to other known methods. C is a design parameter of the algorithm and the selection of its value is discussed in (Katkovnik and Shmulevich, 2000). This ICI procedure for the varying window density estimate can be implemented by Algorithm. Algorithm. Adaptive Window Width Selection L ( ; U ( while ðl 6 UÞ and ði 6 JÞ do L ( f ^ hi ðxþ C Stdff ^ h ðxþg U ( f ^ hi ðxþþc Stdff ^ h ðxþg L ( max½l; LŠ; U ( min½u; UŠ i ( i þ end while h þ ðxþ (h i 3. Simulation examples In this section, we will illustrate the use of the kernel size selection procedure based on the ICI rule. 3.. Qualitative simulation This group of simulations is given in order to demonstrate the ability of the ICI to obtain the optimal (reasonable) window sizes. As a first example, consider estimating the piece-wise constant density function shown in Fig. a. This example is intended to qualitatively demonstrate the behavior of the adaptive kernel size selection procedure, using the symmetric Gaussian kernel as well as non-symmetric right and left kernels 8 < p2 ffiffiffiffiffi exp u2 ; u P 0 j r ðuþ ¼ 2p 2 : 0; u < 0 ð4þ j l ðuþ ¼j r ð uþ: The allowable kernel sizes (H) start with h ¼ 0:0 and increase until h 300 ¼ 3:0 with a step of 0.0. Fig. b d shows the kernel sizes chosen by the ICI rule, corresponding to the three kinds of kernels (jðþ, j r ðþ, and j l ðþ) used. The number of observations N is equal to Especially worthy of notice is the behavior of non-symmetric kernels in the presence of discontinuities in the density function. For instance, the kernel size of the right kernel j r ðþ is rather high at the point of the first discontinuity (x ¼ ) and becomes smaller as it approaches the second discontinuity ðx ¼ 0Þ, after which the situation is similar (Fig. c). This behavior corresponds to the common sense idea that a large window size for the right kernel should be chosen at x ¼ þ while for x ¼ 0, the data available for the right kernel estimator is very small and hence, the kernel size is accordingly small. Here, the notation x means lim e!0;e>0 ðx eþ. For the left kernel j l ðþ, the behavior is the opposite (Fig. d). Immediately after the first discontinuity, the left kernel still contains very few observations and consequently has small size which increases up until x ¼ 0. Similarly, just after this point, the kernel size again becomes quite small, since even small sizes encompass two different density regions. Finally, as shown in Fig. b, the size of the symmetric kernel increases towards the middle between discontinuities, that is, at x ¼ 0:5. Also, very large kernel sizes, in the form of spikes, can be seen exactly at the points of discontinuities. The reason for this phenomenon is the following. It can be shown, by using Eq. (4), that at the point of discontinuity of the density function, the expectation of the estimate satisfies Z lim h!0 jðuþf ðx þ huþdu ¼ f ð xþ Þþfðx Þ : 2 Therefore, the ICI rule behaves correctly and in accordance with this fact. The reason that large kernel sizes are chosen is because the density

5 V. Katkovnik, I. Shmulevich / Pattern Recognition Letters 23 (2002) Fig.. Density f ðxþ to be estimated (a) and adaptive window widths corresponding to a symmetric kernel jðþ (b), right kernel j r ðþ (c), and left kernel j l ðþ (d) C ¼ 4. function on either side of the discontinuity is constant and larger kernel sizes decrease the variance and consequently, the MSE. The combined density estimates obtained by fusion of the left and right estimates are considered in (Katkovnik and Shmulevich, 2000). Another way to evaluate the method of adaptive kernel size selection is to compare it to an ideal varying kernel size. Rather than using Eq. (8), which is an asymptotic result, we shall compare our method to the more stringent criterion, namely, the empirically obtained varying kernel size h ðxþ, which minimizes the MSE between the known density f ðxþ and the estimated density f ^ h ðxþðxþ. In other words, h ðxþ ¼arg min hðxþ 2: E fðxþ f ^ hðxþ ðxþ ð5þ As a second example, we shall consider estimating the density function shown in Fig. 2a. The density is zero outside the interval ½0; Š. The allowable kernel sizes (H) start with h ¼ 0:0 and increase until h 00 ¼ :0 with a step of 0.0. The number of observations N is equal to An ideal kernel size, from the set of allowable kernel sizes, was found for every x, using Eq. (5). This ideal window size is shown in Fig. 2b as a solid line. The dashed line shows the variable kernel size h þ ðxþ obtained by using the ICI rule. As can be seen, their behavior is very similar. As expected, the kernel size is larger in the flat region of the density as compared with regions of the peaks where the kernel size becomes smaller Quantitative simulation For quantitative accuracy analysis, we use a double-peaked mixed density, f ðxþ ¼ð aþnð0; r ÞþaNðm 2 ; r 2 Þ, with m 2 ¼ 7, r ¼, r 2 ¼ 0:05, and a ¼ =2. This type of combined models is most reasonable for demonstrating the advantage of the varying window kernel estimates (Terrell and Scott, 992). The true density to be estimated is depicted in Fig. 3. The allowable kernel sizes (H)

6 646 V. Katkovnik, I. Shmulevich / Pattern Recognition Letters 23 (2002) Fig. 2. (a) Density f ðxþ to be estimated; (b) optimal varying window width h m ðxþ (solid line) minimizing the MSE between the known density and the estimate and the varying window width based on the ICI rule h þ ðxþ (dashed line) (C ¼ 5). point-wise average, over 200 simulation runs, of the Parzen estimate was obtained. That is, f hi ðxþ ¼ X 200 f^ ðjþ h 200 i ðxþ; ð6þ j¼ where f ^ðjþ h i ðxþ is the Parzen estimate with constant kernel size h i using data from the jth realization. For each run, the square root of the average, over 200 runs, of the point-wise squared error was computed. In other words, vffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi u X 200 h e hi ðxþ ¼ f ðxþ f 200 ^ i 2 t ðjþ h i ðxþ : ð7þ j¼ Fig. 3. The true density to be estimated. are logarithmically spaced between h ¼ 0:002 and h 40 ¼ 4:0. The number of observations N is equal to 000. The following computations were performed. For each constant kernel size in H, the The best (constant) kernel size, h, was selected by considering the mean value, over x, of the error in (7). In our simulations, the best kernel size was equal to h ¼ 0:024. We note in passing that the data-driven Sheather Jones plug in method (Sheather and Jones, 99) produces a kernel size roughly five times larger.

7 V. Katkovnik, I. Shmulevich / Pattern Recognition Letters 23 (2002) Fig. 4. The results of Monte Carlo simulations with 200 runs: (a) and (b) show two parts of the density using the ideal constant kernel size as well as the ICI rule. Average of the estimates as well as their confidence intervals are indicated. The true density is shown in dashed lines. Further, the ICI rule was used with values of C between 0.5 and 3.0 with a step size of 0.5. In a similar manner to the above, the point-wise average of the estimate equipped with the ICI rule was obtained. Also, the error, as in (7), was computed. This was performed for every value of C. Finally, the best C was chosen in the same way as h. In our simulations, C ¼ 0:5. It should be mentioned that the accuracy of estimation is not overly sensitive to the selection of C and there exists a range of values all of which result in similar accuracy (Katkovnik and Shmulevich, 2000). Fig. 4a and b depict the two peaks of the density estimated using the ideal constant kernel size h and the ICI rule. The true density is shown as a dashed line. The average of the 200 estimates given in (6) is indicated as mean const. and the corresponding average of the estimates using the ICI rule as mean ICI. Note that in Fig. 4b, the average estimated densities nearly coincide. In addition, these figures also show the upper and lower confidence intervals given by f hi ðxþe hi ðxþ for the ideal constant kernel size as well as for the ICI rule. These are labeled as upper/lower const. and upper/lower ICI respectively. It can be readily seen that on the wide peak of the density, the ICI rule produces smaller confidence intervals and hence, less variability of the error in the estimate. We stress that this is done in comparison with the ideal constant kernel size, which is, of course, unknown. Moreover, adaptive constant bandwidth methods (e.g., Sheather and Jones, 99) produce very different sizes from the ideal. As far as the average of the estimates, the ideal constant kernel size estimate and the estimate based on the ICI rule are comparable, with the ICI rule producing a slightly smoother estimate for the wide peak (Fig. 4a). 4. Conclusions We have proposed a new method for varying the bandwidth in kernel density estimation. This

8 648 V. Katkovnik, I. Shmulevich / Pattern Recognition Letters 23 (2002) method is based on the ICI rule and requires only the knowledge of the variance of the estimate. In our case, as the true density is unknown, the variance of the estimator is approximated by replacing the true density by a pilot estimate with a data-dependent constant kernel size. It is also possible to implement an iterative technique in which successive estimates are used to compute the variance by formula (5), using which in the ICI rule, new estimates are formed. Although we have considered this method for one-dimensional densities, there is no conceptual difficulty in extending it to multi-dimensional densities. In that case, as with other techniques, not only the size, but also the shape of the kernel is an important parameter. We have shown, by means of numerical simulations, that the proposed method can perform significantly better than any constant-bandwidth method. Acknowledgements The authors are grateful for the support and hospitality of Tampere International Center for Signal Processing in Tampere, Finland, where this work was done. References Abramson, I., 982. On bandwidth variation in kernel estimates a square root law. Ann. Stat. 0, Breiman, L., Meisel, W., Purcell, E., 977. Variable kernel estimates of multivariate densities. Technometrics 9, Cacoullos, T., 966. Estimation of a multivariate density. Ann. Inst. Stat. Math. 8, Chiu, S.-T., 992. An automatic bandwidth selector for kernel density estimation. Biometrika 79 (4), Fan, J., Gijbels, I., 996. Local Polynomial Modelling and its Application. Chapman and Hall, London. Fukunaga, K., 990. Statistical Pattern Recognition, second ed. Academic Press, New York. Goldenshluger, A., Nemirovsky, A., 997. On spatial adaptive estimation of nonparametric regression. Math. Meth. Stat. 6 (2), Hall, P., Sheather, S.J., Jones, M.C., Marron, J.S., 99. On optimal data-based bandwidth selection in kernel density estimation. Biometrika 78 (2), Katkovnik, V., 999. A new method for varying adaptive bandwidth selection. IEEE Trans. Signal Process. 47 (9), Katkovnik, V., Shmulevich, I., Kernel density estimation with varying data-driven bandwidth. in: EOS/SPIE Symposium, Image and Signal Processing for Remote Sensing, September 25 29, Barcelona, Spain. Loftsgaarden, D., Quesenberry, C., 965. A nonparametric estimate of a multivariate density function. Ann. Math. Stat. 36, Parzen, E., 962. On the estimation of a probability density function and the mode. Ann. Math. Stat. 33, Raudys, S., 99. On the effectiveness of Parzen window classifier. Informatica 2 (3), Sain, S.R., Scott, D.W., 996. On locally adaptive density estimation. J. Am. Stat. Assoc. 9 (436), Sheather, S.J., Jones, M.C., 99. A reliable data-based bandwidth selection method for kernel density estimation. J. R. Stat. Soc. B 53 (3), Silverman, B.W., 978. Choosing the window width when estimating a density. Biometrika 65,. Sindoukas, D., Laskaris, N., Fotopoulos, S., 997. Algorithms for color image edge enhancement using potential functions. IEEE Signal Process. Lett. 4 (9), Taylor, C.C., 989. Bootstrap choice of the smoothing parameter in kernel density estimation. Biometrika 76 (4), Terrell, G., Scott, D., 992. Variable kernel density estimation. Ann. Stat. 20 (3), Wright, D., Stander, J., Nicolaides, K., 997. Nonparametric density estimation and discrimination from images of shapes. J. Roy. Statistical Soc., Ser. C: Appl. Statistics 46 (3), Zabin, S., Wright, G., Nonparametric density estimation and detection in impulsive interference channels. IEEE Trans. Commun. 42 (2 4),

Least Squares Estimation

Least Squares Estimation Least Squares Estimation SARA A VAN DE GEER Volume 2, pp 1041 1045 in Encyclopedia of Statistics in Behavioral Science ISBN-13: 978-0-470-86080-9 ISBN-10: 0-470-86080-4 Editors Brian S Everitt & David

More information

ON THE EXISTENCE AND LIMIT BEHAVIOR OF THE OPTIMAL BANDWIDTH FOR KERNEL DENSITY ESTIMATION

ON THE EXISTENCE AND LIMIT BEHAVIOR OF THE OPTIMAL BANDWIDTH FOR KERNEL DENSITY ESTIMATION Statistica Sinica 17(27), 289-3 ON THE EXISTENCE AND LIMIT BEHAVIOR OF THE OPTIMAL BANDWIDTH FOR KERNEL DENSITY ESTIMATION J. E. Chacón, J. Montanero, A. G. Nogales and P. Pérez Universidad de Extremadura

More information

Parallel Computing of Kernel Density Estimates with MPI

Parallel Computing of Kernel Density Estimates with MPI Parallel Computing of Kernel Density Estimates with MPI Szymon Lukasik Department of Automatic Control, Cracow University of Technology, ul. Warszawska 24, 31-155 Cracow, Poland Szymon.Lukasik@pk.edu.pl

More information

Bandwidth Selection for Nonparametric Distribution Estimation

Bandwidth Selection for Nonparametric Distribution Estimation Bandwidth Selection for Nonparametric Distribution Estimation Bruce E. Hansen University of Wisconsin www.ssc.wisc.edu/~bhansen May 2004 Abstract The mean-square efficiency of cumulative distribution function

More information

Computational Statistics and Data Analysis

Computational Statistics and Data Analysis Computational Statistics and Data Analysis 53 (2008) 17 26 Contents lists available at ScienceDirect Computational Statistics and Data Analysis journal homepage: www.elsevier.com/locate/csda Coverage probability

More information

Syntax Menu Description Options Remarks and examples Stored results Methods and formulas References Also see. Description

Syntax Menu Description Options Remarks and examples Stored results Methods and formulas References Also see. Description Title stata.com lpoly Kernel-weighted local polynomial smoothing Syntax Menu Description Options Remarks and examples Stored results Methods and formulas References Also see Syntax lpoly yvar xvar [ if

More information

Example: Credit card default, we may be more interested in predicting the probabilty of a default than classifying individuals as default or not.

Example: Credit card default, we may be more interested in predicting the probabilty of a default than classifying individuals as default or not. Statistical Learning: Chapter 4 Classification 4.1 Introduction Supervised learning with a categorical (Qualitative) response Notation: - Feature vector X, - qualitative response Y, taking values in C

More information

A Complete Gradient Clustering Algorithm for Features Analysis of X-ray Images

A Complete Gradient Clustering Algorithm for Features Analysis of X-ray Images A Complete Gradient Clustering Algorithm for Features Analysis of X-ray Images Małgorzata Charytanowicz, Jerzy Niewczas, Piotr A. Kowalski, Piotr Kulczycki, Szymon Łukasik, and Sławomir Żak Abstract Methods

More information

Thresholding technique with adaptive window selection for uneven lighting image

Thresholding technique with adaptive window selection for uneven lighting image Pattern Recognition Letters 26 (2005) 801 808 wwwelseviercom/locate/patrec Thresholding technique with adaptive window selection for uneven lighting image Qingming Huang a, *, Wen Gao a, Wenjian Cai b

More information

Bootstrapping Big Data

Bootstrapping Big Data Bootstrapping Big Data Ariel Kleiner Ameet Talwalkar Purnamrita Sarkar Michael I. Jordan Computer Science Division University of California, Berkeley {akleiner, ameet, psarkar, jordan}@eecs.berkeley.edu

More information

A Learning Algorithm For Neural Network Ensembles

A Learning Algorithm For Neural Network Ensembles A Learning Algorithm For Neural Network Ensembles H. D. Navone, P. M. Granitto, P. F. Verdes and H. A. Ceccatto Instituto de Física Rosario (CONICET-UNR) Blvd. 27 de Febrero 210 Bis, 2000 Rosario. República

More information

Quantitative Methods for Finance

Quantitative Methods for Finance Quantitative Methods for Finance Module 1: The Time Value of Money 1 Learning how to interpret interest rates as required rates of return, discount rates, or opportunity costs. 2 Learning how to explain

More information

Using kernel methods to visualise crime data

Using kernel methods to visualise crime data Submission for the 2013 IAOS Prize for Young Statisticians Using kernel methods to visualise crime data Dr. Kieran Martin and Dr. Martin Ralphs kieran.martin@ons.gov.uk martin.ralphs@ons.gov.uk Office

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.cs.toronto.edu/~rsalakhu/ Lecture 6 Three Approaches to Classification Construct

More information

Department of Economics

Department of Economics Department of Economics On Testing for Diagonality of Large Dimensional Covariance Matrices George Kapetanios Working Paper No. 526 October 2004 ISSN 1473-0278 On Testing for Diagonality of Large Dimensional

More information

Applications of improved grey prediction model for power demand forecasting

Applications of improved grey prediction model for power demand forecasting Energy Conversion and Management 44 (2003) 2241 2249 www.elsevier.com/locate/enconman Applications of improved grey prediction model for power demand forecasting Che-Chiang Hsu a, *, Chia-Yon Chen b a

More information

The CUSUM algorithm a small review. Pierre Granjon

The CUSUM algorithm a small review. Pierre Granjon The CUSUM algorithm a small review Pierre Granjon June, 1 Contents 1 The CUSUM algorithm 1.1 Algorithm............................... 1.1.1 The problem......................... 1.1. The different steps......................

More information

From the help desk: Bootstrapped standard errors

From the help desk: Bootstrapped standard errors The Stata Journal (2003) 3, Number 1, pp. 71 80 From the help desk: Bootstrapped standard errors Weihua Guan Stata Corporation Abstract. Bootstrapping is a nonparametric approach for evaluating the distribution

More information

We discuss 2 resampling methods in this chapter - cross-validation - the bootstrap

We discuss 2 resampling methods in this chapter - cross-validation - the bootstrap Statistical Learning: Chapter 5 Resampling methods (Cross-validation and bootstrap) (Note: prior to these notes, we'll discuss a modification of an earlier train/test experiment from Ch 2) We discuss 2

More information

Local classification and local likelihoods

Local classification and local likelihoods Local classification and local likelihoods November 18 k-nearest neighbors The idea of local regression can be extended to classification as well The simplest way of doing so is called nearest neighbor

More information

Decompose Error Rate into components, some of which can be measured on unlabeled data

Decompose Error Rate into components, some of which can be measured on unlabeled data Bias-Variance Theory Decompose Error Rate into components, some of which can be measured on unlabeled data Bias-Variance Decomposition for Regression Bias-Variance Decomposition for Classification Bias-Variance

More information

DENSITY ESTIMATION FOR STATISTICS AND DATA ANALYSIS

DENSITY ESTIMATION FOR STATISTICS AND DATA ANALYSIS Published in Monographs on Statistics and Applied Probability, London: Chapman and Hall, 1986. DENSITY ESTIMATION FOR STATISTICS AND DATA ANALYSIS B.W. Silverman School of Mathematics University of Bath,

More information

Data analysis in supersaturated designs

Data analysis in supersaturated designs Statistics & Probability Letters 59 (2002) 35 44 Data analysis in supersaturated designs Runze Li a;b;, Dennis K.J. Lin a;b a Department of Statistics, The Pennsylvania State University, University Park,

More information

Basics of Statistical Machine Learning

Basics of Statistical Machine Learning CS761 Spring 2013 Advanced Machine Learning Basics of Statistical Machine Learning Lecturer: Xiaojin Zhu jerryzhu@cs.wisc.edu Modern machine learning is rooted in statistics. You will find many familiar

More information

Lecture 3: Linear methods for classification

Lecture 3: Linear methods for classification Lecture 3: Linear methods for classification Rafael A. Irizarry and Hector Corrada Bravo February, 2010 Today we describe four specific algorithms useful for classification problems: linear regression,

More information

New Approach to Analysis of the Switching Current Data in Ferroelectric Thin Films

New Approach to Analysis of the Switching Current Data in Ferroelectric Thin Films Ferroelectrics, 291: 27 35, 2003 Copyright c Taylor & Francis Inc. ISSN: 0015-0193 print / 1563-5112 online DOI: 10.1080/00150190390222510 New Approach to Analysis of the Switching Current Data in Ferroelectric

More information

Robert Collins CSE598G. More on Mean-shift. R.Collins, CSE, PSU CSE598G Spring 2006

Robert Collins CSE598G. More on Mean-shift. R.Collins, CSE, PSU CSE598G Spring 2006 More on Mean-shift R.Collins, CSE, PSU Spring 2006 Recall: Kernel Density Estimation Given a set of data samples x i ; i=1...n Convolve with a kernel function H to generate a smooth function f(x) Equivalent

More information

Statistics Graduate Courses

Statistics Graduate Courses Statistics Graduate Courses STAT 7002--Topics in Statistics-Biological/Physical/Mathematics (cr.arr.).organized study of selected topics. Subjects and earnable credit may vary from semester to semester.

More information

MATH4427 Notebook 2 Spring 2016. 2 MATH4427 Notebook 2 3. 2.1 Definitions and Examples... 3. 2.2 Performance Measures for Estimators...

MATH4427 Notebook 2 Spring 2016. 2 MATH4427 Notebook 2 3. 2.1 Definitions and Examples... 3. 2.2 Performance Measures for Estimators... MATH4427 Notebook 2 Spring 2016 prepared by Professor Jenny Baglivo c Copyright 2009-2016 by Jenny A. Baglivo. All Rights Reserved. Contents 2 MATH4427 Notebook 2 3 2.1 Definitions and Examples...................................

More information

Automated Learning and Data Visualization

Automated Learning and Data Visualization 1 Automated Learning and Data Visualization William S. Cleveland Department of Statistics Department of Computer Science Purdue University Methods of Statistics, Machine Learning, and Data Mining 2 Mathematical

More information

Optimal replenishment for a periodic review inventory system with two supply modes

Optimal replenishment for a periodic review inventory system with two supply modes European Journal of Operational Research 149 (2003) 229 244 Production, Manufacturing and Logistics Optimal replenishment for a periodic review inventory system with two supply modes Chi Chiang * www.elsevier.com/locate/dsw

More information

The Variability of P-Values. Summary

The Variability of P-Values. Summary The Variability of P-Values Dennis D. Boos Department of Statistics North Carolina State University Raleigh, NC 27695-8203 boos@stat.ncsu.edu August 15, 2009 NC State Statistics Departement Tech Report

More information

Appendix 1: Time series analysis of peak-rate years and synchrony testing.

Appendix 1: Time series analysis of peak-rate years and synchrony testing. Appendix 1: Time series analysis of peak-rate years and synchrony testing. Overview The raw data are accessible at Figshare ( Time series of global resources, DOI 10.6084/m9.figshare.929619), sources are

More information

A THEORETICAL COMPARISON OF DATA MASKING TECHNIQUES FOR NUMERICAL MICRODATA

A THEORETICAL COMPARISON OF DATA MASKING TECHNIQUES FOR NUMERICAL MICRODATA A THEORETICAL COMPARISON OF DATA MASKING TECHNIQUES FOR NUMERICAL MICRODATA Krish Muralidhar University of Kentucky Rathindra Sarathy Oklahoma State University Agency Internal User Unmasked Result Subjects

More information

A Comparative Study of the Pickup Method and its Variations Using a Simulated Hotel Reservation Data

A Comparative Study of the Pickup Method and its Variations Using a Simulated Hotel Reservation Data A Comparative Study of the Pickup Method and its Variations Using a Simulated Hotel Reservation Data Athanasius Zakhary, Neamat El Gayar Faculty of Computers and Information Cairo University, Giza, Egypt

More information

Forecasting methods applied to engineering management

Forecasting methods applied to engineering management Forecasting methods applied to engineering management Áron Szász-Gábor Abstract. This paper presents arguments for the usefulness of a simple forecasting application package for sustaining operational

More information

How To Understand The Theory Of Probability

How To Understand The Theory Of Probability Graduate Programs in Statistics Course Titles STAT 100 CALCULUS AND MATR IX ALGEBRA FOR STATISTICS. Differential and integral calculus; infinite series; matrix algebra STAT 195 INTRODUCTION TO MATHEMATICAL

More information

Regularized Logistic Regression for Mind Reading with Parallel Validation

Regularized Logistic Regression for Mind Reading with Parallel Validation Regularized Logistic Regression for Mind Reading with Parallel Validation Heikki Huttunen, Jukka-Pekka Kauppi, Jussi Tohka Tampere University of Technology Department of Signal Processing Tampere, Finland

More information

Decision-making with the AHP: Why is the principal eigenvector necessary

Decision-making with the AHP: Why is the principal eigenvector necessary European Journal of Operational Research 145 (2003) 85 91 Decision Aiding Decision-making with the AHP: Why is the principal eigenvector necessary Thomas L. Saaty * University of Pittsburgh, Pittsburgh,

More information

Optimization of technical trading strategies and the profitability in security markets

Optimization of technical trading strategies and the profitability in security markets Economics Letters 59 (1998) 249 254 Optimization of technical trading strategies and the profitability in security markets Ramazan Gençay 1, * University of Windsor, Department of Economics, 401 Sunset,

More information

Auxiliary Variables in Mixture Modeling: 3-Step Approaches Using Mplus

Auxiliary Variables in Mixture Modeling: 3-Step Approaches Using Mplus Auxiliary Variables in Mixture Modeling: 3-Step Approaches Using Mplus Tihomir Asparouhov and Bengt Muthén Mplus Web Notes: No. 15 Version 8, August 5, 2014 1 Abstract This paper discusses alternatives

More information

Multiple Linear Regression in Data Mining

Multiple Linear Regression in Data Mining Multiple Linear Regression in Data Mining Contents 2.1. A Review of Multiple Linear Regression 2.2. Illustration of the Regression Process 2.3. Subset Selection in Linear Regression 1 2 Chap. 2 Multiple

More information

Why Taking This Course? Course Introduction, Descriptive Statistics and Data Visualization. Learning Goals. GENOME 560, Spring 2012

Why Taking This Course? Course Introduction, Descriptive Statistics and Data Visualization. Learning Goals. GENOME 560, Spring 2012 Why Taking This Course? Course Introduction, Descriptive Statistics and Data Visualization GENOME 560, Spring 2012 Data are interesting because they help us understand the world Genomics: Massive Amounts

More information

Determining optimal window size for texture feature extraction methods

Determining optimal window size for texture feature extraction methods IX Spanish Symposium on Pattern Recognition and Image Analysis, Castellon, Spain, May 2001, vol.2, 237-242, ISBN: 84-8021-351-5. Determining optimal window size for texture feature extraction methods Domènec

More information

Diagnosis of multi-operational machining processes through variation propagation analysis

Diagnosis of multi-operational machining processes through variation propagation analysis Robotics and Computer Integrated Manufacturing 18 (2002) 233 239 Diagnosis of multi-operational machining processes through variation propagation analysis Qiang Huang, Shiyu Zhou, Jianjun Shi* Department

More information

Gerry Hobbs, Department of Statistics, West Virginia University

Gerry Hobbs, Department of Statistics, West Virginia University Decision Trees as a Predictive Modeling Method Gerry Hobbs, Department of Statistics, West Virginia University Abstract Predictive modeling has become an important area of interest in tasks such as credit

More information

Two Topics in Parametric Integration Applied to Stochastic Simulation in Industrial Engineering

Two Topics in Parametric Integration Applied to Stochastic Simulation in Industrial Engineering Two Topics in Parametric Integration Applied to Stochastic Simulation in Industrial Engineering Department of Industrial Engineering and Management Sciences Northwestern University September 15th, 2014

More information

Probability and Random Variables. Generation of random variables (r.v.)

Probability and Random Variables. Generation of random variables (r.v.) Probability and Random Variables Method for generating random variables with a specified probability distribution function. Gaussian And Markov Processes Characterization of Stationary Random Process Linearly

More information

SENSITIVITY ANALYSIS AND INFERENCE. Lecture 12

SENSITIVITY ANALYSIS AND INFERENCE. Lecture 12 This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike License. Your use of this material constitutes acceptance of that license and the conditions of use of materials on this

More information

MATH. ALGEBRA I HONORS 9 th Grade 12003200 ALGEBRA I HONORS

MATH. ALGEBRA I HONORS 9 th Grade 12003200 ALGEBRA I HONORS * Students who scored a Level 3 or above on the Florida Assessment Test Math Florida Standards (FSA-MAFS) are strongly encouraged to make Advanced Placement and/or dual enrollment courses their first choices

More information

Implementation and validation of an opportunistic stock market timing heuristic: One-day share volume spike as buy signal

Implementation and validation of an opportunistic stock market timing heuristic: One-day share volume spike as buy signal Available online at www.sciencedirect.com Expert Systems with Applications Expert Systems with Applications 35 (2008) 1628 1637 www.elsevier.com/locate/eswa Implementation and validation of an opportunistic

More information

Linear Threshold Units

Linear Threshold Units Linear Threshold Units w x hx (... w n x n w We assume that each feature x j and each weight w j is a real number (we will relax this later) We will study three different algorithms for learning linear

More information

Kernel Testing as an Alternative to χ 2 Analysis for Investigating the Distribution of Quantitative Traits

Kernel Testing as an Alternative to χ 2 Analysis for Investigating the Distribution of Quantitative Traits Kernel Testing as an Alternative to χ 2 Analysis for Investigating the Distribution of Quantitative Traits Jeffrey D. Hart, Anna Hale, J. Creighton Miller, Jr. Summary. Chi-square analysis is a popular

More information

Maximum Likelihood Estimation of ADC Parameters from Sine Wave Test Data. László Balogh, Balázs Fodor, Attila Sárhegyi, and István Kollár

Maximum Likelihood Estimation of ADC Parameters from Sine Wave Test Data. László Balogh, Balázs Fodor, Attila Sárhegyi, and István Kollár Maximum Lielihood Estimation of ADC Parameters from Sine Wave Test Data László Balogh, Balázs Fodor, Attila Sárhegyi, and István Kollár Dept. of Measurement and Information Systems Budapest University

More information

Statistics Review PSY379

Statistics Review PSY379 Statistics Review PSY379 Basic concepts Measurement scales Populations vs. samples Continuous vs. discrete variable Independent vs. dependent variable Descriptive vs. inferential stats Common analyses

More information

A logistic approximation to the cumulative normal distribution

A logistic approximation to the cumulative normal distribution A logistic approximation to the cumulative normal distribution Shannon R. Bowling 1 ; Mohammad T. Khasawneh 2 ; Sittichai Kaewkuekool 3 ; Byung Rae Cho 4 1 Old Dominion University (USA); 2 State University

More information

On Density Based Transforms for Uncertain Data Mining

On Density Based Transforms for Uncertain Data Mining On Density Based Transforms for Uncertain Data Mining Charu C. Aggarwal IBM T. J. Watson Research Center 19 Skyline Drive, Hawthorne, NY 10532 charu@us.ibm.com Abstract In spite of the great progress in

More information

Clustering Time Series Based on Forecast Distributions Using Kullback-Leibler Divergence

Clustering Time Series Based on Forecast Distributions Using Kullback-Leibler Divergence Clustering Time Series Based on Forecast Distributions Using Kullback-Leibler Divergence Taiyeong Lee, Yongqiao Xiao, Xiangxiang Meng, David Duling SAS Institute, Inc 100 SAS Campus Dr. Cary, NC 27513,

More information

Nonparametric adaptive age replacement with a one-cycle criterion

Nonparametric adaptive age replacement with a one-cycle criterion Nonparametric adaptive age replacement with a one-cycle criterion P. Coolen-Schrijner, F.P.A. Coolen Department of Mathematical Sciences University of Durham, Durham, DH1 3LE, UK e-mail: Pauline.Schrijner@durham.ac.uk

More information

Environmental Remote Sensing GEOG 2021

Environmental Remote Sensing GEOG 2021 Environmental Remote Sensing GEOG 2021 Lecture 4 Image classification 2 Purpose categorising data data abstraction / simplification data interpretation mapping for land cover mapping use land cover class

More information

TCOM 370 NOTES 99-4 BANDWIDTH, FREQUENCY RESPONSE, AND CAPACITY OF COMMUNICATION LINKS

TCOM 370 NOTES 99-4 BANDWIDTH, FREQUENCY RESPONSE, AND CAPACITY OF COMMUNICATION LINKS TCOM 370 NOTES 99-4 BANDWIDTH, FREQUENCY RESPONSE, AND CAPACITY OF COMMUNICATION LINKS 1. Bandwidth: The bandwidth of a communication link, or in general any system, was loosely defined as the width of

More information

SAS Software to Fit the Generalized Linear Model

SAS Software to Fit the Generalized Linear Model SAS Software to Fit the Generalized Linear Model Gordon Johnston, SAS Institute Inc., Cary, NC Abstract In recent years, the class of generalized linear models has gained popularity as a statistical modeling

More information

Penalized Logistic Regression and Classification of Microarray Data

Penalized Logistic Regression and Classification of Microarray Data Penalized Logistic Regression and Classification of Microarray Data Milan, May 2003 Anestis Antoniadis Laboratoire IMAG-LMC University Joseph Fourier Grenoble, France Penalized Logistic Regression andclassification

More information

Multiple Imputation for Missing Data: A Cautionary Tale

Multiple Imputation for Missing Data: A Cautionary Tale Multiple Imputation for Missing Data: A Cautionary Tale Paul D. Allison University of Pennsylvania Address correspondence to Paul D. Allison, Sociology Department, University of Pennsylvania, 3718 Locust

More information

Introduction to General and Generalized Linear Models

Introduction to General and Generalized Linear Models Introduction to General and Generalized Linear Models General Linear Models - part I Henrik Madsen Poul Thyregod Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby

More information

Why High-Order Polynomials Should Not be Used in Regression Discontinuity Designs

Why High-Order Polynomials Should Not be Used in Regression Discontinuity Designs Why High-Order Polynomials Should Not be Used in Regression Discontinuity Designs Andrew Gelman Guido Imbens 2 Aug 2014 Abstract It is common in regression discontinuity analysis to control for high order

More information

Regression III: Advanced Methods

Regression III: Advanced Methods Lecture 16: Generalized Additive Models Regression III: Advanced Methods Bill Jacoby Michigan State University http://polisci.msu.edu/jacoby/icpsr/regress3 Goals of the Lecture Introduce Additive Models

More information

Maximum likelihood estimation of mean reverting processes

Maximum likelihood estimation of mean reverting processes Maximum likelihood estimation of mean reverting processes José Carlos García Franco Onward, Inc. jcpollo@onwardinc.com Abstract Mean reverting processes are frequently used models in real options. For

More information

Introduction to nonparametric regression: Least squares vs. Nearest neighbors

Introduction to nonparametric regression: Least squares vs. Nearest neighbors Introduction to nonparametric regression: Least squares vs. Nearest neighbors Patrick Breheny October 30 Patrick Breheny STA 621: Nonparametric Statistics 1/16 Introduction For the remainder of the course,

More information

No-Arbitrage Condition of Option Implied Volatility and Bandwidth Selection

No-Arbitrage Condition of Option Implied Volatility and Bandwidth Selection Kamla-Raj 2014 Anthropologist, 17(3): 751-755 (2014) No-Arbitrage Condition of Option Implied Volatility and Bandwidth Selection Milos Kopa 1 and Tomas Tichy 2 1 Institute of Information Theory and Automation

More information

D-optimal plans in observational studies

D-optimal plans in observational studies D-optimal plans in observational studies Constanze Pumplün Stefan Rüping Katharina Morik Claus Weihs October 11, 2005 Abstract This paper investigates the use of Design of Experiments in observational

More information

Computing with Finite and Infinite Networks

Computing with Finite and Infinite Networks Computing with Finite and Infinite Networks Ole Winther Theoretical Physics, Lund University Sölvegatan 14 A, S-223 62 Lund, Sweden winther@nimis.thep.lu.se Abstract Using statistical mechanics results,

More information

Chapter 6: Point Estimation. Fall 2011. - Probability & Statistics

Chapter 6: Point Estimation. Fall 2011. - Probability & Statistics STAT355 Chapter 6: Point Estimation Fall 2011 Chapter Fall 2011 6: Point1 Estimat / 18 Chap 6 - Point Estimation 1 6.1 Some general Concepts of Point Estimation Point Estimate Unbiasedness Principle of

More information

T test as a parametric statistic

T test as a parametric statistic KJA Statistical Round pissn 2005-619 eissn 2005-7563 T test as a parametric statistic Korean Journal of Anesthesiology Department of Anesthesia and Pain Medicine, Pusan National University School of Medicine,

More information

Asymmetry and the Cost of Capital

Asymmetry and the Cost of Capital Asymmetry and the Cost of Capital Javier García Sánchez, IAE Business School Lorenzo Preve, IAE Business School Virginia Sarria Allende, IAE Business School Abstract The expected cost of capital is a crucial

More information

Learning Vector Quantization: generalization ability and dynamics of competing prototypes

Learning Vector Quantization: generalization ability and dynamics of competing prototypes Learning Vector Quantization: generalization ability and dynamics of competing prototypes Aree Witoelar 1, Michael Biehl 1, and Barbara Hammer 2 1 University of Groningen, Mathematics and Computing Science

More information

Statistics in Retail Finance. Chapter 7: Fraud Detection in Retail Credit

Statistics in Retail Finance. Chapter 7: Fraud Detection in Retail Credit Statistics in Retail Finance Chapter 7: Fraud Detection in Retail Credit 1 Overview > Detection of fraud remains an important issue in retail credit. Methods similar to scorecard development may be employed,

More information

Comparative study of the performance of the CuSum and EWMA control charts

Comparative study of the performance of the CuSum and EWMA control charts Computers & Industrial Engineering 46 (2004) 707 724 www.elsevier.com/locate/dsw Comparative study of the performance of the CuSum and EWMA control charts Vera do Carmo C. de Vargas*, Luis Felipe Dias

More information

Publication List. Chen Zehua Department of Statistics & Applied Probability National University of Singapore

Publication List. Chen Zehua Department of Statistics & Applied Probability National University of Singapore Publication List Chen Zehua Department of Statistics & Applied Probability National University of Singapore Publications Journal Papers 1. Y. He and Z. Chen (2014). A sequential procedure for feature selection

More information

Enhancing the SNR of the Fiber Optic Rotation Sensor using the LMS Algorithm

Enhancing the SNR of the Fiber Optic Rotation Sensor using the LMS Algorithm 1 Enhancing the SNR of the Fiber Optic Rotation Sensor using the LMS Algorithm Hani Mehrpouyan, Student Member, IEEE, Department of Electrical and Computer Engineering Queen s University, Kingston, Ontario,

More information

Time Series Analysis

Time Series Analysis JUNE 2012 Time Series Analysis CONTENT A time series is a chronological sequence of observations on a particular variable. Usually the observations are taken at regular intervals (days, months, years),

More information

Midwest Symposium On Circuits And Systems, 2004, v. 2, p. II137-II140. Creative Commons: Attribution 3.0 Hong Kong License

Midwest Symposium On Circuits And Systems, 2004, v. 2, p. II137-II140. Creative Commons: Attribution 3.0 Hong Kong License Title Adaptive window selection and smoothing of Lomb periodogram for time-frequency analysis of time series Author(s) Chan, SC; Zhang, Z Citation Midwest Symposium On Circuits And Systems, 2004, v. 2,

More information

A Study on the Comparison of Electricity Forecasting Models: Korea and China

A Study on the Comparison of Electricity Forecasting Models: Korea and China Communications for Statistical Applications and Methods 2015, Vol. 22, No. 6, 675 683 DOI: http://dx.doi.org/10.5351/csam.2015.22.6.675 Print ISSN 2287-7843 / Online ISSN 2383-4757 A Study on the Comparison

More information

Moving Least Squares Approximation

Moving Least Squares Approximation Chapter 7 Moving Least Squares Approimation An alternative to radial basis function interpolation and approimation is the so-called moving least squares method. As we will see below, in this method the

More information

A Review of Cross Sectional Regression for Financial Data You should already know this material from previous study

A Review of Cross Sectional Regression for Financial Data You should already know this material from previous study A Review of Cross Sectional Regression for Financial Data You should already know this material from previous study But I will offer a review, with a focus on issues which arise in finance 1 TYPES OF FINANCIAL

More information

2 Sample t-test (unequal sample sizes and unequal variances)

2 Sample t-test (unequal sample sizes and unequal variances) Variations of the t-test: Sample tail Sample t-test (unequal sample sizes and unequal variances) Like the last example, below we have ceramic sherd thickness measurements (in cm) of two samples representing

More information

Fuzzy regression model with fuzzy input and output data for manpower forecasting

Fuzzy regression model with fuzzy input and output data for manpower forecasting Fuzzy Sets and Systems 9 (200) 205 23 www.elsevier.com/locate/fss Fuzzy regression model with fuzzy input and output data for manpower forecasting Hong Tau Lee, Sheu Hua Chen Department of Industrial Engineering

More information

**BEGINNING OF EXAMINATION** The annual number of claims for an insured has probability function: , 0 < q < 1.

**BEGINNING OF EXAMINATION** The annual number of claims for an insured has probability function: , 0 < q < 1. **BEGINNING OF EXAMINATION** 1. You are given: (i) The annual number of claims for an insured has probability function: 3 p x q q x x ( ) = ( 1 ) 3 x, x = 0,1,, 3 (ii) The prior density is π ( q) = q,

More information

Jitter Measurements in Serial Data Signals

Jitter Measurements in Serial Data Signals Jitter Measurements in Serial Data Signals Michael Schnecker, Product Manager LeCroy Corporation Introduction The increasing speed of serial data transmission systems places greater importance on measuring

More information

Smoothing and Non-Parametric Regression

Smoothing and Non-Parametric Regression Smoothing and Non-Parametric Regression Germán Rodríguez grodri@princeton.edu Spring, 2001 Objective: to estimate the effects of covariates X on a response y nonparametrically, letting the data suggest

More information

Data Mining and Data Warehousing. Henryk Maciejewski. Data Mining Predictive modelling: regression

Data Mining and Data Warehousing. Henryk Maciejewski. Data Mining Predictive modelling: regression Data Mining and Data Warehousing Henryk Maciejewski Data Mining Predictive modelling: regression Algorithms for Predictive Modelling Contents Regression Classification Auxiliary topics: Estimation of prediction

More information

Linear Models and Conjoint Analysis with Nonlinear Spline Transformations

Linear Models and Conjoint Analysis with Nonlinear Spline Transformations Linear Models and Conjoint Analysis with Nonlinear Spline Transformations Warren F. Kuhfeld Mark Garratt Abstract Many common data analysis models are based on the general linear univariate model, including

More information

THE NUMBER OF GRAPHS AND A RANDOM GRAPH WITH A GIVEN DEGREE SEQUENCE. Alexander Barvinok

THE NUMBER OF GRAPHS AND A RANDOM GRAPH WITH A GIVEN DEGREE SEQUENCE. Alexander Barvinok THE NUMBER OF GRAPHS AND A RANDOM GRAPH WITH A GIVEN DEGREE SEQUENCE Alexer Barvinok Papers are available at http://www.math.lsa.umich.edu/ barvinok/papers.html This is a joint work with J.A. Hartigan

More information

The Kelly criterion for spread bets

The Kelly criterion for spread bets IMA Journal of Applied Mathematics 2007 72,43 51 doi:10.1093/imamat/hxl027 Advance Access publication on December 5, 2006 The Kelly criterion for spread bets S. J. CHAPMAN Oxford Centre for Industrial

More information

STATISTICA Formula Guide: Logistic Regression. Table of Contents

STATISTICA Formula Guide: Logistic Regression. Table of Contents : Table of Contents... 1 Overview of Model... 1 Dispersion... 2 Parameterization... 3 Sigma-Restricted Model... 3 Overparameterized Model... 4 Reference Coding... 4 Model Summary (Summary Tab)... 5 Summary

More information

Descriptive Statistics

Descriptive Statistics Descriptive Statistics Primer Descriptive statistics Central tendency Variation Relative position Relationships Calculating descriptive statistics Descriptive Statistics Purpose to describe or summarize

More information

IN current film media, the increase in areal density has

IN current film media, the increase in areal density has IEEE TRANSACTIONS ON MAGNETICS, VOL. 44, NO. 1, JANUARY 2008 193 A New Read Channel Model for Patterned Media Storage Seyhan Karakulak, Paul H. Siegel, Fellow, IEEE, Jack K. Wolf, Life Fellow, IEEE, and

More information

The Assumption(s) of Normality

The Assumption(s) of Normality The Assumption(s) of Normality Copyright 2000, 2011, J. Toby Mordkoff This is very complicated, so I ll provide two versions. At a minimum, you should know the short one. It would be great if you knew

More information

Application of discriminant analysis to predict the class of degree for graduating students in a university system

Application of discriminant analysis to predict the class of degree for graduating students in a university system International Journal of Physical Sciences Vol. 4 (), pp. 06-0, January, 009 Available online at http://www.academicjournals.org/ijps ISSN 99-950 009 Academic Journals Full Length Research Paper Application

More information

Two-Sample T-Tests Assuming Equal Variance (Enter Means)

Two-Sample T-Tests Assuming Equal Variance (Enter Means) Chapter 4 Two-Sample T-Tests Assuming Equal Variance (Enter Means) Introduction This procedure provides sample size and power calculations for one- or two-sided two-sample t-tests when the variances of

More information