Signal detection and goodness-of-fit: the Berk-Jones statistics revisited
|
|
- Alexis Cain
- 8 years ago
- Views:
Transcription
1 Signal detection and goodness-of-fit: the Berk-Jones statistics revisited Jon A. Wellner (Seattle) INET Big Data Conference
2 INET Big Data Conference, Cambridge September 29-30, 2015 Based on joint work with: Lutz Dümbgen (Bern) Leah Jager (U.S. Naval Academy)
3 Outline 1: 2: 3: 4: 5: 6: Introduction: some history Testing for sparse normal means: optimal detection boundary Consequences and tradeoffs. The LIL and strong LIL for Brownian motion and Brownian bridge A new class of test statistics: LIL adjusted higher criticism & Berk-Jones. Power properties and confidence bands. INET Big Data Conference, Cambridge, September 29-30,
4 1. Introduction: some history Setting: classical goodness - of - fit X 1,..., X n i.i.d. with distribution function F F n (x) = 1 n ni=1 1 [Xi x] Test H : F = F 0 versus K : F F 0, F 0 continuous Without loss of generality F 0 (x) = x, the U(0, 1) distribution Break hypotheses down into family of pointwise hypotheses: H x : F (x) = F 0 (x) versus K x : F (x) F 0 (x) H = x H x, K = x K x INET Big Data Conference, Cambridge, September 29-30,
5 1. Introduction: some history nf n (x) Binomial(n, F (x)). Likelihood ratio statistic for testing H x versus K x : λ n (x) = sup F (x) L n(f (x)) = L n(f n (x)) L n (F 0 (x)) L n (F 0 (x)) = F n(x) nfn(x) (1 F n (x)) n(1 F n(x)) F 0 (x) nfn(x) (1 F 0 (x)) n(1 F n(x)) ( ) nfn (x) Fn (x) ( 1 F n (x) = F 0 (x) 1 F 0 (x) ) n(1 Fn (x)) INET Big Data Conference, Cambridge, September 29-30,
6 1. Introduction: some history Thus logλ n (x) = nf n (x)log ( Fn (x) F 0 (x) ) + n(1 F n (x))log = nk(f n (x), F 0 (x)) K(u, v) ulog ( u v ) + (1 u)log ( 1 u 1 v Kullback - Leibler distance Bernoulli(u), Bernoulli(v) Berk-Jones (1979) test statistic: via S.N. Roy s union intersection principle, R n sup x ), ( 1 Fn (x) 1 F 0 (x) n 1 logλ n (x) = sup x K(F n (x), F 0 (x)). ) INET Big Data Conference, Cambridge, September 29-30,
7 1. Introduction: some history History: Berk and Jones (1979) Groeneboom and Shorack (1981) Shorack and Wellner (1986, p. 786) Owen (1995): inversion of R n to get confidence bands; finite - sample distribution via Noé s recursion Einmahl and McKeague (2002): integral version of R n Donoho and Jin (2004): Tukey s Higher-Criticism statistic for testing sparse normal means model with comparisons to Berk - Jones statistic R n INET Big Data Conference, Cambridge, September 29-30,
8 2. Testing for sparse normal means Initial setting: multiple testing of normal means For i = 1,..., n consider testing versus H 0,i : X i N(0, 1) H 1,i : X i N(µ i, 1) with µ i > 0. Sparsity: proportion ɛ n n 1 #{i n : µ i > 0} is small; ɛ n n β with 0 < β < 1. Three questions (in increasing order of difficulty): Q1: Can we tell if at least one null hypothesis is false? Q2: What is the proportion of false null hypotheses? Q3: Which null hypotheses are false? Main focus here: Q1. INET Big Data Conference, Cambridge, September 29-30,
9 2. Testing for sparse normal means Previous work: Q1: is there any signal? Ingster (1997, 1999) Jin (2004) Donoho and Jin (2004) Jager and Wellner (2007) Hall and Jin (2007) Cai and Wu (2014) INET Big Data Conference, Cambridge, September 29-30,
10 2. Testing for sparse normal means Change of setting: Ingster - Donoho - Jin testing problem Suppose Y 1,..., Y n i.i.d. G on R test H : G = N(0, 1) versus H 1 : G = (1 ɛ)n(0, 1) + ɛn(µ, 1), and, in particular, against H (n) 1 : G = (1 ɛ n )N(0, 1) + ɛ n N(µ n, 1). for ɛ n = n β, µ n = 2rlogn 0 < β < 1, 0 < r < 1. Let Φ(z) P (Z z) = z (2π) 1/2 exp( x 2 /2)dx, Z N(0, 1). transform to X i 1 Φ(Y i ) [0, 1] i.i.d. F = 1 G(Φ 1 (1 )). INET Big Data Conference, Cambridge, September 29-30,
11 2. Testing for sparse normal means Then the testing problem becomes: test H 0 : F = F 0 = U(0, 1) versus H (n) 1 : F (u) = u + ɛ n {(1 u) Φ(Φ 1 (1 u) µ n )} = (1 ɛ n )u + ɛ n {1 Φ(Φ 1 (1 u) µ n )} Test statistics: Donoho and Jin (2004) proposed HCn n(fn (u) u) sup X (1) u X ([n/2]) u(1 u) Tukey s higher criticism statistic where F n (u) n 1 n i=1 1 [0,u] (X i ) = empirical distribution function of the X i s. Let K 2 (u, v) = 2 1 (u v) 2 /(v(1 v)) 2 ; then 2 1 (HC n) 2 = sup X (1) x X [n/2] K + 2 (F n(x), x) where K + 2 (u, v) = K 2(u, v)1 [v u]. INET Big Data Conference, Cambridge, September 29-30,
12 2. Testing for sparse normal means Optimal detection boundary ρ (β) defined by: ρ (β) = { β 1/2, 1/2 < β 3/4 (1 1 β) 2, 3/4 < β < 1. Theorem 1: (Donoho - Jin, 2004). For r > ρ (β) the tests T n based on 2 1 (HCn) 2 or R n + are both size and power consistent for testing H 0 versus H (n) 1. With t n (α n ) = loglog(n)(1 + o(1)) P H0 (T n > t n (α n )) = α n 0, and P (n)(t n > t n (α n )) 1, as n. H 1 INET Big Data Conference, Cambridge, September 29-30,
13 2. Testing for sparse normal means 1.0 r detectable 0.4 ρ * (β) 0.2 undetectable β INET Big Data Conference, Cambridge, September 29-30,
14 2. Testing for sparse normal means A family of statistics via phi-divergences For s R, x 0 define Then define φ s (x) = 1 s+sx x s s(1 s), s 0, 1 xlog(x) x + 1, s = 1 x log(x) 1, s = 0. K s (u, v) = vφ s (u/v) + (1 v)φ s ((1 u)/(1 v)). INET Big Data Conference, Cambridge, September 29-30,
15 2. Testing for sparse normal means Special cases: K 2 (u, v) = 1 (u v) 2 2v(1 v) K 1 (u, v) = K(u, v) = ulog(u/v) + (1 u)log((1 u)/(1 v)) K 1/2 (u, v) = 2{( u v) 2 + ( 1 u 1 v) 2 } = 4{1 uv (1 u)(1 v)}. K 0 (u, v) = K(v, u) K 1 (u, v) = K 2 (v, u) = 1 (u v) 2 2u(1 u) INET Big Data Conference, Cambridge, September 29-30,
16 2. Testing for sparse normal means The φ divergence family of statistics (Jager & W, 2007): S n (s) = { supx R K s (F n (x), F 0 (x)), s 1 sup x [X(1),X (n) ) K s(f n (x), F 0 (x)), s < 1, Thus, with F 0 (x) = x, S n (2) = 1 2 sup (F n (x) x) 2, x R x(1 x) S n (1) = R n = Berk-Jones statistic S n (1/2) = 4 sup {1 x [X (1),X (n) ) F n (x)x S n (0) = reversed Berk-Jones R n S n ( 1) = 1 (F n (x) x) 2 sup 2 x [X (1),X (n) ) F n (x)(1 F n (x)) (1 F n (x))(1 x)} INET Big Data Conference, Cambridge, September 29-30,
17 2. Testing for sparse normal means Null hypothesis distribution theory: Owen (1995) and Jager (2006): finite sample critical points via Noé s recursion for n 3000 For n 3000, asymptotic theory via Jaeschke (1979) and Eicker (1979) (cf. SW p ), together with so where K s (u, v) 2 1 (u v) 2 /[v(1 v)] nk s (F n (x), x) 1 2 Z n (x) n(fn (x) x) x(1 x) n(f n (x) x) 2 x(1 x) f.d. 1 2 Z n(x) 2 U(x) Z(x) x(1 x) with U a standard Brownian bridge process on [0, 1]. INET Big Data Conference, Cambridge, September 29-30,
18 2. Testing for sparse normal means Let r n loglog(n) + (1/2)logloglog(n) (1/2)log(4π) = loglog(n)(1 + o(1)). Theorem 1. If F = F 0, the uniform distribution on [0, 1], then for 1 s 2 ns n (s) r n d Y 4 where P (Y 4 x) = exp( 4exp( x)). Theorem 2. If F = F n, the Ingster - Donoho - Jin sparse normal means model, then for each s [ 1, 2] P H0 (ns n (s) > t n (α n )) 0, and P (n)(ns n (s) > t n (α n )) 1 if r > ρ (β). H 1 INET Big Data Conference, Cambridge, September 29-30,
19 3. Consequences and tradeoffs: trouble in the middle! Although the test statistics ns n (s) (and their one-sided, onetailed counterparts) have excellent power behavior against sparse normal means and other tail alternatives, we have lost something in the middle: if {F n } is a sequence of distribution functions satisfying 1/2 n(f n f 1/2 0 ) 2 1 af 1/2 0, in L 2 (λ) n(fn F 0 ) A uniformly where A(x) = x a(y)df 0 (y), then it is not hard to see that for any κ > 0 P Fn (ns n (s) r n > κ) 0. Can we find a new family of test statistics which have good power properties for alternatives of both the tail and central type? INET Big Data Conference, Cambridge, September 29-30,
20 4. The LIL and Strong LIL for Brownian Motion and Bridge Standard Brownian motion W = (W(t)) t 0 LIL for BM: lim sup t 0 ±W(t) 2tloglog(t 1 ) = 1 almost surely, lim sup t ±W(t) 2tloglog(t) = 1 almost surely. Refined (upper) strong LIL for BM: For arbitrary constants ν > 3/2, ( W(t) 2 lim sup t {0, } 2t almost surely. ) loglog(t + t 1 ) ν logloglog(t + t 1 ) < 0 INET Big Data Conference, Cambridge, September 29-30,
21 4. The LIL and Strong LIL for Brownian Motion and Bridge Reformulation for standard Brownian bridge U = (U(t)) t (0,1) ( ) t (0, 1) t logit(t) := log R, 1 t R x l(x) := ex 1 + e x (0, 1). C(t) := log 1 + logit(t) 2 /2 loglog(1/t) as t 0, D(t) := log 1 + C(t) 2 /2 logloglog(1/t) as t 0. almost surely. ( U(t) 2 lim sup t {0,1} 2t(1 t) ) C(t) ν D(t) < 0 INET Big Data Conference, Cambridge, September 29-30,
22 5: A new class of test statistics: LIL adjusted higher criticism & Berk-Jones. This suggests a new class of test statistics as follows: Fix s [1, 2] and ν > 3/2. Define X n,s (t) nk s (G n (t), t). Set T n (s, ν) sup 0<t<1 {X n,s (t) C(t) νd(t)}. If s [ 1, 1) replace the supremum over (0, 1) by the sup over [X (1), X (n) ). Theorem. For all s [ 1, 2] and ν > 3/2, if H 0 holds then { U 2 } (t) T n (s, ν) d T ν C(t) νd(t) 2t(1 t) sup 0<t<1 where T ν is finite almost surely. Proof: Careful use of strong approximation methods: Csörgő, Csörgő, Horvath, and Mason (1986). INET Big Data Conference, Cambridge, September 29-30,
23 In particular for the LIL adjusted higher criticism & Berk-Jones statistics: { } U 2 T n (2, ν) sup n (t) C(t) νd(t) d T ν, 0<t<1 2t(1 t) T n (1, ν) {nk 1 (G n (t), t) C(t) νd(t)} d T ν. sup 0<t<1 where K 1 (u, v) K(u, v) = ulog(u/v)+(1 u)log((1 u)/(1 v)). What about power? Tail alternatives, e.g. sparse normal means? Central (or contiguous) alternatives? Widths of confidence bands? INET Big Data Conference, Cambridge, September 29-30,
24 6. Confidence bands and power properties Let U 1, U 2,..., U n be i.i.d. Unif[0, 1]. Auxiliary function K : [0, 1] (0, 1) [0, ], K(x, p) := xlog ( ) x p ( 1 x + (1 x)log 1 p i.e. Kullback-Leibler divergence between Bin(1, x) and Bin(1, p). Two key properties: (x p)2 ( ) K(x, p) = 1 + o(1) as x p. 2p(1 p) K(x, p) c implies x p ) 2c p(1 p) + c 2c x(1 x) + c INET Big Data Conference, Cambridge, September 29-30,
25 6. Confidence bands and power properties Uniform order statistics 0 < U n:1 < U n:2 < < U n:n < 1. T n := {t n1, t n2,..., t nn } with t ni := IE(U n:i ) = i n + 1. Theorem 2. For the process X n = ( X n (t)) t Tn with for ν > 3/2 we have X n (t ni ) (n + 1)K(t ni, U n:i ) T n (1, ν) sup T n { X n C νd } d T ν. Some realizations of X n C νd for n = 5000 and ν = 3: INET Big Data Conference, Cambridge, September 29-30,
26 X n (t) INET Big Data Conference, Cambridge, September 29-30, t
27 X n (t) INET Big Data Conference, Cambridge, September 29-30, t
28 X n (t) INET Big Data Conference, Cambridge, September 29-30, t
29 X n (t) INET Big Data Conference, Cambridge, September 29-30, t
30 Distribution function of arg max t X n (t): t INET Big Data Conference, Cambridge, September 29-30,
31 Let X 1, X 2,..., X n be i.i.d. with unknown c.d.f. F on R. Empirical df: F n (x) := n 1 ni=1 1 [Xi x]. Testing problem: H 0 : F F 0 versus K : F F 0. Berk-Jones statistic: R n (F 0 ) := sup R n K(F n, F 0 ). Critical value: κ BJ n,α (1 α) quantile of sup t (0,1) = loglog(n) + O ( logloglog(n) ). n K n (G n (t), t) New proposal: T n (F 0 ) sup R ( n K(Fn, F 0 ) C(F 0 ) νd(f 0 ) ). INET Big Data Conference, Cambridge, September 29-30,
32 ... with critical value κ new n,α (1 α) quantile of ( n K(Gn (t), t) C(t) νd(t) ) sup t (0,1) (1 α) quantile of ( U(t) 2 sup t (0,1) 2t(1 t) ) C(t) νd(t). INET Big Data Conference, Cambridge, September 29-30,
33 Lemma. such that For any critical value κ > 0 there exists a constant B κ IP F ( Tn (F o ) κ ) B κ n (F, F o ) 4/5 where n (F, F o ) := sup R and Γ( ) := C( ) + 1. n F Fo Γ(F o ) F o (1 F o ) + Γ(F o )/ n Note: Γ(t) t(1 t) 0 as t {0, 1}. INET Big Data Conference, Cambridge, September 29-30,
34 Special case: Detecting heterogeneous Gaussian mixtures (Donoho Jin 2004) F o := Φ, F n := (1 ε n ) Φ + ε n Φ( µ n ), ε n (0, 1), µ n > 0. Setting 1 (Donoho Jin, 2004): ε n = n β+o(1) for some β (1/2, 1). Setting 2: ε n = n 1/2+o(1) but π n := n 1/2 ε n 0. INET Big Data Conference, Cambridge, September 29-30,
35 Theorem A. For any fixed κ > 0, IP Fn ( Tn (F o ) > κ ) 1 provided that µ n satisfies the following conditions: Setting 1 (ε n = n β+o(1), β (1/2, 1)): µ n = 2rlog(n) with r > β 1/2 if β 3/4, ( ) β if β 3/4. Setting 2 (ε n = n 1/2+o(1), π n = n 1/2 ε n 0): µ n = 2slog(1/π n ) with s > 1. INET Big Data Conference, Cambridge, September 29-30,
36 Setting 2 (contiguous alternatives): For fixed π, µ > 0, ε n = πn 1/2 and µ n = µ. Optimal test of F versus F n has asymptotic power ( Φ Theorem B. As π 0 and µ = ( Φ Φ 1 (α) + π2 (exp(µ 2 ) 1) 4 Φ 1 (α) + π2 (exp(µ 2 ) 1) 4 ). 2slog(1/π) for fixed s > 0, ) α if s < 1, 1 if s > 1, lim sup n IP F n ( Tn (F 0 ) > κ n,α ) 1 if s > 1. INET Big Data Conference, Cambridge, September 29-30,
37 Confidence Bands Owen (1995) proposed (1 α)-confidence band { } F : sup n K(F n, F ) κ BJ n,α. R New proposal: With order statistics X n:1 X n:2 X n:n, { ( F : max (n + 1)K(tni, F (X n:i )) C(t ni ) νd(t ni ) ) } κ n,α 1 i n Resulting bounds for F (x): with confidence 1 α, on [X n:i, X n:i+1 ), 0 i n, F [a BJO ni [a new ni, b BJO ni ] with Owen s (1995) proposal,, b new ni ] with new proposal, while F n (X n:i ) = INET Big Data Conference, Cambridge, September 29-30,
38 n = 500: i a new ni, s ni, b new ni INET Big Data Conference, Cambridge, September 29-30,
39 n = 2000: i a ni s ni, b ni s ni INET Big Data Conference, Cambridge, September 29-30,
40 n = 8000: i a ni s ni, b ni s ni INET Big Data Conference, Cambridge, September 29-30,
41 Theorem C. For any fixed α (0, 1), while max 0 i n (bbjo ni max 0 i n (bnew ni max 0 i n b new ni b BJO ni a new ni a BJO ni 1, a BJO ni ) = (1 + o(1)) a new ni ) = O(n 1/2 ). 2loglogn, n INET Big Data Conference, Cambridge, September 29-30,
42 Final comments, extensions One can replace K(u, v) = K 1 (u, v) with more general φdivergences K s (u, v) as in Jager Wellner (2007) under the null hypothesis. Power behavior of the family T n (s, ν) for s / {1, 2} is still unknown. Numerical experiments of Walther (2013) and Siegmund and Li (2015) indicate that K = K 1 has the best small/moderate sample performance in the sparse normal means model of Donoho Jin (2004). Results for more general mixture models: Cai and Wu (2014). INET Big Data Conference, Cambridge, September 29-30,
43 Proof of Jaeschke - Eicker theorem for sup 0<t<1 U n (t) t(1 t) : d n = (logn)5 n < 1/2 if n > Number theory: Littlewood showed that Li(x) π(x) changes sign infinitely often for x large. Skewes (1933): first sign change of Li(x) π(x) before if the Riemann hypothesis holds Current estimate: e first sign change of Li(x) π(x) before INET Big Data Conference, Cambridge, September 29-30,
Some stability results of parameter identification in a jump diffusion model
Some stability results of parameter identification in a jump diffusion model D. Düvelmeyer Technische Universität Chemnitz, Fakultät für Mathematik, 09107 Chemnitz, Germany Abstract In this paper we discuss
More informationLecture 13: Martingales
Lecture 13: Martingales 1. Definition of a Martingale 1.1 Filtrations 1.2 Definition of a martingale and its basic properties 1.3 Sums of independent random variables and related models 1.4 Products of
More informationBANACH AND HILBERT SPACE REVIEW
BANACH AND HILBET SPACE EVIEW CHISTOPHE HEIL These notes will briefly review some basic concepts related to the theory of Banach and Hilbert spaces. We are not trying to give a complete development, but
More informationMATH4427 Notebook 2 Spring 2016. 2 MATH4427 Notebook 2 3. 2.1 Definitions and Examples... 3. 2.2 Performance Measures for Estimators...
MATH4427 Notebook 2 Spring 2016 prepared by Professor Jenny Baglivo c Copyright 2009-2016 by Jenny A. Baglivo. All Rights Reserved. Contents 2 MATH4427 Notebook 2 3 2.1 Definitions and Examples...................................
More informationA Uniform Asymptotic Estimate for Discounted Aggregate Claims with Subexponential Tails
12th International Congress on Insurance: Mathematics and Economics July 16-18, 2008 A Uniform Asymptotic Estimate for Discounted Aggregate Claims with Subexponential Tails XUEMIAO HAO (Based on a joint
More information. (3.3) n Note that supremum (3.2) must occur at one of the observed values x i or to the left of x i.
Chapter 3 Kolmogorov-Smirnov Tests There are many situations where experimenters need to know what is the distribution of the population of their interest. For example, if they want to use a parametric
More informationDongfeng Li. Autumn 2010
Autumn 2010 Chapter Contents Some statistics background; ; Comparing means and proportions; variance. Students should master the basic concepts, descriptive statistics measures and graphs, basic hypothesis
More informationINDISTINGUISHABILITY OF ABSOLUTELY CONTINUOUS AND SINGULAR DISTRIBUTIONS
INDISTINGUISHABILITY OF ABSOLUTELY CONTINUOUS AND SINGULAR DISTRIBUTIONS STEVEN P. LALLEY AND ANDREW NOBEL Abstract. It is shown that there are no consistent decision rules for the hypothesis testing problem
More informationCHAPTER IV - BROWNIAN MOTION
CHAPTER IV - BROWNIAN MOTION JOSEPH G. CONLON 1. Construction of Brownian Motion There are two ways in which the idea of a Markov chain on a discrete state space can be generalized: (1) The discrete time
More informationEstimating the Degree of Activity of jumps in High Frequency Financial Data. joint with Yacine Aït-Sahalia
Estimating the Degree of Activity of jumps in High Frequency Financial Data joint with Yacine Aït-Sahalia Aim and setting An underlying process X = (X t ) t 0, observed at equally spaced discrete times
More informationPacific Journal of Mathematics
Pacific Journal of Mathematics GLOBAL EXISTENCE AND DECREASING PROPERTY OF BOUNDARY VALUES OF SOLUTIONS TO PARABOLIC EQUATIONS WITH NONLOCAL BOUNDARY CONDITIONS Sangwon Seo Volume 193 No. 1 March 2000
More informationSTAT 830 Convergence in Distribution
STAT 830 Convergence in Distribution Richard Lockhart Simon Fraser University STAT 830 Fall 2011 Richard Lockhart (Simon Fraser University) STAT 830 Convergence in Distribution STAT 830 Fall 2011 1 / 31
More informationDetermining distribution parameters from quantiles
Determining distribution parameters from quantiles John D. Cook Department of Biostatistics The University of Texas M. D. Anderson Cancer Center P. O. Box 301402 Unit 1409 Houston, TX 77230-1402 USA cook@mderson.org
More informationSummary of Formulas and Concepts. Descriptive Statistics (Ch. 1-4)
Summary of Formulas and Concepts Descriptive Statistics (Ch. 1-4) Definitions Population: The complete set of numerical information on a particular quantity in which an investigator is interested. We assume
More informationChapter 4: Statistical Hypothesis Testing
Chapter 4: Statistical Hypothesis Testing Christophe Hurlin November 20, 2015 Christophe Hurlin () Advanced Econometrics - Master ESA November 20, 2015 1 / 225 Section 1 Introduction Christophe Hurlin
More informationMetric Spaces. Chapter 1
Chapter 1 Metric Spaces Many of the arguments you have seen in several variable calculus are almost identical to the corresponding arguments in one variable calculus, especially arguments concerning convergence
More informationInfinitely Imbalanced Logistic Regression
Journal of Machine Learning Research () 1 13 Submitted 9/06;Revised 12/06; Published Infinitely Imbalanced Logistic Regression Art B. Owen Department of Statistics Stanford Unversity Stanford CA, 94305,
More informationChapter 2: Binomial Methods and the Black-Scholes Formula
Chapter 2: Binomial Methods and the Black-Scholes Formula 2.1 Binomial Trees We consider a financial market consisting of a bond B t = B(t), a stock S t = S(t), and a call-option C t = C(t), where the
More informationPrinciple of Data Reduction
Chapter 6 Principle of Data Reduction 6.1 Introduction An experimenter uses the information in a sample X 1,..., X n to make inferences about an unknown parameter θ. If the sample size n is large, then
More informationINSURANCE RISK THEORY (Problems)
INSURANCE RISK THEORY (Problems) 1 Counting random variables 1. (Lack of memory property) Let X be a geometric distributed random variable with parameter p (, 1), (X Ge (p)). Show that for all n, m =,
More informationExact Confidence Intervals
Math 541: Statistical Theory II Instructor: Songfeng Zheng Exact Confidence Intervals Confidence intervals provide an alternative to using an estimator ˆθ when we wish to estimate an unknown parameter
More informationIs a Brownian motion skew?
Is a Brownian motion skew? Ernesto Mordecki Sesión en honor a Mario Wschebor Universidad de la República, Montevideo, Uruguay XI CLAPEM - November 2009 - Venezuela 1 1 Joint work with Antoine Lejay and
More informationCHAPTER II THE LIMIT OF A SEQUENCE OF NUMBERS DEFINITION OF THE NUMBER e.
CHAPTER II THE LIMIT OF A SEQUENCE OF NUMBERS DEFINITION OF THE NUMBER e. This chapter contains the beginnings of the most important, and probably the most subtle, notion in mathematical analysis, i.e.,
More informationAsymptotics of discounted aggregate claims for renewal risk model with risky investment
Appl. Math. J. Chinese Univ. 21, 25(2: 29-216 Asymptotics of discounted aggregate claims for renewal risk model with risky investment JIANG Tao Abstract. Under the assumption that the claim size is subexponentially
More informationFor a partition B 1,..., B n, where B i B j = for i. A = (A B 1 ) (A B 2 ),..., (A B n ) and thus. P (A) = P (A B i ) = P (A B i )P (B i )
Probability Review 15.075 Cynthia Rudin A probability space, defined by Kolmogorov (1903-1987) consists of: A set of outcomes S, e.g., for the roll of a die, S = {1, 2, 3, 4, 5, 6}, 1 1 2 1 6 for the roll
More informationThe Heat Equation. Lectures INF2320 p. 1/88
The Heat Equation Lectures INF232 p. 1/88 Lectures INF232 p. 2/88 The Heat Equation We study the heat equation: u t = u xx for x (,1), t >, (1) u(,t) = u(1,t) = for t >, (2) u(x,) = f(x) for x (,1), (3)
More informationHYPOTHESIS TESTING: POWER OF THE TEST
HYPOTHESIS TESTING: POWER OF THE TEST The first 6 steps of the 9-step test of hypothesis are called "the test". These steps are not dependent on the observed data values. When planning a research project,
More informationThe Exponential Distribution
21 The Exponential Distribution From Discrete-Time to Continuous-Time: In Chapter 6 of the text we will be considering Markov processes in continuous time. In a sense, we already have a very good understanding
More informationTHE CENTRAL LIMIT THEOREM TORONTO
THE CENTRAL LIMIT THEOREM DANIEL RÜDT UNIVERSITY OF TORONTO MARCH, 2010 Contents 1 Introduction 1 2 Mathematical Background 3 3 The Central Limit Theorem 4 4 Examples 4 4.1 Roulette......................................
More information1 if 1 x 0 1 if 0 x 1
Chapter 3 Continuity In this chapter we begin by defining the fundamental notion of continuity for real valued functions of a single real variable. When trying to decide whether a given function is or
More informationMATH 425, PRACTICE FINAL EXAM SOLUTIONS.
MATH 45, PRACTICE FINAL EXAM SOLUTIONS. Exercise. a Is the operator L defined on smooth functions of x, y by L u := u xx + cosu linear? b Does the answer change if we replace the operator L by the operator
More informationMessage-passing sequential detection of multiple change points in networks
Message-passing sequential detection of multiple change points in networks Long Nguyen, Arash Amini Ram Rajagopal University of Michigan Stanford University ISIT, Boston, July 2012 Nguyen/Amini/Rajagopal
More informationTrading regret rate for computational efficiency in online learning with limited feedback
Trading regret rate for computational efficiency in online learning with limited feedback Shai Shalev-Shwartz TTI-C Hebrew University On-line Learning with Limited Feedback Workshop, 2009 June 2009 Shai
More informationAn Introduction to Partial Differential Equations
An Introduction to Partial Differential Equations Andrew J. Bernoff LECTURE 2 Cooling of a Hot Bar: The Diffusion Equation 2.1. Outline of Lecture An Introduction to Heat Flow Derivation of the Diffusion
More informationMetric Spaces. Lecture Notes and Exercises, Fall 2015. M.van den Berg
Metric Spaces Lecture Notes and Exercises, Fall 2015 M.van den Berg School of Mathematics University of Bristol BS8 1TW Bristol, UK mamvdb@bristol.ac.uk 1 Definition of a metric space. Let X be a set,
More informationUn point de vue bayésien pour des algorithmes de bandit plus performants
Un point de vue bayésien pour des algorithmes de bandit plus performants Emilie Kaufmann, Telecom ParisTech Rencontre des Jeunes Statisticiens, Aussois, 28 août 2013 Emilie Kaufmann (Telecom ParisTech)
More informationNotes on metric spaces
Notes on metric spaces 1 Introduction The purpose of these notes is to quickly review some of the basic concepts from Real Analysis, Metric Spaces and some related results that will be used in this course.
More information1 The Brownian bridge construction
The Brownian bridge construction The Brownian bridge construction is a way to build a Brownian motion path by successively adding finer scale detail. This construction leads to a relatively easy proof
More informationMonte Carlo Simulation
1 Monte Carlo Simulation Stefan Weber Leibniz Universität Hannover email: sweber@stochastik.uni-hannover.de web: www.stochastik.uni-hannover.de/ sweber Monte Carlo Simulation 2 Quantifying and Hedging
More informationBasics of Statistical Machine Learning
CS761 Spring 2013 Advanced Machine Learning Basics of Statistical Machine Learning Lecturer: Xiaojin Zhu jerryzhu@cs.wisc.edu Modern machine learning is rooted in statistics. You will find many familiar
More informationSimple Linear Regression Inference
Simple Linear Regression Inference 1 Inference requirements The Normality assumption of the stochastic term e is needed for inference even if it is not a OLS requirement. Therefore we have: Interpretation
More informationUndergraduate Notes in Mathematics. Arkansas Tech University Department of Mathematics
Undergraduate Notes in Mathematics Arkansas Tech University Department of Mathematics An Introductory Single Variable Real Analysis: A Learning Approach through Problem Solving Marcel B. Finan c All Rights
More informationCorrected Diffusion Approximations for the Maximum of Heavy-Tailed Random Walk
Corrected Diffusion Approximations for the Maximum of Heavy-Tailed Random Walk Jose Blanchet and Peter Glynn December, 2003. Let (X n : n 1) be a sequence of independent and identically distributed random
More informationIEOR 6711: Stochastic Models I Fall 2012, Professor Whitt, Tuesday, September 11 Normal Approximations and the Central Limit Theorem
IEOR 6711: Stochastic Models I Fall 2012, Professor Whitt, Tuesday, September 11 Normal Approximations and the Central Limit Theorem Time on my hands: Coin tosses. Problem Formulation: Suppose that I have
More informationBandwidth Selection for Nonparametric Distribution Estimation
Bandwidth Selection for Nonparametric Distribution Estimation Bruce E. Hansen University of Wisconsin www.ssc.wisc.edu/~bhansen May 2004 Abstract The mean-square efficiency of cumulative distribution function
More informationStatistical Machine Learning from Data
Samy Bengio Statistical Machine Learning from Data 1 Statistical Machine Learning from Data Gaussian Mixture Models Samy Bengio IDIAP Research Institute, Martigny, Switzerland, and Ecole Polytechnique
More informationPart II Redundant Dictionaries and Pursuit Algorithms
Aisenstadt Chair Course CRM September 2009 Part II Redundant Dictionaries and Pursuit Algorithms Stéphane Mallat Centre de Mathématiques Appliquées Ecole Polytechnique Sparsity in Redundant Dictionaries
More informationParabolic Equations. Chapter 5. Contents. 5.1.2 Well-Posed Initial-Boundary Value Problem. 5.1.3 Time Irreversibility of the Heat Equation
7 5.1 Definitions Properties Chapter 5 Parabolic Equations Note that we require the solution u(, t bounded in R n for all t. In particular we assume that the boundedness of the smooth function u at infinity
More informationEXISTENCE AND NON-EXISTENCE RESULTS FOR A NONLINEAR HEAT EQUATION
Sixth Mississippi State Conference on Differential Equations and Computational Simulations, Electronic Journal of Differential Equations, Conference 5 (7), pp. 5 65. ISSN: 7-669. UL: http://ejde.math.txstate.edu
More informationA SURVEY ON CONTINUOUS ELLIPTICAL VECTOR DISTRIBUTIONS
A SURVEY ON CONTINUOUS ELLIPTICAL VECTOR DISTRIBUTIONS Eusebio GÓMEZ, Miguel A. GÓMEZ-VILLEGAS and J. Miguel MARÍN Abstract In this paper it is taken up a revision and characterization of the class of
More informationModern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh
Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh Peter Richtárik Week 3 Randomized Coordinate Descent With Arbitrary Sampling January 27, 2016 1 / 30 The Problem
More informationBrownian Motion and Stochastic Flow Systems. J.M Harrison
Brownian Motion and Stochastic Flow Systems 1 J.M Harrison Report written by Siva K. Gorantla I. INTRODUCTION Brownian motion is the seemingly random movement of particles suspended in a fluid or a mathematical
More informationSupplement to Call Centers with Delay Information: Models and Insights
Supplement to Call Centers with Delay Information: Models and Insights Oualid Jouini 1 Zeynep Akşin 2 Yves Dallery 1 1 Laboratoire Genie Industriel, Ecole Centrale Paris, Grande Voie des Vignes, 92290
More informationA Short and Elementary proof of the main Bahadur-Kiefer Theorem
EINDHOVEN UNIVERSITY OF TECHNOLOGY Department of Mathematics and Computing Science Memorandum COSOR 94-26 A Short and Elementary proof of the main Bahadur-Kiefer Theorem John H.J. Einmahl Eindhoven, August
More informationTesting against a Change from Short to Long Memory
Testing against a Change from Short to Long Memory Uwe Hassler and Jan Scheithauer Goethe-University Frankfurt This version: December 9, 2007 Abstract This paper studies some well-known tests for the null
More informationUncertainty quantification for the family-wise error rate in multivariate copula models
Uncertainty quantification for the family-wise error rate in multivariate copula models Thorsten Dickhaus (joint work with Taras Bodnar, Jakob Gierl and Jens Stange) University of Bremen Institute for
More informationSOLUTIONS TO EXERCISES FOR. MATHEMATICS 205A Part 3. Spaces with special properties
SOLUTIONS TO EXERCISES FOR MATHEMATICS 205A Part 3 Fall 2008 III. Spaces with special properties III.1 : Compact spaces I Problems from Munkres, 26, pp. 170 172 3. Show that a finite union of compact subspaces
More informationRANDOM INTERVAL HOMEOMORPHISMS. MICHA L MISIUREWICZ Indiana University Purdue University Indianapolis
RANDOM INTERVAL HOMEOMORPHISMS MICHA L MISIUREWICZ Indiana University Purdue University Indianapolis This is a joint work with Lluís Alsedà Motivation: A talk by Yulij Ilyashenko. Two interval maps, applied
More informationt := maxγ ν subject to ν {0,1,2,...} and f(x c +γ ν d) f(x c )+cγ ν f (x c ;d).
1. Line Search Methods Let f : R n R be given and suppose that x c is our current best estimate of a solution to P min x R nf(x). A standard method for improving the estimate x c is to choose a direction
More informationUNIFORM ASYMPTOTICS FOR DISCOUNTED AGGREGATE CLAIMS IN DEPENDENT RISK MODELS
Applied Probability Trust 2 October 2013 UNIFORM ASYMPTOTICS FOR DISCOUNTED AGGREGATE CLAIMS IN DEPENDENT RISK MODELS YANG YANG, Nanjing Audit University, and Southeast University KAIYONG WANG, Southeast
More informationTesting against a Change from Short to Long Memory
Testing against a Change from Short to Long Memory Uwe Hassler and Jan Scheithauer Goethe-University Frankfurt This version: January 2, 2008 Abstract This paper studies some well-known tests for the null
More information4. Expanding dynamical systems
4.1. Metric definition. 4. Expanding dynamical systems Definition 4.1. Let X be a compact metric space. A map f : X X is said to be expanding if there exist ɛ > 0 and L > 1 such that d(f(x), f(y)) Ld(x,
More informationThe Variability of P-Values. Summary
The Variability of P-Values Dennis D. Boos Department of Statistics North Carolina State University Raleigh, NC 27695-8203 boos@stat.ncsu.edu August 15, 2009 NC State Statistics Departement Tech Report
More informationDetekce změn v autoregresních posloupnostech
Nové Hrady 2012 Outline 1 Introduction 2 3 4 Change point problem (retrospective) The data Y 1,..., Y n follow a statistical model, which may change once or several times during the observation period
More information0 <β 1 let u(x) u(y) kuk u := sup u(x) and [u] β := sup
456 BRUCE K. DRIVER 24. Hölder Spaces Notation 24.1. Let Ω be an open subset of R d,bc(ω) and BC( Ω) be the bounded continuous functions on Ω and Ω respectively. By identifying f BC( Ω) with f Ω BC(Ω),
More informationAsymptotics for a discrete-time risk model with Gamma-like insurance risks. Pokfulam Road, Hong Kong
Asymptotics for a discrete-time risk model with Gamma-like insurance risks Yang Yang 1,2 and Kam C. Yuen 3 1 Department of Statistics, Nanjing Audit University, Nanjing, 211815, China 2 School of Economics
More informationLarge Sample Theory. Consider a sequence of random variables Z 1, Z 2,..., Z n. Convergence in probability: Z n
Large Samle Theory In statistics, we are interested in the roerties of articular random variables (or estimators ), which are functions of our data. In ymtotic analysis, we focus on describing the roerties
More information1. Let X and Y be normed spaces and let T B(X, Y ).
Uppsala Universitet Matematiska Institutionen Andreas Strömbergsson Prov i matematik Funktionalanalys Kurs: NVP, Frist. 2005-03-14 Skrivtid: 9 11.30 Tillåtna hjälpmedel: Manuella skrivdon, Kreyszigs bok
More informationSection 7.1. Introduction to Hypothesis Testing. Schrodinger s cat quantum mechanics thought experiment (1935)
Section 7.1 Introduction to Hypothesis Testing Schrodinger s cat quantum mechanics thought experiment (1935) Statistical Hypotheses A statistical hypothesis is a claim about a population. Null hypothesis
More informationChapter 4 Statistical Inference in Quality Control and Improvement. Statistical Quality Control (D. C. Montgomery)
Chapter 4 Statistical Inference in Quality Control and Improvement 許 湘 伶 Statistical Quality Control (D. C. Montgomery) Sampling distribution I a random sample of size n: if it is selected so that the
More information1 Sufficient statistics
1 Sufficient statistics A statistic is a function T = rx 1, X 2,, X n of the random sample X 1, X 2,, X n. Examples are X n = 1 n s 2 = = X i, 1 n 1 the sample mean X i X n 2, the sample variance T 1 =
More informationNon Parametric Inference
Maura Department of Economics and Finance Università Tor Vergata Outline 1 2 3 Inverse distribution function Theorem: Let U be a uniform random variable on (0, 1). Let X be a continuous random variable
More informationα α λ α = = λ λ α ψ = = α α α λ λ ψ α = + β = > θ θ β > β β θ θ θ β θ β γ θ β = γ θ > β > γ θ β γ = θ β = θ β = θ β = β θ = β β θ = = = β β θ = + α α α α α = = λ λ λ λ λ λ λ = λ λ α α α α λ ψ + α =
More informationGambling Systems and Multiplication-Invariant Measures
Gambling Systems and Multiplication-Invariant Measures by Jeffrey S. Rosenthal* and Peter O. Schwartz** (May 28, 997.. Introduction. This short paper describes a surprising connection between two previously
More informationCredit Risk Models: An Overview
Credit Risk Models: An Overview Paul Embrechts, Rüdiger Frey, Alexander McNeil ETH Zürich c 2003 (Embrechts, Frey, McNeil) A. Multivariate Models for Portfolio Credit Risk 1. Modelling Dependent Defaults:
More informationProbability Calculator
Chapter 95 Introduction Most statisticians have a set of probability tables that they refer to in doing their statistical wor. This procedure provides you with a set of electronic statistical tables that
More informationA LOGNORMAL MODEL FOR INSURANCE CLAIMS DATA
REVSTAT Statistical Journal Volume 4, Number 2, June 2006, 131 142 A LOGNORMAL MODEL FOR INSURANCE CLAIMS DATA Authors: Daiane Aparecida Zuanetti Departamento de Estatística, Universidade Federal de São
More informationNonparametric adaptive age replacement with a one-cycle criterion
Nonparametric adaptive age replacement with a one-cycle criterion P. Coolen-Schrijner, F.P.A. Coolen Department of Mathematical Sciences University of Durham, Durham, DH1 3LE, UK e-mail: Pauline.Schrijner@durham.ac.uk
More informationPlanning And Scheduling Maintenance Resources In A Complex System
Planning And Scheduling Maintenance Resources In A Complex System Martin Newby School of Engineering and Mathematical Sciences City University of London London m.j.newbycity.ac.uk Colin Barker QuaRCQ Group
More information6.2 Permutations continued
6.2 Permutations continued Theorem A permutation on a finite set A is either a cycle or can be expressed as a product (composition of disjoint cycles. Proof is by (strong induction on the number, r, of
More informationT ( a i x i ) = a i T (x i ).
Chapter 2 Defn 1. (p. 65) Let V and W be vector spaces (over F ). We call a function T : V W a linear transformation form V to W if, for all x, y V and c F, we have (a) T (x + y) = T (x) + T (y) and (b)
More informationIs Infrastructure Capital Productive? A Dynamic Heterogeneous Approach.
Is Infrastructure Capital Productive? A Dynamic Heterogeneous Approach. César Calderón a, Enrique Moral-Benito b, Luis Servén a a The World Bank b CEMFI International conference on Infrastructure Economics
More informationInequalities of Analysis. Andrejs Treibergs. Fall 2014
USAC Colloquium Inequalities of Analysis Andrejs Treibergs University of Utah Fall 2014 2. USAC Lecture: Inequalities of Analysis The URL for these Beamer Slides: Inequalities of Analysis http://www.math.utah.edu/~treiberg/inequalitiesslides.pdf
More informationStatistical Machine Learning
Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes
More informationPrime numbers and prime polynomials. Paul Pollack Dartmouth College
Prime numbers and prime polynomials Paul Pollack Dartmouth College May 1, 2008 Analogies everywhere! Analogies in elementary number theory (continued fractions, quadratic reciprocity, Fermat s last theorem)
More informationWHEN DOES A RANDOMLY WEIGHTED SELF NORMALIZED SUM CONVERGE IN DISTRIBUTION?
WHEN DOES A RANDOMLY WEIGHTED SELF NORMALIZED SUM CONVERGE IN DISTRIBUTION? DAVID M MASON 1 Statistics Program, University of Delaware Newark, DE 19717 email: davidm@udeledu JOEL ZINN 2 Department of Mathematics,
More informationproblem arises when only a non-random sample is available differs from censored regression model in that x i is also unobserved
4 Data Issues 4.1 Truncated Regression population model y i = x i β + ε i, ε i N(0, σ 2 ) given a random sample, {y i, x i } N i=1, then OLS is consistent and efficient problem arises when only a non-random
More informationMonitoring Structural Change in Dynamic Econometric Models
Monitoring Structural Change in Dynamic Econometric Models Achim Zeileis Friedrich Leisch Christian Kleiber Kurt Hornik http://www.ci.tuwien.ac.at/~zeileis/ Contents Model frame Generalized fluctuation
More informationThe one dimensional heat equation: Neumann and Robin boundary conditions
The one dimensional heat equation: Neumann and Robin boundary conditions Ryan C. Trinity University Partial Differential Equations February 28, 2012 with Neumann boundary conditions Our goal is to solve:
More informationUniversity of Maryland Fraternity & Sorority Life Spring 2015 Academic Report
University of Maryland Fraternity & Sorority Life Academic Report Academic and Population Statistics Population: # of Students: # of New Members: Avg. Size: Avg. GPA: % of the Undergraduate Population
More informationCombining Weak Statistical Evidence in Cyber Security
Combining Weak Statistical Evidence in Cyber Security Nick Heard Department of Mathematics, Imperial College London; Heilbronn Institute for Mathematical Research, University of Bristol Intelligent Data
More informationThe Math Circle, Spring 2004
The Math Circle, Spring 2004 (Talks by Gordon Ritter) What is Non-Euclidean Geometry? Most geometries on the plane R 2 are non-euclidean. Let s denote arc length. Then Euclidean geometry arises from the
More informationStirling s formula, n-spheres and the Gamma Function
Stirling s formula, n-spheres and the Gamma Function We start by noticing that and hence x n e x dx lim a 1 ( 1 n n a n n! e ax dx lim a 1 ( 1 n n a n a 1 x n e x dx (1 Let us make a remark in passing.
More informationA Coefficient of Variation for Skewed and Heavy-Tailed Insurance Losses. Michael R. Powers[ 1 ] Temple University and Tsinghua University
A Coefficient of Variation for Skewed and Heavy-Tailed Insurance Losses Michael R. Powers[ ] Temple University and Tsinghua University Thomas Y. Powers Yale University [June 2009] Abstract We propose a
More informationChapter 7: Proportional Play and the Kelly Betting System
Chapter 7: Proportional Play and the Kelly Betting System Proportional Play and Kelly s criterion: Investing in the stock market is, in effect, making a series of bets. Contrary to bets in a casino though,
More informationSales forecasting # 1
Sales forecasting # 1 Arthur Charpentier arthur.charpentier@univ-rennes1.fr 1 Agenda Qualitative and quantitative methods, a very general introduction Series decomposition Short versus long term forecasting
More informationarxiv:1412.1183v1 [q-fin.rm] 3 Dec 2014
Regulatory Capital Modelling for Credit Risk arxiv:1412.1183v1 [q-fin.rm] 3 Dec 2014 Marek Rutkowski a, Silvio Tarca a, a School of Mathematics and Statistics F07, University of Sydney, NSW 2006, Australia.
More informationI. Pointwise convergence
MATH 40 - NOTES Sequences of functions Pointwise and Uniform Convergence Fall 2005 Previously, we have studied sequences of real numbers. Now we discuss the topic of sequences of real valued functions.
More informationMetric Spaces. Chapter 7. 7.1. Metrics
Chapter 7 Metric Spaces A metric space is a set X that has a notion of the distance d(x, y) between every pair of points x, y X. The purpose of this chapter is to introduce metric spaces and give some
More informationCounting Primes whose Sum of Digits is Prime
2 3 47 6 23 Journal of Integer Sequences, Vol. 5 (202), Article 2.2.2 Counting Primes whose Sum of Digits is Prime Glyn Harman Department of Mathematics Royal Holloway, University of London Egham Surrey
More information