THE STATISTICAL TREATMENT OF EXPERIMENTAL DATA 1



Similar documents
INTRODUCTION TO ERRORS AND ERROR ANALYSIS

AP Physics 1 and 2 Lab Investigations

Welcome to Physics 40!

EXPERIMENTAL ERROR AND DATA ANALYSIS

99.37, 99.38, 99.38, 99.39, 99.39, 99.39, 99.39, 99.40, 99.41, cm

Nuclear Physics Lab I: Geiger-Müller Counter and Nuclear Counting Statistics

Figure 1. A typical Laboratory Thermometer graduated in C.

= δx x + δy y. df ds = dx. ds y + xdy ds. Now multiply by ds to get the form of the equation in terms of differentials: df = y dx + x dy.

Experimental Uncertainties (Errors)

CHI-SQUARE: TESTING FOR GOODNESS OF FIT

Experiment #1, Analyze Data using Excel, Calculator and Graphs.

An Innovative Method for Dead Time Correction in Nuclear Spectroscopy

Introduction to Error Analysis

Carbon Dioxide and an Argon + Nitrogen Mixture. Measurement of C p /C v for Argon, Nitrogen, Stephen Lucas 05/11/10

Standard Deviation Estimator

Experiment 10. Radioactive Decay of 220 Rn and 232 Th Physics 2150 Experiment No. 10 University of Colorado

Biostatistics: DESCRIPTIVE STATISTICS: 2, VARIABILITY

An Introduction to Basic Statistics and Probability

Introduction to Hypothesis Testing

71M6521 Energy Meter IC. Real Time Clock Compensation. The Challenge. The RTC in the 71M6521D/F. Theory of Operation APPLICATION NOTE

Balance Accuracy. FlexiWeigh

Northumberland Knowledge

Physics 40 Lab 1: Tests of Newton s Second Law

Physics Lab Report Guidelines

CALCULATIONS & STATISTICS

Problem Solving and Data Analysis

Evaluation of Quantitative Data (errors/statistical analysis/propagation of error)

Measurement with Ratios

" Y. Notation and Equations for Regression Lecture 11/4. Notation:

Vilnius University. Faculty of Mathematics and Informatics. Gintautas Bareikis

PS Chapter 1 Review. Name: Class: Date: Multiple Choice Identify the choice that best completes the statement or answers the question.

Chi Square Tests. Chapter Introduction

Statistics 104: Section 6!

CORRELATED TO THE SOUTH CAROLINA COLLEGE AND CAREER-READY FOUNDATIONS IN ALGEBRA

Radioactivity III: Measurement of Half Life.

The accurate calibration of all detectors is crucial for the subsequent data

Data Acquisition And Analysis

Tutorial 4.6 Gamma Spectrum Analysis

MEASURES OF VARIATION

NEW MEXICO Grade 6 MATHEMATICS STANDARDS

On Correlating Performance Metrics

Simple Regression Theory II 2010 Samuel L. Baker

6 EXTENDING ALGEBRA. 6.0 Introduction. 6.1 The cubic equation. Objectives

Why is Insurance Good? An Example Jon Bakija, Williams College (Revised October 2013)

UNCERTAINTY AND CONFIDENCE IN MEASUREMENT

Chapter 1 Units, Physical Quantities, and Vectors

Descriptive Statistics

Implementing the combing method in the dynamic Monte Carlo. Fedde Kuilman PNR_131_2012_008

Determining the Acceleration Due to Gravity

9.2 Summation Notation

A Short Guide to Significant Figures

Measurement. Stephanie Bell. Issue 2

Measuring Line Edge Roughness: Fluctuations in Uncertainty

Financial Mathematics and Simulation MATH Spring 2011 Homework 2

The Method of Least Squares

3 More on Accumulation and Discount Functions

AMS 5 CHANCE VARIABILITY

Chapter 20: chance error in sampling

Wentzville School District Algebra 1: Unit 8 Stage 1 Desired Results

3.1. Solving linear equations. Introduction. Prerequisites. Learning Outcomes. Learning Style

Concepts in Investments Risks and Returns (Relevant to PBE Paper II Management Accounting and Finance)

Chapter 3 Review Math 1030

S-Parameters and Related Quantities Sam Wetterlin 10/20/09

In mathematics, there are four attainment targets: using and applying mathematics; number and algebra; shape, space and measures, and handling data.

AN_6521_035 APRIL 2009

Interactive comment on Total cloud cover from satellite observations and climate models by P. Probst et al.

Revised Version of Chapter 23. We learned long ago how to solve linear congruences. ax c (mod m)

Lies My Calculator and Computer Told Me

Introduction to CCDs and CCD Data Calibration

Antoine Henri Becquerel was born in Paris on December 15, 1852

Measurement of Length, Mass, Volume and Density

Numerical Matrix Analysis

Using simulation to calculate the NPV of a project

Unamended Quantum Mechanics Rigorously Implies Awareness Is Not Based in the Physical Brain

Summary of Formulas and Concepts. Descriptive Statistics (Ch. 1-4)

H. The Study Design. William S. Cash and Abigail J. Moss National Center for Health Statistics

Notes on Factoring. MA 206 Kurt Bryan

Conservation of Energy Physics Lab VI

Profit Forecast Model Using Monte Carlo Simulation in Excel

Theory versus Experiment. Prof. Jorgen D Hondt Vrije Universiteit Brussel jodhondt@vub.ac.be

Estimation and Confidence Intervals

Introduction to. Hypothesis Testing CHAPTER LEARNING OBJECTIVES. 1 Identify the four steps of hypothesis testing.

17. SIMPLE LINEAR REGRESSION II

Probability and Expected Value

The Deadly Sins of Algebra

Equations, Inequalities & Partial Fractions

Why Taking This Course? Course Introduction, Descriptive Statistics and Data Visualization. Learning Goals. GENOME 560, Spring 2012

Toothpick Squares: An Introduction to Formulas

ELECTRIC FIELD LINES AND EQUIPOTENTIAL SURFACES

PHYS 1624 University Physics I. PHYS 2644 University Physics II

ARTIFICIAL RADIOACTIVITY

A Comparative Study of the Pickup Method and its Variations Using a Simulated Hotel Reservation Data

Prelab Exercises: Hooke's Law and the Behavior of Springs

Current Standard: Mathematical Concepts and Applications Shape, Space, and Measurement- Primary

4. Continuous Random Variables, the Pareto and Normal Distributions

OPTIONS TRADING AS A BUSINESS UPDATE: Using ODDS Online to Find A Straddle s Exit Point

Local Sidereal Time is the hour angle of the First Point of Aries, and is equal to the hour angle plus right ascension of any star.

Calculate Highest Common Factors(HCFs) & Least Common Multiples(LCMs) NA1

OBJECTIVE ASSESSMENT OF FORECASTING ASSIGNMENTS USING SOME FUNCTION OF PREDICTION ERRORS

Transcription:

THE STATISTICAL TREATMET OF EXPERIMETAL DATA Introduction The subject of statistical data analysis is regarded as crucial by most scientists, since error-free measurement is impossible in virtually all experimental sciences, natural or social. Experimentalists gather data with the aim of formulating a physically reasonable model to describe a particular phenomenon. But, given an experimental data set, how does one know whether it agrees with a theoretical prediction? There is no guarantee that a theory will describe the real situation -- the basis of science is the demand that any idea be validated experimentally. To know whether your measurement is consistent with theory or not, you need to know whether the uncertainty in the measurement is smaller or larger than the discrepancy between the measurement and the theoretical prediction. In symbols we would say that if v th denotes the theoretically predicted value of a given quantity, v ex denotes the experimental result, and v denotes the uncertainty expected in the measurement, then the theory is consistent with experiment only if v ex - Δv v th v ex + Δv (S-) The next question is how to determine v for a particular experiment and measurement apparatus. The candid truth is that v is often estimated from one's knowledge of the accuracy of the measuring device (sometimes this is called an external determination). For example in measuring the length of an object using a meter stick whose finest marks are mm apart, one should attempt to make measurements to a fraction of mm. Depending upon the object being measured, and the ability to clearly align the ruler with the object, one might estimate that v would be about 0.5 mm (for example). However, you should be aware that when scientists refer to experimental error, they usually mean the uncertainty in their measurement, which is a measure of precision and not accuracy. With this in mind, we realize that the spread observed in a series of repeated measurements allows us to determine uncertainties based upon statistics (sometimes called an internal determination). Some of this advice is based upon a handout that originated in the Physics Department at Haverford College, Haverford, PA. The title and some of the material is based upon the treatment found in H. D. Young, Statistical Treatment of Experimental Data (McGraw-Hill, ew York, 96), which is a fine reference to the subject in general.

The Statistical Treatment of Experimental Data It is important to keep in mind the distinction between systematic and statistical uncertainty. Imagine that because of the humidity in Central Illinois at this time of year, the meter stick you are using has swelled in length by a small amount. In that case, it will underestimate by a corresponding amount the length of any object it is used to measure. We call that sort of uncertainty systematic and try to design experiments to avoid it, because if it is overlooked, it will lead to an incorrect (inaccurate) result, despite a full statistical analysis which might give us a measure of the precision of our technique. You may end up finding some example of insidious systematic errors in this lab. Statistical uncertainty, on the other hand, is the uncertainty that is due to small random events beyond our control, that do not lead to bias in either direction. Statistical uncertainty can be reduced by repeating measurements, and it can be accurately determined from the spread of values that occur. Therefore we prefer it to systematic error. ;) The following material is devoted mostly to determination of statistical uncertainty and something called the propagation of uncertainty. The theoretical results you will learn below are mathematically exactly true for a certain, commonly encountered type of statistical uncertainty (i.e., based on a Gaussian distribution). [See note at end of write-up] Dealing with Systematic Error Generally speaking, a series of independent measurements allow us to reduce random errors, but independent experiments are required to gauge systematic errors. Still there are often simple checks that allow you to subtract offsets or drift which might otherwise be included in your dataset. Imagine you have been assigned to determine whether a particular radiation source represents a health hazard. To determine the health hazard, you might use a radiation detector called a Geiger counter to measure incoming high-energy particles given off by atomic nuclei as they undergo radioactive decay. However, if you turned on your detector, even with no source of radioactivity nearby, you would detect background radiation. This is because high-energy particles, mostly electrons and gamma rays, arrive at the earth from space every moment. These cosmic rays set off your detector, and they cannot be eliminated. In addition, there are some radioactive nuclei in building materials, so that the bricks and mortar of our lab building are emitting radiation, too Even our bodies contain unstable 40 K nuclei that contribute slightly to the background radiation level. This background intensity (number of counts per unit time) represents a source of systematic uncertainty, since it adds to any measurement of radiation we make in the room. We would need to subtract off this background intensity from our

The Statistical Treatment of Experimental Data 3 measurement of any extra intensity caused by our radioactive source. The quantity of interest, the source intensity, is Is = It - Ib (S-) where Is indicates the source intensity, It the total intensity, and Ib the background intensity. The uncertainty in our measurement of Is still needs to be determined. Determining Statistical Uncertainty It stands to reason that repeating measurements is a good way to increase the accuracy of the overall process. (You would never poll only three people to establish an approval rating for the President.) It turns out that from a reasonably large sample of measured values we can also accurately estimate the uncertainty Given measurements (here, is the number of groups reporting measurements){i, I, I }, of some quantity I (such as the background radiation intensity), the best estimate of the actual value of I is the mean or average of the values. I = " I i. (S-3) We also want to deduce I, the uncertainty associated with this measured value. We can estimate this quantity from the spread of individual values of the I i s that occurred. It should also depend on how many measurements were made, but we ll deal with that after we decide how to characterize the spread of values obtained. Statisticians and physicists have devised an accepted measure of the spread of values that occur in a distribution and call it the standard deviation, σ. It is defined in terms of an average of the deviation of the individual measurements from the average value I. (If we just averaged the deviations, we would get roughly zero, since the measurements above and below the mean would almost cancel out. Therefore we average the square of the deviations and take the square root at the end.) The definition of σ will be " = $ ( I i # I ). (S-4) Eq. (S-4) is a bit of a pain even on a calculator and it can be shown that the following expression is totally equivalent, so we usually use it: " = I i # $ ' $ I i * (S-5) & )

4 The Statistical Treatment of Experimental Data The second term under the square root sign is just I, so that part is easy to obtain. The only real new punching of calculator buttons is the accumulation of the sum of the squares of the I i s in the first term. OTE: If is a relatively small number, the uncertainty estimated from (S-4) or (S- 5) is slightly underestimated. (After all, if =, you would get σ = 0 erroneously.) To approximately correct for this underestimate, use (-) instead of in equations S-4 and S-5 if is less than 0 or so. One might be tempted to assume that σ is also a good estimate of I. That is, if you took many identical data sets and used them to compute I, one might suppose that the standard deviation of I would be σ. This is not the case. Particularly for large we will know I considerably more accurately than I ± σ. In fact you would find that if we all did another measurement thereby doubling, we would find that σ would not change appreciably since it is just a measure of the natural spread in values resulting from our measuring procedure, and yet the measurement of I would be improved. It can be shown that a better estimate for the error bar is the so-called standard deviation of the mean, σ m, given by " m = " (S-6) Thus we would typically say we have determined I to be I = I ± " m. The exact meaning of this statement is discussed in statistics texts; the following interpretation is good enough for most purposes: there is about a /3 chance that the actual value of I lies in the interval [ I "# m,i + # m ], a 95% chance that it is in the interval [ I " # m,i + # m ], and a 99.7% chance that it is in the interval [ I " 3# m,i + 3# m ], etc. Most scientists are rather conservative, and would report the uncertainty, "I, as being larger than σ m, e.g. σ m. Propagation of Uncertainties It frequently happens in experimental science that the desired quantity depends on several other quantities, each of which has an associated uncertainty. In that case we need to know how to compute the uncertainty in the final quantity from those of the measured quantities. This process is called propagation of uncertainty. This topic is treated in numerous texts, such as Experimentation: An Introduction to Measurement Theory and Experiment Design, by D.C. Baird or Data Reduction and Error Analysis for the Physical Sciences, by P.R. Bevington and D.K. Robinson. We will simply summarize several simple results that will be useful to you in this course. First, suppose that you have measured two quantities, a and b, that are added or subtracted to produce a final result. If we may presume that there are no correlations between the fluctuations in our measurements of a

The Statistical Treatment of Experimental Data 5 and the fluctuations in our measurement of b, then the best estimate of the uncertainty of the sum or difference is "( a ± b) = ("a) + ("b). (S-7) We say that the uncertainties are added in quadrature. ote that if the two uncertainties are equal, the resultant uncertainty is larger than either one alone by a factor. Also, if the uncertainty in one is less than about /4 of the other, its contribution to the overall uncertainty is negligible. In other experiments, you may be multiplying or dividing two quantities. In that case, the relative or fractional uncertainties that are added in quadrature: "( ab) ab = #"a& $ a ' # + "b & $ b '. (S-8) This implies, for example, that if a is.0% of a and b is 4.0% of b, then ab is [(0.00) + (0.040) ] /, or 4.5%, of ab. There are more complicated rules for other manipulations. For example, if you raise a measured value to a power the relative uncertainty is scaled by that power "( a n ) = n "a a n a summarized as follows. For function of two independent variables, z = f(x,y), (S-9) where it has been assumed that there is no uncertainty in n, only in a. Such rules can be "z = $ #z ' & ) "x + #z $ ' & ) "y (S-0) %#x( %#y ( If z is a function of more than two variables the equation is extended by adding similar terms. OTE: These results are mathematically exact only for Gaussian distributions. Some sorts of data, radioactive counting measurements for example, actually follow "Poisson" distributions, which are slightly different. However, the error introduced by this fact is often relatively small.