What is Effect Size? effect size in Information Technology, Learning, and Performance Journal manuscripts.
|
|
|
- Claude Banks
- 10 years ago
- Views:
Transcription
1 The Incorporation of Effect Size in Information Technology, Learning, and Performance Research Joe W. Kotrlik Heather A. Williams Research manuscripts published in the Information Technology, Learning, and Performance Journal are expected to adhere to the publication guidelines of the American Psychological Association (2001) and generally accepted research and statistical methodology. This manuscript describes the rationale supporting the reporting of effect size in quantitative research and also provides examples of how to calculate effect size for some of the most common statistical analyses. We include a table of recommendations for effect size interpretation. We also address basic assumptions and cautions on the reporting of effect size. Introduction Effect size is a term used to describe a family of indices that measure the magnitude of a treatment effect. Effect size is different from significance tests because these measures focus on the meaningfulness of the results and allow comparison between studies, furthering the ability of researchers to judge the practical significance of results presented. It is almost always necessary to include some index of effect size or strength of relationship in your results section so that the consumers of your research will be able to truly understand the importance of your findings and make comparisons among studies. A review of manuscripts in the Information Technology, Learning, and Performance Journal from the past three years revealed that effect size has not been consistently reported. Out of the 16 manuscripts published from , effect size should have been reported in 11 manuscripts. An analysis of these 11 manuscripts revealed that effect size was reported correctly in one manuscript, not reported at all in six manuscripts, and reported incorrectly or inappropriately in the remaining four manuscripts. The need to improve the reporting of effect size is an issue being addressed by numerous journals. This article will address the use and reporting of effect size in Information Technology, Learning, and Performance Journal manuscripts. Research/Theoretical Base What is Effect Size? The term effect size has become increasingly popular throughout educational literature in recent decades; and even more prevalent in recent years. The APA Task Force emphasized that researchers should always provide some effect-size estimate when reporting a p value.... reporting and interpreting effect sizes in the context of previously reported effects is essential to good research (Wilkinson & APA Task Force on Statistical Inference, 1999, p. 599). Kirk (1996) also emphasized the importance of the effect size concept. Joe W. Kotrlik is Professor, School of Human Resource Management and Workforce Development, Louisiana State University, Baton Rouge, Louisiana. Heather A. Williams is performance improvement consultant, Reserve, Louisiana and a recent graduate, School of Human Resource Education and Workforce Development, Louisiana State University, Baton Rouge, Louisiana. Information Technology, Learning, and Performance Journal, Vol. 21, No. 1, Spring
2 2 Kotrlik and Williams With these discussions focusing on effect size, the literature presents a variety of effect size definitions. Common definitions include: a standardized value that estimates the magnitude of the differences between groups (Thomas, Salazar, & Landers, 1991), a standardized mean difference (Olejnik & Algina, 2000; Vacha-Haase, 2001), the degree to which sample results diverge from the null hypothesis (Cohen, 1988, 1994), a measure of the degree of difference or association deemed large enough to be of practical significance (Morse, 1998), an estimate of the degree to which the phenomenon being studied exists in the population (e.g., a correlation or difference in means) (Hair, Anderson, Tatham, & Black, 1995), and strength of relationship (American Psychological Association, 2001). Among the numerous effect size definitions, the majority of definitions include the descriptors standardized difference between means or standardized measure of association. In practice, researchers commonly use two categories of measures of effect size in the literature, namely, measures of effect size (according to group mean differences), and measures of strength of association (according to variance accounted for) (Maxwell & Delaney, 1990). Why is Effect Size Important? An early discussion of effect size by Karl Pearson (1901) addressed the idea that statistical significance provides the reader with only part of the story and therefore must be supplemented. Fisher (1925) followed up on this discussion by proposing that researchers present correlation ratios or measures of the strength of association when reporting research findings. Since that time, numerous researchers have argued for the use of effect size statistics to complement or even replace statistical significance testing results, allowing the reader to interpret the results presented as well as providing a method of comparison of results between studies (Cohen, 1965, 1990, 1994; Hays, 1963; Kirk, 1996, 2001; Thompson, 1998, 2002). Effect size can also characterize the degree to which sample results diverge from the null hypothesis (Cohen, 1988, 1994). Therefore, reporting effect size allows a researcher to judge the magnitude of the differences present between groups, increasing the capability of the researcher to compare current research results to previous research and to judge the practical significance of the results derived. Today, the most widely referenced impetus for authors to report effect size is the following statement printed in the 2001 Publication Manual of the American Psychological Association: For the reader to fully understand the importance of your findings, it is almost always necessary to include some index of effect size or strength of relationship in your Results section.... The general principle to be followed... is to provide the reader not only with information about statistical significance but also with enough information to assess the magnitude of the observed effect or relationship. (p ) These statements followed the American Psychological Association Task Force s earlier recommendation strongly urging authors to report effect sizes such as Cohen s d, Glass s delta, eta 2, or adjusted R 2 (Wilkinson & APA Task Force on Statistical Inference, 1999, p. 599). Support for this direction by such a prominent group is witnessed throughout the literature. For example, Fan (2001) described good research as presenting both statistical significance testing results as well as effect sizes. Baugh (2002) declared, Effect size reporting is increasingly recognized as a necessary and responsible practice (p. 255). It is a researcher s responsibility to adhere to the most stringent analytical and reporting methods possible in order to ensure the proper interpretation and application of research results. Effect Size versus Statistical Significance So why the confusion and what is the difference between statistical significance testing and effect sizes? The confusion stems from misconceptions of what statistical significance testing tells us. First, as Nickerson (2000) explains, there is a belief that a small value of p means a treatment effect of large
3 The Incorporation of Effect Size in Information Technology, Learning, and Performance Research 3 magnitude, and that statistical significance means theoretical or practical significance. Statistical significance testing does not imply meaningfulness (Olejnik & Algina, 2000, p. 241). Table 1: Percentage of Time College Faculty Actually Spent Teaching Compared to Percentage of Time College Faculty Preferred to Spend Teaching Actual % Preferred % Comparison Time spent M SD M SD t df p Cohen s d Teaching Note. Information in table taken from Williams, H. A. (2003). Copyright 2003 by Heather A. Williams. Adapted with permission. Statistical significance testing evaluates the probability of obtaining the sampling outcome by chance, while effect size provides some indication of practical meaningfulness (Fan, 2001). Statistical significance relies heavily on sample size, while effect size assists in the interpretation of results and makes trivial effects harder to ignore, further assisting researchers to decide whether results are practically significant (Kirk, 2001). Arguments surrounding the frequent misuse of statistical significance testing populate education and related literature. One argument is that statistical significance is a function of sample size and, therefore, the larger the sample size, the more likely it is that significant results will occur (Fan, 2001; Kirk, 1996; Thompson, 1996). Fan (2001) furthers this argument by positing that the large reliance on statistical significance testing often limits understanding and applicability of research findings in education practice (p. 275). A second argument is that researchers often associate statistical significance with practical significance (Daniel, 1998; Nickerson, 2000). A third argument states that it is the researcher who inappropriately uses and interprets this statistical methodology and, therefore, it is the researcher and not the method that is the source of the deficiency (Cortina & Dunlap, 1997; Kirk, 1996). An example in a recent study demonstrates why the reporting of effect size is important. In this example, failing to report effect size would have resulted in the researcher using the results of a statistically significant t-test to conclude that important differences existed. In this study, Williams (2003) compared the percent of time that faculty actually spent in teaching with the percent of time they preferred to spend in teaching. The data in Table 1 show that even though the t-test was statistically significant (t=2.20, p=.03, df=154), Cohen s effect size value (d=.09) did not meet the standard of even a small effect size. This indicated that the difference has low practical significance, so Williams made no substantive recommendations based on the results of this t-test. When one considers the discussion surrounding effect size in the literature, it is apparent that it is the researcher s responsibility to understand the test results and properly use the reporting methodology which best represents the true findings of the research effort. This responsibility increases in importance because, although a statistically significant outcome may or may not be practically meaningful, a practically meaningful outcome may or may not have occurred by chance (Fan, 2001). These possibilities add challenge to the presentation of results, leading to the recommendation that a researcher should present both the statistical significance test results and an appropriate effect size measure (Carver, 1993; Fagley & McKinney, 1983; Robinson & Levin, 1997; Shaver, 1993; Thompson, 1996, 1998). Basic Assumptions and Cautions on the Use of Effect Size Interpretation We have established the justification for the use of effect sizes. There are a few issues that a researcher must keep in mind, however, when choosing an effect size measure and reporting that measure, aside from which measure goes best with the type of significance test used. First, a researcher must determine the appropriate statistical test to analyze the data. This should be based on sampling assumptions, sample design, and research objectives. Following the determination of the most appropriate significance test, the researcher should
4 4 Kotrlik and Williams decide how the effect size should be calculated and should report information on both the selection of the analytical method and the calculation and selection of the effect size measure. The consistent application of these two steps will further improve the quality of research reporting in the Information Technology, Learning, and Performance Journal. Another issue concerns how effect size is described or interpreted. Selecting the set of descriptors to use for effect size interpretations for correlation coefficients, for example, is similar to deciding the statistical significance level. Just as a researcher might have a set of assumptions or other logical reasons that were used to choose an alpha level of.01 rather than.05, a researcher may decide to use a more conservative set of descriptors when he or she desires more confidence in the declaration of the existence of practical relationships. The set of descriptors by Davis (1971), for example, describes a correlation of.10 as having a low relationship, while Hinkle, Wiersma, and Jurs (1979) describes this same relationship as negligible. Authors should cite the basis for their method of reporting effect sizes. As a final caution, this manuscript provides a rather basic discussion of the appropriate use of effect sizes. It is the researcher s responsibility to investigate the appropriate use of effect sizes in his or her individual research efforts. Measures of Effect Size Although Cohen (1988) and other authors did not anticipate that their scales for the interpretation of effect size magnitude would be so widely used and accepted, these scales have become a required facet of quality research. The acceptance of the use of effect size interpretation scales permeates education literature. A word of caution it is the researcher who is the closest to the data and the design of the study; therefore, it is the researcher who should describe the magnitude of the results based on the study itself and previous research in that area. It is the researcher s responsibility to choose the appropriate method with which to describe the magnitude of an effect size. We present below methods for determining effect size for the most commonly used statistical analyses. Parametric Correlations (Simple and Multiple) Perhaps the simplest measures and reporting of effect sizes exist for correlations. The correlation coefficient itself is a measure of effect size. The most commonly used statistics for parametric correlations are Pearson s r, Spearman s rho, and point bi-serial. The size of these correlation coefficients must be interpreted using a set of descriptors for correlation coefficients as described in Table 2. Regression (Simple and Multiple) Another simple measure of effect size is the multiple regression coefficient, R 2. All standard statistical analysis packages calculate this coefficient automatically, and this coefficient represents the proportion of variance in the dependent variable explained by the independent variable(s). The effect size for this coefficient may be interpreted using a set of descriptors derived from Cohen s f 2 statistic (see Table 2). Independent Samples t-tests One method to estimate effect size for independent samples t-tests is to calculate and interpret Cohen s d statistic (Cohen, 1988). If your statistical analysis program does not calculate Cohen s d, you will need three pieces of information to calculate this statistic: the means of the two samples and the pooled variance estimate. If your program does not calculate Cohen s d, it probably does not provide the pooled standard deviation either. Use the following formulas to calculate Cohen s d: Pooled standard = sqrt[((n 1-1)s 1 2 +(n 2-1)s 2 2 )/((n 1-1)+(n 2-1))] deviation Then, Cohen s d = Difference between sample means Pooled standard deviation After calculating Cohen s d, use the descriptors in Table 2 to interpret this coefficient.
5 The Incorporation of Effect Size in Information Technology, Learning, and Performance Research 5 Table 2: Descriptors for Interpreting Effect Size Source Statistic Value Interpretation Rea & Parker, 1992 Cohen, 1988 Cohen, 1988 R 2 Kirk, 1996 Davis, 1971 a Hinkle, Wiersma, & Jurs, 1979 a Hopkins (1997) a Phi or Cramér s V Cohen s d Cohen s f Omega squared Correlation coefficients Correlation coefficients Correlation coefficients.00 and under.01 Negligible association.10 and under.20 Weak association.20 and under.40 Moderate association.40 and under.60 Relatively strong association.60 and under.80 Strong association.80 and under 1.00 Very strong association.20 Small effect size.50 Medium effect size.80 Large effect size.10 Small effect size.25 Medium effect size.40 Large effect size.0196 Small effect size.1300 Medium effect size.2600 Large effect size.010 Small effect size.059 Medium effect size.138 Large effect size.70 or higher Very strong association.50 to.69 Substantial association.30 to.49 Moderate association.10 to.29 Low association.01 to.09 Negligible association.90 to 1.00 Very high correlation.70 to.90 High correlation.50 to.70 Moderate correlation.30 to.50 Low correlation.00 to.30 Little if any correlation.90 to 1.00 Nearly, practically, or almost: perfect, distinct, infinite.70 to.90 Very large, very high, huge.50 to.70 Large, high, major.30 to.50 Moderate, medium.10 to.30 Small, low, minor.00 to.10 Trivial, very small, insubstantial, tiny, practically zero a Several authors have published guidelines for interpreting the magnitude of correlation coefficients. Three of those sets of guidelines are provided in this table.
6 6 Kotrlik and Williams Paired Samples t-tests To calculate effect size for paired t-tests, Cohen s d statistic may be calculated. If your statistical analysis program does not calculate Cohen s d, you will need two pieces of information to calculate this statistic: the mean of the differences between the pairs and the standard deviation of the differences. Use the following formula to calculate Cohen s d: Cohen s d = Mean difference between the pairs Standard deviation of the differences After calculating Cohen s d, use the descriptors in Table 2 to interpret this coefficient. Analysis of Variance Two methods of reporting effect sizes for analyses of variance are Cohen s f (Cohen, 1988) and omega squared (Ω 2 ). Both provide an estimate of the proportion of variance explained by the categorical variable, with Cohen s f estimating the proportion of variance explained for the sample, while Omega squared estimates the proportion of variance explained for the population. To calculate Cohen s f, eta squared (ω 2 ) must be calculated as follows: w 2 = SS Between /SS Total Then, use the following formula to calculate Cohen s f : Square root of (w 2 /1-w 2 ) Calculate omega squared (Ω 2 ) as follows: W 2 =SS Between-(k-1)MS Within (k=number of groups) SS Total+MS Within The sum of square and mean square information is provided by most statistical programs. After calculating Cohen s f or omega squared, use the descriptors in Table 2 to interpret these coefficients. Contingency Tables In estimating the magnitude of association in contingency tables, use the phi coefficient for a 2 x 2 table. Phi is a Pearson product-moment coefficient calculated on two nominal, dichotomous variables when the categories of both variables are coded 0 and 1. To describe the magnitude of association between categorical variables for a contingency table larger than 2 x 2, use Cramér s V. Many statistical analysis programs, including SPSS and SAS, will automatically calculate either the phi or Cramér s V coefficients. Use the descriptors in Table 2 to interpret these coefficients. Conclusion The effect size measures described are not intended to represent all effect size measures available. Effect size measures were presented only for the most commonly used statistical analyses. The references provided at the end of this article will serve as a good starting point for authors attempting to identify appropriate measures of effect size for other types of statistical analyses. The authors main purpose for writing this article was their hope that Information Technology, Learning, and Performance Journal authors will improve the reporting of their research by including effect size measures when appropriate. We hope that this article will serve as a resource for researchers. References American Psychological Association. (2001). Publication manual of the American Psychological Association (5 th ed.). Washington, DC: Author. Baugh, F. (2002). Correcting effect sizes for score reliability: A reminder that measurement and substantive issues are linked inextricably. Educational and Psychological Measurement, 62(2), Carver, R. (1993). The case against statistical significance testing revisited. Journal of Experimental Education, 61, Cohen, J. (1965). Some statistical issues in psychological research. In B. B. Wolman (Ed.), Handbook of clinical psychology. New York: Academic Press.
7 The Incorporation of Effect Size in Information Technology, Learning, and Performance Research 7 Cohen. J. (1988). Statistical power analysis for the behavioral sciences (2 nd ed.). Hillsdale, NJ: Lawrence Erlbaum. Cohen. J. (1990). Things I have learned (so far). American Psychologist, 45, Cohen, J. (1994). The earth is round (p<.05). American Psychologist, 49, Cortina, J. M., & Dunlap, W. P. (1997). Logic and purpose of significance testing. Psychological Methods, 2, Daniel, L. G. (1998). Statistical significance testing: A historical overview of misuse and mis-interpretation with implication for the editorial policies of educational journals. Research in the Schools, 5(2), Davis, J. A. (1971). Elementary survey analysis. Englewood Cliffs, NJ: Prentice-Hall. Fagley, N. S., & McKinney, I. J. (1983). Review bias for statistically significant results: A reexamination. Journal of Counseling Psychology, 30, Fan, X. (2001). Statistical significance and effect size in education research: Two sides of a coin. The Journal of Educational Research (Washington, D.C.), 94(5), Fisher, R. A. (1925). Statistical methods for research workers. London: Oliver & Boyd. Hair, J. F, Anderson, R. E., Tatham, R. L., & Black, W. C. (1995). Multivariate data analysis (4 th ed.). NJ: Prentice Hall. Hays, W. L. (1963). Statistics for psychologists. New York: Holt, Rinehart & Winston. Hinkle, D. E., Wiersma, W., & Jurs, S.G. (1979). Applied statistics for the behavioral sciences. Chicago: Rand McNally College Publishing. Hopkins, W. G. (1997). New view of statistics. Retrieved August 23, 2002 from Kirk, R. E. (1996). Practical significance: A concept whose time has come. Educational and Psychological Measurement, 56, Kirk, R. E. (2001). Promoting good statistical practices: Some suggestions. Educational and Psychological Measurement, 61(2), Maxwell, S. E., & Delaney, H. D. (1990). Designing experiments and analyzing data: A model comparison perspective. Belmont, CA: Wadsworth. Morse, D. T. (1998). MINSIZE: A computer program for obtaining minimum sample size as an indicator of effect size. Educational and Psychological Measurement, 58(1), Material published as part of this journal, either on-line or in print, is copyrighted by the Organizational Systems Research Association. Permission to make digital or paper copy of part or all of these works for personal or classroom use is granted without fee provided that the copies are not made or distributed for profit or commercial advantage AND that copies 1) bear this notice in full and 2) give the full citation. It is permissible to abstract these works so long as credit is given. To copy in all other cases or to republish or to post on a server or to redistribute to lists requires specific permission and payment of a fee. Contact Donna Everett, [email protected] to request redistribution permission. Nickerson, R. S. (2000). Null hypothesis significance testing: A review of an old and continuing controversy. Psychological Methods, 5, Olejnik, S., & Algina, J. (2000). Measures of effect size for comparative studies: Applications, interpretations, and limitations. Contemporary Educational Psychology, 25, Pearson, K. (1901). On the correlation of characters not quantitatively measurable. Philosophical Transactions of The Royal Society of London, 195, Rea, L. M., & Parker, R. A. (1992). Designing and conducting survey research. San Francisco: Jossey- Bass. Robinson, D. H., & Levin, J. R. (1997). Reflections on statistical and substantive significance, with a slice of replication. Educational Researcher, 26(5), Shaver, J. P. (1993). What statistical significance testing is and what it is not. The Journal of Experimental Education, 61, Thomas, J. R., Salazar, W., & Landers, D. M. (1991). What is missing in p<.05? Effect size. Research Quarterly for Exercise and Sport, 62(3), Thompson, B. (1996). AERA editorial polices regarding statistical significance testing: Three suggested reforms. Educational Researcher, 25(2), Thompson, B. (1998). In praise of brilliance: Where that praise really belongs. American Psychologist, 53, Thompson, B. (2002). What future quantitative social science research could look like: Confidence intervals for effect sizes. Educational Researcher, 31(3), Vacha-Haase, T. (2001). Statistical significance should not be considered one of life s guarantees: Effect sizes are needed. Educational and Psychological Measurement, 61(2), Wilkinson, L., & APA Task Force on Statistical Inference. (1999). Statistical methods in psychology journals: Guidelines and explanations. American Psychologist, 54, Williams, H. A. (2003). A mediated hierarchical regression analysis of factors related to research productivity of human resource development postsecondary faculty (Doctoral Dissertation, Louisiana State University, 2003). Available: / (Retrieved April 9, 2003).
Calculating, Interpreting, and Reporting Estimates of Effect Size (Magnitude of an Effect or the Strength of a Relationship)
1 Calculating, Interpreting, and Reporting Estimates of Effect Size (Magnitude of an Effect or the Strength of a Relationship) I. Authors should report effect sizes in the manuscript and tables when reporting
James E. Bartlett, II is Assistant Professor, Department of Business Education and Office Administration, Ball State University, Muncie, Indiana.
Organizational Research: Determining Appropriate Sample Size in Survey Research James E. Bartlett, II Joe W. Kotrlik Chadwick C. Higgins The determination of sample size is a common task for many organizational
UNDERSTANDING THE DEPENDENT-SAMPLES t TEST
UNDERSTANDING THE DEPENDENT-SAMPLES t TEST A dependent-samples t test (a.k.a. matched or paired-samples, matched-pairs, samples, or subjects, simple repeated-measures or within-groups, or correlated groups)
UNDERSTANDING THE INDEPENDENT-SAMPLES t TEST
UNDERSTANDING The independent-samples t test evaluates the difference between the means of two independent or unrelated groups. That is, we evaluate whether the means for two independent groups are significantly
Consider a study in which. How many subjects? The importance of sample size calculations. An insignificant effect: two possibilities.
Consider a study in which How many subjects? The importance of sample size calculations Office of Research Protections Brown Bag Series KB Boomer, Ph.D. Director, [email protected] A researcher conducts
Contrast Coding in Multiple Regression Analysis: Strengths, Weaknesses, and Utility of Popular Coding Structures
Journal of Data Science 8(2010), 61-73 Contrast Coding in Multiple Regression Analysis: Strengths, Weaknesses, and Utility of Popular Coding Structures Matthew J. Davis Texas A&M University Abstract: The
Additional sources Compilation of sources: http://lrs.ed.uiuc.edu/tseportal/datacollectionmethodologies/jin-tselink/tselink.htm
Mgt 540 Research Methods Data Analysis 1 Additional sources Compilation of sources: http://lrs.ed.uiuc.edu/tseportal/datacollectionmethodologies/jin-tselink/tselink.htm http://web.utk.edu/~dap/random/order/start.htm
Understanding and Quantifying EFFECT SIZES
Understanding and Quantifying EFFECT SIZES Karabi Nandy, Ph.d. Assistant Adjunct Professor Translational Sciences Section, School of Nursing Department of Biostatistics, School of Public Health, University
THE VALUE AND USEFULNESS OF INFORMATION TECHNOLOGY IN FAMILY AND CONSUMER SCIENCES EDUCATION AS PERCEIVED BY SECONDARY FACS TEACHERS
Journal of Family and Consumer Sciences Education, Vol. 18, No. 1, Spring/Summer 2000 THE VALUE AND USEFULNESS OF INFORMATION TECHNOLOGY IN FAMILY AND CONSUMER SCIENCES EDUCATION AS PERCEIVED BY SECONDARY
Eta Squared, Partial Eta Squared, and Misreporting of Effect Size in Communication Research
612 HUMAN COMMUNICATION RESEARCH /October 2002 Eta Squared, Partial Eta Squared, and Misreporting of Effect Size in Communication Research TIMOTHY R. LEVINE Michigan State University CRAIG R. HULLETT University
A Power Primer. Jacob Cohen New York University ABSTRACT
Psychological Bulletin July 1992 Vol. 112, No. 1, 155-159 1992 by the American Psychological Association For personal use only--not for distribution. A Power Primer Jacob Cohen New York University ABSTRACT
Statistics. Measurement. Scales of Measurement 7/18/2012
Statistics Measurement Measurement is defined as a set of rules for assigning numbers to represent objects, traits, attributes, or behaviors A variableis something that varies (eye color), a constant does
Using Excel for inferential statistics
FACT SHEET Using Excel for inferential statistics Introduction When you collect data, you expect a certain amount of variation, just caused by chance. A wide variety of statistical tests can be applied
UNDERSTANDING ANALYSIS OF COVARIANCE (ANCOVA)
UNDERSTANDING ANALYSIS OF COVARIANCE () In general, research is conducted for the purpose of explaining the effects of the independent variable on the dependent variable, and the purpose of research design
Descriptive Statistics
Descriptive Statistics Primer Descriptive statistics Central tendency Variation Relative position Relationships Calculating descriptive statistics Descriptive Statistics Purpose to describe or summarize
DATA ANALYSIS. QEM Network HBCU-UP Fundamentals of Education Research Workshop Gerunda B. Hughes, Ph.D. Howard University
DATA ANALYSIS QEM Network HBCU-UP Fundamentals of Education Research Workshop Gerunda B. Hughes, Ph.D. Howard University Quantitative Research What is Statistics? Statistics (as a subject) is the science
Study Guide for the Final Exam
Study Guide for the Final Exam When studying, remember that the computational portion of the exam will only involve new material (covered after the second midterm), that material from Exam 1 will make
TABLE OF CONTENTS. About Chi Squares... 1. What is a CHI SQUARE?... 1. Chi Squares... 1. Hypothesis Testing with Chi Squares... 2
About Chi Squares TABLE OF CONTENTS About Chi Squares... 1 What is a CHI SQUARE?... 1 Chi Squares... 1 Goodness of fit test (One-way χ 2 )... 1 Test of Independence (Two-way χ 2 )... 2 Hypothesis Testing
Sample Size and Power in Clinical Trials
Sample Size and Power in Clinical Trials Version 1.0 May 011 1. Power of a Test. Factors affecting Power 3. Required Sample Size RELATED ISSUES 1. Effect Size. Test Statistics 3. Variation 4. Significance
STATISTICS FOR PSYCHOLOGISTS
STATISTICS FOR PSYCHOLOGISTS SECTION: STATISTICAL METHODS CHAPTER: REPORTING STATISTICS Abstract: This chapter describes basic rules for presenting statistical results in APA style. All rules come from
II. DISTRIBUTIONS distribution normal distribution. standard scores
Appendix D Basic Measurement And Statistics The following information was developed by Steven Rothke, PhD, Department of Psychology, Rehabilitation Institute of Chicago (RIC) and expanded by Mary F. Schmidt,
WED 594. Advanced Research Methods SYLLABUS
WED 594 Spring 2016 Advanced Research Methods Thursdays 6:00-8:30 p.m. Pulliam Hall Rm. 204c SYLLABUS Instructor: C. Keith Waugh, Ph.D. Associate Professor and Director of WED Graduate Programs 219C Pulliam
A Performance Comparison of Native and Non-native Speakers of English on an English Language Proficiency Test ...
technical report A Performance Comparison of Native and Non-native Speakers of English on an English Language Proficiency Test.......... Agnes Stephenson, Ph.D. Hong Jiao Nathan Wall This paper was presented
ACT Research Explains New ACT Test Writing Scores and Their Relationship to Other Test Scores
ACT Research Explains New ACT Test Writing Scores and Their Relationship to Other Test Scores Wayne J. Camara, Dongmei Li, Deborah J. Harris, Benjamin Andrews, Qing Yi, and Yong He ACT Research Explains
A Hands-On Exercise Improves Understanding of the Standard Error. of the Mean. Robert S. Ryan. Kutztown University
A Hands-On Exercise 1 Running head: UNDERSTANDING THE STANDARD ERROR A Hands-On Exercise Improves Understanding of the Standard Error of the Mean Robert S. Ryan Kutztown University A Hands-On Exercise
Overview of Factor Analysis
Overview of Factor Analysis Jamie DeCoster Department of Psychology University of Alabama 348 Gordon Palmer Hall Box 870348 Tuscaloosa, AL 35487-0348 Phone: (205) 348-4431 Fax: (205) 348-8648 August 1,
UNDERSTANDING THE TWO-WAY ANOVA
UNDERSTANDING THE e have seen how the one-way ANOVA can be used to compare two or more sample means in studies involving a single independent variable. This can be extended to two independent variables
Assessment of Online Learning Environments: Using the OCLES(20) with Graduate Level Online Classes
www.ncolr.org/jiol Volume 7, Number 3, Winter 2008 ISSN: 15414914 Assessment of Online Learning Environments: Using the OCLES(20) with Graduate Level Online Classes Thomas A. DeVaney Nan B. Adams Cynthia
Learner Self-efficacy Beliefs in a Computer-intensive Asynchronous College Algebra Course
Learner Self-efficacy Beliefs in a Computer-intensive Asynchronous College Algebra Course Charles B. Hodges Georgia Southern University Department of Leadership, Technology, & Human Development P.O. Box
Statistics in Medicine Research Lecture Series CSMC Fall 2014
Catherine Bresee, MS Senior Biostatistician Biostatistics & Bioinformatics Research Institute Statistics in Medicine Research Lecture Series CSMC Fall 2014 Overview Review concept of statistical power
Correlational Research. Correlational Research. Stephen E. Brock, Ph.D., NCSP EDS 250. Descriptive Research 1. Correlational Research: Scatter Plots
Correlational Research Stephen E. Brock, Ph.D., NCSP California State University, Sacramento 1 Correlational Research A quantitative methodology used to determine whether, and to what degree, a relationship
Understanding Power and Rules of Thumb for Determining Sample Sizes
Tutorials in Quantitative Methods for Psychology 2007, vol. 3 (2), p. 43 50. Understanding Power and Rules of Thumb for Determining Sample Sizes Carmen R. Wilson VanVoorhis and Betsy L. Morgan University
" Y. Notation and Equations for Regression Lecture 11/4. Notation:
Notation: Notation and Equations for Regression Lecture 11/4 m: The number of predictor variables in a regression Xi: One of multiple predictor variables. The subscript i represents any number from 1 through
UNIVERSITY OF NAIROBI
UNIVERSITY OF NAIROBI MASTERS IN PROJECT PLANNING AND MANAGEMENT NAME: SARU CAROLYNN ELIZABETH REGISTRATION NO: L50/61646/2013 COURSE CODE: LDP 603 COURSE TITLE: RESEARCH METHODS LECTURER: GAKUU CHRISTOPHER
There are three kinds of people in the world those who are good at math and those who are not. PSY 511: Advanced Statistics for Psychological and Behavioral Research 1 Positive Views The record of a month
1/27/2013. PSY 512: Advanced Statistics for Psychological and Behavioral Research 2
PSY 512: Advanced Statistics for Psychological and Behavioral Research 2 Introduce moderated multiple regression Continuous predictor continuous predictor Continuous predictor categorical predictor Understand
Simple Linear Regression Inference
Simple Linear Regression Inference 1 Inference requirements The Normality assumption of the stochastic term e is needed for inference even if it is not a OLS requirement. Therefore we have: Interpretation
Trust, Job Satisfaction, Organizational Commitment, and the Volunteer s Psychological Contract
Trust, Job Satisfaction, Commitment, and the Volunteer s Psychological Contract Becky J. Starnes, Ph.D. Austin Peay State University Clarksville, Tennessee, USA [email protected] Abstract Studies indicate
Chapter 9. Two-Sample Tests. Effect Sizes and Power Paired t Test Calculation
Chapter 9 Two-Sample Tests Paired t Test (Correlated Groups t Test) Effect Sizes and Power Paired t Test Calculation Summary Independent t Test Chapter 9 Homework Power and Two-Sample Tests: Paired Versus
individualdifferences
1 Simple ANalysis Of Variance (ANOVA) Oftentimes we have more than two groups that we want to compare. The purpose of ANOVA is to allow us to compare group means from several independent samples. In general,
DATA COLLECTION AND ANALYSIS
DATA COLLECTION AND ANALYSIS Quality Education for Minorities (QEM) Network HBCU-UP Fundamentals of Education Research Workshop Gerunda B. Hughes, Ph.D. August 23, 2013 Objectives of the Discussion 2 Discuss
Quantitative research methods in business studies (doctoral course; 7,5 ECTS credits)
Quantitative research methods in business studies (doctoral course; 7,5 ECTS credits) Course director Mandar Dabhilkar, associate professor of operations management, email: [email protected] Learning objectives
Association Between Variables
Contents 11 Association Between Variables 767 11.1 Introduction............................ 767 11.1.1 Measure of Association................. 768 11.1.2 Chapter Summary.................... 769 11.2 Chi
Running head: SAMPLE FOR STUDENTS 1. Sample APA Paper for Students Interested in Learning APA Style 6th Edition. Jeffrey H. Kahn
Running head: SAMPLE FOR STUDENTS 1 Sample APA Paper for Students Interested in Learning APA Style 6th Edition Jeffrey H. Kahn Illinois State University Author Note Jeffrey H. Kahn, Department of Psychology,
Assessing a theoretical model on EFL college students
ABSTRACT Assessing a theoretical model on EFL college students Yu-Ping Chang Yu Da University, Taiwan This study aimed to (1) integrate relevant language learning models and theories, (2) construct a theoretical
HYPOTHESIS TESTING: CONFIDENCE INTERVALS, T-TESTS, ANOVAS, AND REGRESSION
HYPOTHESIS TESTING: CONFIDENCE INTERVALS, T-TESTS, ANOVAS, AND REGRESSION HOD 2990 10 November 2010 Lecture Background This is a lightning speed summary of introductory statistical methods for senior undergraduate
ANOVA ANOVA. Two-Way ANOVA. One-Way ANOVA. When to use ANOVA ANOVA. Analysis of Variance. Chapter 16. A procedure for comparing more than two groups
ANOVA ANOVA Analysis of Variance Chapter 6 A procedure for comparing more than two groups independent variable: smoking status non-smoking one pack a day > two packs a day dependent variable: number of
NORTHERN VIRGINIA COMMUNITY COLLEGE PSYCHOLOGY 211 - RESEARCH METHODOLOGY FOR THE BEHAVIORAL SCIENCES Dr. Rosalyn M.
NORTHERN VIRGINIA COMMUNITY COLLEGE PSYCHOLOGY 211 - RESEARCH METHODOLOGY FOR THE BEHAVIORAL SCIENCES Dr. Rosalyn M. King, Professor DETAILED TOPICAL OVERVIEW AND WORKING SYLLABUS CLASS 1: INTRODUCTIONS
Research Methods & Experimental Design
Research Methods & Experimental Design 16.422 Human Supervisory Control April 2004 Research Methods Qualitative vs. quantitative Understanding the relationship between objectives (research question) and
EMR 6410 Fundamentals of Measurement in the Behavioral Sciences. Course Pack Topics
EMR 6410 Fundamentals of Measurement in the Behavioral Sciences http://www.gowmu.wmich.edu http://homepages.wmich.edu/~applegab/courses/emr6410.html Course Pack Topics Course Catalogue Description 2 Course
ROCHESTER INSTITUTE OF TECHNOLOGY COURSE OUTLINE FORM COLLEGE OF SCIENCE. School of Mathematical Sciences
! ROCHESTER INSTITUTE OF TECHNOLOGY COURSE OUTLINE FORM COLLEGE OF SCIENCE School of Mathematical Sciences New Revised COURSE: COS-MATH-252 Probability and Statistics II 1.0 Course designations and approvals:
MASTER COURSE SYLLABUS-PROTOTYPE PSYCHOLOGY 2317 STATISTICAL METHODS FOR THE BEHAVIORAL SCIENCES
MASTER COURSE SYLLABUS-PROTOTYPE THE PSYCHOLOGY DEPARTMENT VALUES ACADEMIC FREEDOM AND THUS OFFERS THIS MASTER SYLLABUS-PROTOTYPE ONLY AS A GUIDE. THE INSTRUCTORS ARE FREE TO ADAPT THEIR COURSE SYLLABI
interpretation and implication of Keogh, Barnes, Joiner, and Littleton s paper Gender,
This essay critiques the theoretical perspectives, research design and analysis, and interpretation and implication of Keogh, Barnes, Joiner, and Littleton s paper Gender, Pair Composition and Computer
Analyzing Research Data Using Excel
Analyzing Research Data Using Excel Fraser Health Authority, 2012 The Fraser Health Authority ( FH ) authorizes the use, reproduction and/or modification of this publication for purposes other than commercial
Test Item Analysis & Decision Making Offered by the Measurement and Evaluation Center
Test Item Analysis & Decision Making Offered by the Measurement and Evaluation Center 1 Analyzing Multiple-Choice Item Responses Understanding how to interpret and use information based on student test
Evaluation in Online STEM Courses
Evaluation in Online STEM Courses Lawrence O. Flowers, PhD Assistant Professor of Microbiology James E. Raynor, Jr., PhD Associate Professor of Cellular and Molecular Biology Erin N. White, PhD Assistant
Predicting the Performance of a First Year Graduate Student
Predicting the Performance of a First Year Graduate Student Luís Francisco Aguiar Universidade do Minho - NIPE Abstract In this paper, I analyse, statistically, if GRE scores are a good predictor of the
The Effect of Static Visual Instruction on Students Online Learning: A Pilot Study
Interdisciplinary Journal of Information, Knowledge, and Management Volume 5, 2010 The Effect of Static Visual Instruction on Students Online Learning: A Pilot Study Pao-Nan Chou and Hsi-Chi Hsiao (The
Applications of Structural Equation Modeling in Social Sciences Research
American International Journal of Contemporary Research Vol. 4 No. 1; January 2014 Applications of Structural Equation Modeling in Social Sciences Research Jackson de Carvalho, PhD Assistant Professor
USING READING IN CONTENT AREA STRATEGIES TO IMPROVE STUDENT UNDERSTANDING IN FAMILY AND CONSUMER SCIENCES
Journal of Family and Consumer Sciences Education, Vol. 23, No. 2, Fall/Winter, 2005 USING READING IN CONTENT AREA STRATEGIES TO IMPROVE STUDENT UNDERSTANDING IN FAMILY AND CONSUMER SCIENCES Lynne Bell
CORRELATIONAL ANALYSIS: PEARSON S r Purpose of correlational analysis The purpose of performing a correlational analysis: To discover whether there
CORRELATIONAL ANALYSIS: PEARSON S r Purpose of correlational analysis The purpose of performing a correlational analysis: To discover whether there is a relationship between variables, To find out the
Consumer s Guide to Research on STEM Education. March 2012. Iris R. Weiss
Consumer s Guide to Research on STEM Education March 2012 Iris R. Weiss Consumer s Guide to Research on STEM Education was prepared with support from the National Science Foundation under grant number
Shiken: JLT Testing & Evlution SIG Newsletter. 5 (3) October 2001 (pp. 13-17)
Statistics Corner: Questions and answers about language testing statistics: Point- biserial correlation coefficients James Dean Brown (University of Hawai'i at Manoa) QUESTION: Recently on the email forum
THE UNIVERSITY OF TEXAS AT TYLER COLLEGE OF NURSING COURSE SYLLABUS NURS 5317 STATISTICS FOR HEALTH PROVIDERS. Fall 2013
THE UNIVERSITY OF TEXAS AT TYLER COLLEGE OF NURSING 1 COURSE SYLLABUS NURS 5317 STATISTICS FOR HEALTH PROVIDERS Fall 2013 & Danice B. Greer, Ph.D., RN, BC [email protected] Office BRB 1115 (903) 565-5766
research/scientific includes the following: statistical hypotheses: you have a null and alternative you accept one and reject the other
1 Hypothesis Testing Richard S. Balkin, Ph.D., LPC-S, NCC 2 Overview When we have questions about the effect of a treatment or intervention or wish to compare groups, we use hypothesis testing Parametric
DO PROGRAMS DESIGNED TO INCREASE EMOTIONAL INTELLIGENCE AT WORK WORK? Fabio Sala, Ph.D. Hay/McBer http://ei.haygroup.com. Methods
Consortium for Research on Emotional Intelligence in Organizations EI Program Eval 1 DO PROGRAMS DESIGNED TO INCREASE EMOTIONAL INTELLIGENCE AT WORK WORK? Fabio Sala, Ph.D. Hay/McBer http://ei.haygroup.com
The Relationship between Social Intelligence and Job Satisfaction among MA and BA Teachers
Kamla-Raj 2012 Int J Edu Sci, 4(3): 209-213 (2012) The Relationship between Social Intelligence and Job Satisfaction among MA and BA Teachers Soleiman Yahyazadeh-Jeloudar 1 and Fatemeh Lotfi-Goodarzi 2
Analysis of Data. Organizing Data Files in SPSS. Descriptive Statistics
Analysis of Data Claudia J. Stanny PSY 67 Research Design Organizing Data Files in SPSS All data for one subject entered on the same line Identification data Between-subjects manipulations: variable to
The Method of Least Squares
The Method of Least Squares Steven J. Miller Mathematics Department Brown University Providence, RI 0292 Abstract The Method of Least Squares is a procedure to determine the best fit line to data; the
Factorial Invariance in Student Ratings of Instruction
Factorial Invariance in Student Ratings of Instruction Isaac I. Bejar Educational Testing Service Kenneth O. Doyle University of Minnesota The factorial invariance of student ratings of instruction across
GMAC. Predicting Success in Graduate Management Doctoral Programs
GMAC Predicting Success in Graduate Management Doctoral Programs Kara O. Siegert GMAC Research Reports RR-07-10 July 12, 2007 Abstract An integral part of the test evaluation and improvement process involves
Section Format Day Begin End Building Rm# Instructor. 001 Lecture Tue 6:45 PM 8:40 PM Silver 401 Ballerini
NEW YORK UNIVERSITY ROBERT F. WAGNER GRADUATE SCHOOL OF PUBLIC SERVICE Course Syllabus Spring 2016 Statistical Methods for Public, Nonprofit, and Health Management Section Format Day Begin End Building
Introduction to. Hypothesis Testing CHAPTER LEARNING OBJECTIVES. 1 Identify the four steps of hypothesis testing.
Introduction to Hypothesis Testing CHAPTER 8 LEARNING OBJECTIVES After reading this chapter, you should be able to: 1 Identify the four steps of hypothesis testing. 2 Define null hypothesis, alternative
INTERPRETING THE ONE-WAY ANALYSIS OF VARIANCE (ANOVA)
INTERPRETING THE ONE-WAY ANALYSIS OF VARIANCE (ANOVA) As with other parametric statistics, we begin the one-way ANOVA with a test of the underlying assumptions. Our first assumption is the assumption of
The correlation coefficient
The correlation coefficient Clinical Biostatistics The correlation coefficient Martin Bland Correlation coefficients are used to measure the of the relationship or association between two quantitative
Homework 11. Part 1. Name: Score: / null
Name: Score: / Homework 11 Part 1 null 1 For which of the following correlations would the data points be clustered most closely around a straight line? A. r = 0.50 B. r = -0.80 C. r = 0.10 D. There is
The Data Analysis Dilemma: Ban or Abandon. A Review of Null Hypothesis Significance Testing
Copyright 1998 by the RESEARCH IN THE SCHOOLS Mid-South Educational Research Association 1998, Vol. 5, No. 2, 3-14 The Data Analysis Dilemma: Ban or Abandon. A Review of Null Hypothesis Significance Testing
The Usability of Electronic Stores based on the Organization of Information and Features
The Usability of Electronic Stores based on the Organization of Information and Features CHAN KAH-SING Singapore Polytechnic This paper describes an investigation on how the perceived usability of electronic
Section 13, Part 1 ANOVA. Analysis Of Variance
Section 13, Part 1 ANOVA Analysis Of Variance Course Overview So far in this course we ve covered: Descriptive statistics Summary statistics Tables and Graphs Probability Probability Rules Probability
Projects Involving Statistics (& SPSS)
Projects Involving Statistics (& SPSS) Academic Skills Advice Starting a project which involves using statistics can feel confusing as there seems to be many different things you can do (charts, graphs,
Introduction to Quantitative Methods
Introduction to Quantitative Methods October 15, 2009 Contents 1 Definition of Key Terms 2 2 Descriptive Statistics 3 2.1 Frequency Tables......................... 4 2.2 Measures of Central Tendencies.................
Moderation. Moderation
Stats - Moderation Moderation A moderator is a variable that specifies conditions under which a given predictor is related to an outcome. The moderator explains when a DV and IV are related. Moderation
Terminating Sequential Delphi Survey Data Collection
A peer-reviewed electronic journal. Copyright is retained by the first or sole author, who grants right of first publication to the Practical Assessment, Research & Evaluation. Permission is granted to
DOCTOR OF PHILOSOPHY DEGREE. Educational Leadership Doctor of Philosophy Degree Major Course Requirements. EDU721 (3.
DOCTOR OF PHILOSOPHY DEGREE Educational Leadership Doctor of Philosophy Degree Major Course Requirements EDU710 (3.0 credit hours) Ethical and Legal Issues in Education/Leadership This course is an intensive
School Psychology Doctoral Program Dissertation Outline 1 Final Version 6/2/2006
School Psychology Doctoral Program Dissertation Outline 1 Final Version 6/2/2006 Instructions: Double Underline means the item should be a title or heading in your dissertation. Do not deviate from the
The Effect of Varied Visual Scaffolds on Engineering Students Online Reading. Abstract. Introduction
Interdisciplinary Journal of E-Learning and Learning Objects Volume 6, 2010 The Effect of Varied Visual Scaffolds on Engineering Students Online Reading Pao-Nan Chou and Hsi-Chi Hsiao (The authors contributed
[This document contains corrections to a few typos that were found on the version available through the journal s web page]
Online supplement to Hayes, A. F., & Preacher, K. J. (2014). Statistical mediation analysis with a multicategorical independent variable. British Journal of Mathematical and Statistical Psychology, 67,
COMPARISONS OF CUSTOMER LOYALTY: PUBLIC & PRIVATE INSURANCE COMPANIES.
277 CHAPTER VI COMPARISONS OF CUSTOMER LOYALTY: PUBLIC & PRIVATE INSURANCE COMPANIES. This chapter contains a full discussion of customer loyalty comparisons between private and public insurance companies
California University of Pennsylvania Guidelines for New Course Proposals University Course Syllabus Approved: 2/4/13. Department of Psychology
California University of Pennsylvania Guidelines for New Course Proposals University Course Syllabus Approved: 2/4/13 Department of Psychology A. Protocol Course Name: Statistics and Research Methods in
Statistics Review PSY379
Statistics Review PSY379 Basic concepts Measurement scales Populations vs. samples Continuous vs. discrete variable Independent vs. dependent variable Descriptive vs. inferential stats Common analyses
Analysing Questionnaires using Minitab (for SPSS queries contact -) [email protected]
Analysing Questionnaires using Minitab (for SPSS queries contact -) [email protected] Structure As a starting point it is useful to consider a basic questionnaire as containing three main sections:
