For example, enter the following data in three COLUMNS in a new View window.


 Deirdre Rogers
 10 months ago
 Views:
Transcription
1 Statistics with Statview  18 Paired ttest A paired ttest compares two groups of measurements when the data in the two groups are in some way paired between the groups (e.g., before and after on the same individual or plot, measurements taken for each group in the same site or on the same day). When the variation among individuals, sites, or days is large compared to the variation between treatments, you should use a paired ttest. If your data do not have a paired structure, or if the variation among the data points within groups is not large, use a twosample ttest (it is more powerful). The paired ttest essentially checks whether the mean of the differences between paired measurements is significantly different from zero. To run the paired ttest: Have your data in separate columns such that rows contain paired observations from each group. Also construct a site or subject variable that is basically the row numbers. For example, enter the following data in three COLUMNS in a new View window. day kangaroo rats packrats Create a new View window, and in the Analysis browser, open Bivariate plots. Then doubleclick on Scattergram and a dialog box will appear. Click on OK. In the Variables browser click on day and then click X variable. Then drag through both rat variables and click Y variable. A graph will appear like the one below. Do the data look as though there is a consistent pattern within each pair but a lot of variance among pairs? Y Variables Bivariate Scattergram day kangaroo rats packrats Now doubleclick on Paired Comparisons in the Analysis browser. All the defaults are fine so click OK. If you still had the graph selected, three comparisons came up. Remove day. If the analysis didn t already appear, Add each species in the Variables browser. The output will show you the Pvalue in a brief table. In our example, the difference between kangaroo rat and packrat numbers is highly significant; there are on average 1.3 more packrats than kangaroo rats per plot.
2 Statistics with Statview  19 Paired ttest Hypothesized Difference = 0 Mean Diff. DF tvalue PValue kangaroo rats, packrats <.0001 It is always a good idea to try a twosample ttest on the data as well. That way you can be sure whether the use of the paired ttest was necessary. The means for the rat data don t differ significantly when tested with a twosample ttest, so the paired ttest was necessary. If you had run only the twosample ttest, you would have concluded that there wasn t a significant difference in numbers/plot between the two species. Unpaired ttest for density Grouping Variable: species Hypothesized Difference = 0 Mean Di DF tvalue PValue kangaroo rat, packrat Group Info for density Grouping Variable: species kangaroo rat packrat Cou Mean Variance Std. Dev. Std. Err MannWhitney U test (compares two groups when assumptions aren t met) A MannWhitney U test doesn t require normality or homogeneous variances, but it is somewhat less powerful than the ttest (which means the MannWhitney U test is less likely to find a significant difference between the two populations). So, if you have approximately normal data, then you should use a twosample ttest. The following example assumes your data for the two iguana populations are grouped under a compact variable named Density. To run a MannWhitney U test: Pull down the Analyze menu and choose new View window. In the analysis browser, double click on Nonparametrics. A dialog box will appear. Click on the button for Unpaired two group/ MannWhitney, then click OK. In the variable browser, select the continuous variable (e.g., Density), and then click Add. Then click on the arrow in front of the continuous variable to reveal the compact variable name (Isl...). Click on the compact variable name and then click Add. A table showing the results of the MannWhitney U test will be displayed. The output consists of a table showing the parameters used in the calculation of the test, but the main entry to examine is the Tied Pvalue. If P<0.05, then the medians of the two groups differ. In this example P<0.05 so the two medians are significantly different. The next table shows the Rank information for the test, which you can just ignore, unless you are a real stats nerd and want to know how it calculated the result.
3 Statistics with Statview  20 MannWhitney U for Density Grouping Variable: Islands U U Prime ZValue PValue Tied ZValue Tied PValue # Ties MannWhitney Rank Info for Density Grouping Variable: Islands Count Sum Ranks Mean Rank Island A Island B So again, this statistical test provides strong support for your original hypothesis that Island B has more iguanas than Island A. Oneway ANOVA and Posthoc paired comparisons (Fisher PLSD test) The oneway analysis of variance (ANOVA) is used when you are comparing the means of three or more groups and only one independent variable. It tests the significance of the difference among the means of the groups, and not which means are different from each other. In order to find out which means are significantly different from each other, you have to conduct Posthoc paired comparisons. They are called Posthoc, because you conduct the tests after you have completed an ANOVA and it shows that there is a significant difference among the means. One of the Posthoc tests is the Fisher PLSD (Protected Least Sig. Difference) test, which gives you a test of all pairwise combinations. To conduct these two tests your data can be in arranged in either of two ways: 1) data for each group can be in a separate column, and be under a compact variable heading, or 2) data can be in a single column with a second column containing codes to show which data values belong in each group. In our example, you would have a single column of density data and a column that has A, B, and C in it for each population. The following example assumes that you are using logtransformed data from the iguana population density data example above for Islands A, B, C. To run the ANOVA test: Pull down the Analyze menu and choose New view. In the analysis browser, click on arrow in front of ANOVA to reveal the choices below. Choose the ANOVA table and Means table, and click on Create Analysis. A dialog box will appear showing you options for the analysis. Pull down the bar at the bottom of box that says no error bars and choose % confidence interval, which will give the 95% confidence intervals for the data. Then click OK. In the variable browser, select the continuous variable (e.g., log Density), and then click dependent.
4 Statistics with Statview  21 Then click on the arrow in front of the continuous variable to reveal the compact variable name (Island). Click on the compact variable name and then click independent. An ANOVA table will be displayed along with a summary table of counts, means, std. dev. and std. errors. ANOVA Table for log Density DF Sum of Squares Mean Square FValue PValue Lambda Power Islands Residual Means Table for log Density Effect: Islands Island A Island B Island C Count Mean Std. Dev. Std. Err The next thing to do is interpret the results of the ANOVA. The ANOVA tested whether there were any differences in mean populations densities among the three populations. Look at the Pvalue. If this Pvalue is greater than 0.05, then there are no significant differences among any of the means. If the Pvalue is less than 0.05, then at least one mean is different from the others. In this example, P<0.05 in the ANOVA table, so the mean densities are significantly different. The second summary statistic table contains the counts, means, std. dev. and std. errors. Examine this information to make sure that it corresponds with the information that you have in the attribute pane in your original data set window. Now that you know the means are different, you want to find out which pairs of means are different from each other. e.g., is the density on Island A greater than B or C? To examine all pairwise comparisons of means, click on the arrow in front of Posthoc tests in the analysis window to reveal the choices below. Then doubleclick on Fisher PLSD (Protected Least Sig. Difference) to get a table of all pairwise comparisons among the means. A table will appear in the window on the right showing the mean difference (or difference between the means), the critical difference (or the mean difference required for P<0.05), and the Pvalue. Fisher's PLSD for log Density Effect: Islands Significance Level: 5 % Island A, Island B Island A, Island C Island B, Island C Mean Diff. Crit. Diff PValue S S The table shows that the mean population density of iguanas on Island A is significantly different from the mean density on Island B, but not Island C. Also, the density on Island B is significantly different from the density on Island C. Or more simply, Island B has more iguanas than either Island A or Island C, which don't differ. Plotting a graph (cell point chart) of these results might be helpful.
5 Statistics with Statview  22 Twoway ANOVA In many studies the data collected are associated with groups defined by more than one independent variable. For example, say you want to know whether the bill size of finches is different between individuals of different species and between males and females of the same species. These two 2 factors (Species and Sex) can be examined simultaneously in a Twoway ANOVA. Also, you can find out whether the independent variables have joint effects on the dependent variable (bill size), or whether they act independently of each other (i.e., does bill size depend on sex in one species but not in the other species?). Again, this test requires that the data are normally distributed and all of the groups have homogeneous variances. So you need to check these assumptions first. If you want to compare means from two (or more) grouping variables simultaneously, as ANOVA does, there is no satisfactory nonparametric alternative. So you may need to transform your data. Your data may be set up with the groups entered in separate columns, and then compacted (see p. 8). Alternatively, you can use the stacked format. You will choose the same variables in the Variables browser, but they will be presented somewhat differently. Before we run a twoway ANOVA, first do a ttest on bill size just between species, then do a t test on bill size just between sexes. Note the results. Do you think these results accurately represent the data? This exercise will show you how useful a twoway ANOVA can be in telling you more about the patterns in your data. Now run a twoway ANOVA on the same data. The procedure is much the same as for a Oneway ANOVA with one added step to include the second variable to the analysis. Pull down the Analyze menu and choose New View. In the analysis browser, click on arrow in front of ANOVA to reveal the choices below. Choose the ANOVA table and Means table, and click on Create Analysis. A dialog box will appear showing you options for the analysis. Click on the All effects button under Means tables and plots, and then pull down the bar where it says no error bars and choose % confidence interval, which will give the 95% confidence intervals for the data. Then click OK. In the variable browser, select the continuous variable (e.g., Bill size), and then click dependent. Then click on the arrow in front of the continuous variable to reveal the compact variable names. Click on the compact variable name Species and then click independent. Do the same for the variable name Sex. An ANOVA table will be displayed along with a summary table of counts, means, std. deviations and std. errors.
6 Statistics with Statview  23 ANOVA Table for Species DF Sum of Squares Mean Square FValue PValue Lambda Power Species Sex Species * Sex < Residual Means Table for Species Effect: Species Sp. A Sp. B Count Mean Std. Dev. Std. Err Means Table for Species Effect: Sex female male Sp. A, female Sp. A, male Sp. B, female Sp. B, male Count Mean Std. Dev. Std. Err. Means Table for Species Effect: Species * Sex Count Mean Std. Dev. Std. Err The output consists of an ANOVA table that shows the significance of each of the independent variables and the interaction. The Pvalue associated with each independent variable tells you the probability that the means of the different levels of that variable are the same. So, again, if p<0.05, the levels of that variable are different. The Pvalue of the interaction term tells you the probability that the two variables act independently of each other. In our bill size example, there is a significant difference in bill size between species but not between sexes. There is also a significant speciessex interaction. This indicates that the effect of sex differs between the species. To get a better impression of what the interaction term means, make a graph showing the interaction. In the Analysis browser, doubleclick on interaction line under the ANOVA heading to make an interaction line graph. Three graphs should appear, the first two showing the effects of each variable separately, and the last combining the effects of both variables into one graph. If a dialog box appears instead of a graph, you have to first make sure that one of the tables in the view is selected first before you doubleclick on interaction line.
7 Statistics with Statview  24 Cell Mean Interaction Line Plot for Bill size Effect: Species Error Bars: 95% Confidence Interval Sp. A Cell Mean Cell Sp. B Cell Mean Interaction Line Plot for Bill size Effect: Species * Sex Error Bars: 95% Confidence Interval Sp. A Cell Sp. B Interaction Line Plot for Bill size Effect: Sex Error Bars: 95% Confidence Interval female. female male Cell male To change symbols, click on the symbol in the key, then pull down the "Draw" menu. Choose "point" and pick the new symbol. If you look at the data, the interaction is apparent. In species A, the males have larger bills but in species B the females have larger bills. So simply looking at sex doesn t tell us anything (as you saw when you did the ttest) and neither sex has a consistently larger bill. When you did the ttest, you found that species was not significant, but the twoway ANOVA found that species was significant, over and above the interaction. This means that species A has larger bills overall. INDEPENDENT VARIABLE(S ALL) CONTINUOUS: CORRELATION & REGRESSION Correlation If the values of two CONTINUOUS variables appear to be related to one another, but neither causes the other, they are considered to be correlated. For example, the number of leaves falling in autumn and the number of geese flying south are generally correlated, but neither variable causes the other. The correlation coefficient, r, provides a quantitative measurement of how closely two variables are related. It ranges from 0 (no correlation) to 1 or 1 (the two variables are perfectly related, positively or negatively). Let s examine the correlation between bird weight and bill length, using the data displayed below. Bird #
8 Statistics with Statview bird weight (g) bill length (mm) Enter the data above in two COLUMNS in a new spreadsheet and name the columns Weight and Length. To visualize what the correlation represents, make a plot of the data: Create a new View window, and in the Analysis browser click on the arrow in front of Bivariate plots. Then doubleclick on Scattergram and a dialog box will appear. Click on OK. In the Variables browser click on the variable Weight and then click X variable. Then click on the variable Length and click Y variable. A graph will appear as below. 18 Bivariate Scattergram 16 Length Weight From this plot you can see that as weight increases there is also an increase in bird bill length. Thus, these two variables appear to be correlated. To quantify the extent of the correlation and see if it is statistically significant: Go to the View window and doubleclick on Correlation/Covar... in the Analysis browser. This will open a dialog box titled Correlation/Covariance. Click on the button for Fisher s r to z (pvalues) and then click OK. A box will appear in the view window. In the Variables browser click on the variable Weight and then click Add (not necessary if you still have the graph selected). Then click on the variable Length and click Add. A table will appear as shown on the next page.
9 Statistics with Statview  26 Correlation Matrix Weight Length Weight Length observations were used in this computation. Fisher's r to z Correlation PValue Weight, Length observations were used in this computation. The value in the table is the Correlation matrix table is the Pearson correlation coefficient (0.66), which shows there is a positive correlation between Weight and Length. (A negative correlation would have a negative Pearson correlation coefficient.) The results of the Fisher s test shows that it is a statistically significant correlation (i.e., P<0.05). Regression Regressions and correlations are both used to test whether two variables are related to each other, and if so, how closely they are related. However, whereas correlation demands NO causal link, regression does. With regression we can determine if a change in one variable can be predicted from the change in another variable. A simple example is a onetoone relationship between two variables, such as the relationship between the age and the number of growth rings of a tree. Another example is the relationship between the age and length of a fish. Age (years) Length (cm) The data consist of a value for the independent variable (x) and the associated value for the dependent variable (y). Think of these as on an xaxis and a yaxis. In our example, given x (the age of the fish) one can predict y (length). Generally, the independent variable (x) is controlled or standardized by the investigator, and the y variable is dependent on the value of x. A regression calculates the equation of the best fitting straight line through the (x,y) points that the data pairs define. In the equation of a line (y = a + bx), a is the yintercept (where x=0) and b is the slope. The output of a regression will give you estimates for both of these values. If we wanted to predict the length of a fish at a given age, we could do so using the regression equation that best fits these data. Enter the data above into a new spreadsheet and name the two data columns Age and Length. To visualize the relationship between these two variables, make a plot of the data: Create a new View window, then in the Analysis browser click on the arrow in front of Regression.
10 Statistics with Statview  27 Then doubleclick on Regression Plot and a dialog box will appear. Click OK, and a new box will appear in the window on the right. In the Variables browser, click on the variable Age and then click Independent variable. Then click on the variable Length and click Dependent variable. A graph will appear as below. Length Bivariate Scattergram with Regression Age Length = * Age; R^2 =.969 The graph shows that there is a strong positive relationship between fish Age and Length. The equation below the graph is the equation for the regression line that best describes the relationship between the two variables. The R^2 (or R 2 ) value is the coefficient of determination, and can be interpreted as the proportion of the variation in the data that is explained by the factors you have used in the analysis. R 2 ranges from 0 to 1. If it is close to 1, it means that your independent variable has explained almost all of why the value of your dependent variable differs from observation to observation. If R 2 is close to 0, it means that they have explained almost none of the variation in your dependent variable. In this example it appears that 97% of the variation in the variation in Length is explained by variation in Age. Now what you want to do is determine whether the relationship is statistically significant. To run a regression analysis: In the Analysis browser click on the arrow in front of Regression. Then select Regression sum..., ANOVA table, and Regression coef..., and click Create Analysis. A summary table should appear along with an ANOVA table and a coefficient table. If it does not appear then you should select your variables for the analysis in the Variables browser. The regression summary provides the basic data for the analysis along with the R 2 values. The results of the ANOVA table indicate whether the relationship between the two variables is significant. P<0.05.The output of the Regression Coefficient table consists of the estimates (which it calls coefficients) of yintercept and the slope (which is your independent variable). Each of these has an associated Pvalue shown at the far right. If the Pvalue for the yintercept is less than 0.05, then it is significantly different from zero. If the Pvalue for the slope is less than 0.05, then the slope is significantly different from zero. Examine the output for the regression analysis of the example data on the next page.
11 Statistics with Statview  28 Regression Summary Length vs. Age Count Num. Missing R R Squared Adjusted R Squared RMS Residual ANOVA Table Length vs. Age DF Sum of Squares Mean Square FValue PValue Regression <.0001 Residual Total Regression Coefficients Length vs. Age Coefficient Std. Error Std. Coeff. tvalue PValue Intercept <.0001 Age <.0001 From the output, we can see the very high R 2 value reveals that 97% of the variation in length (dependent variable) can be explained by variation in age (independent variable). The low pvalue (<0.05) in the ANOVA table indicates that the relationship is highly significant, and thus very unlikely to occur by chance alone. The output also indicates that the yintercept and the slope are significantly different from zero. Analysis of Covariance (ANCOVA) ANCOVAs are hybrids between regressions and ANOVAs because they combine both continuous ("covariate") and categorical INDEPENDENT variables. The most common case in ecology is when you want to compare the slopes and/or intercepts of regression lines. In this case, your data will consist of a categorical variable that separates two or more sets of pairs of points that define separate regression lines. For ANCOVA analysis, your data must be "stacked" (one column for the continuous independent variable, one column for the categorical independent variable, and one column for the response or dependent variable). For example, assume we want to compare the growth rates of fish in warm and cool tanks. We will compare the regressions of length on age from the two temperatures, where age is our continuous independent variable (our covariate), temperature is our categorical independent variable, and length is our dependent or response variable. If you go back to the data set you had previously on fish age and length, add seven more rows and one column as shown on the next page. Age (years) Temperature Length (cm)
12 Statistics with Statview warm warm warm warm warm warm warm cool cool cool cool cool cool cool 20.4 First, graph these data to understand the relationships among the variables. Create a new View window, and in the Analysis browser, click on the arrow in front of Bivariate Plot. Doubleclick on Scattergram. In the dialog box, click the button for Regression with a... (don't worry about "means" or "slopes"), and under "When split, show lines for," choose "each group separately." Click OK. In the Variables browser, click on Age and then click independent variable. Then click on Length and click dependent variable. A graph of both data sets will appear with the combined regression line shown. Finally, click on Temperature and click split by. The graph will sort into the two separate data sets and show each regression line individually. The individual regression equations are shown below the graph. Length Bivariate Scattergram with Regression Split By: Temperature Age Length = * Age; R^2 =.969 (cool) Length = * Age; R^2 =.996 (warm) cool warm To run the ANCOVA analysis to see whether the slopes or the intercepts of these two lines differ:
13 Statistics with Statview  30 In the Analysis browser, click on the arrow in front of ANOVA, doubleclick on ANOVA Table, and click OK in the dialog box to accept the default parameters. Analysis will appear for each regression line separately, but you want to compare the two lines in the same analysis. In the Variables browser, click on Temperature and then click on Remove (to recombine the date from the separate regressions) and then click on Independent (to include the variable in the analysis as an independent variable). The new analysis table will have a row for Age (the covariate or Xaxis variable), Temperature (the categorical variable), the interaction between these two, and the residual or error term. ANOVA Table for Length DF Sum of Squares Mean Square FValue PValue Lambda Power Age < Temperature Age * Temperature Residual There are four possible types of outcomes in an ANCOVA analysis (see Figure 1 next page). To interpret the ANCOVA output, look at the interaction term first: If the interaction is significant (P<0.05), then the two slopes are different. If the two slopes are different, then the significance of the main effects for the categorical and continuous variables may not be very useful to interpret. In this case the continuous variable Pvalue is combining the data for the two regression lines and checking to see if the slope of that composite line is significant. It may or may not be significant, depending on whether the slopes of the separate lines are in a similar direction. The interpretation of the categorical variable is even harder. If the slopes are different, the categorical Pvalue tells you whether the mean Yvalue of one line is different from the mean Yvalue of the other. But since the slopes differ, this is pretty meaningless. If the interaction term is not significant (P>0.05), then the slopes are not different and both of the main effects of the continuous and categorical variables are easy to interpret. The continuous variable Pvalue tells you if the composite slope is different from zero, and the categorical variable Pvalue tells you whether the two Yintercepts differ from each other (i.e., whether the two lines are coincident or parallel). In our example, the interaction term (age*temperature) is significant (P=0.0051) so the slopes of the two lines differ. That makes it more difficult to directly assess what else is going on. To sort this out, look at the graph you constructed. Given the slopes and intercepts we saw in the regressions, the outcome is most similar to Case 4. Both lines have positive slopes (which is why the composite slope tested by the "age" main effect is significant (p<0.0001)). The two lines also apparently have different yvalues a the mean value of x (temperature main effect p=0.0310), but of course, how different the yvalues are depends on where you look along the axis.
14 Statistics with Statview  31 Figure 1. Possible outcomes of an ANCOVA. y Case 1 y Case 2 x variable (covariate) x variable (covariate) y Case 3 y Case 4 x variable (covariate) x variable (covariate) In Case 1, both the slopes and the intercepts are equal. This is the outcome you might expect if your treatment variable (temperature) had no effect on either the relationship of the X and Y variables (the effect of age on length) or the value of Y when X is 0 (the length of newlyhatched fish). In Case 2, the slopes are the same but the intercepts differ. This is the outcome you might expect if the relationship between X and Y (the effect of age on length) was the same for both treatments, but the values of Y were uniformly larger under one treatment or another (fish are generally bigger at a given age in warm water). In Case 3, the intercepts are the same but the slopes differ. This is the outcome you might expect if the treatments start the same but the treatment causes the relationship to become more different as the value of the covariate increases (fish hatch at the same size, but grow faster in warm water than in cool water). In Case 4, both the intercepts and the slopes differ. This is the outcome you might expect if the treatment causes changes in both the relationship of the X and Y variables (the effect of age on length) and the value of X when Y is 0 (the length of newlyhatched fish).
Using SPSS version 14 Joel Elliott, Jennifer Burnaford, Stacey Weiss
Using SPSS version 14 Joel Elliott, Jennifer Burnaford, Stacey Weiss SPSS is a program that is very easy to learn and is also very powerful. This manual is designed to introduce you to the program however,
More informationA Guide for a Selection of SPSS Functions
A Guide for a Selection of SPSS Functions IBM SPSS Statistics 19 Compiled by Beth Gaedy, Math Specialist, Viterbo University  2012 Using documents prepared by Drs. Sheldon Lee, Marcus Saegrove, Jennifer
More informationSPSS Tests for Versions 9 to 13
SPSS Tests for Versions 9 to 13 Chapter 2 Descriptive Statistic (including median) Choose Analyze Descriptive statistics Frequencies... Click on variable(s) then press to move to into Variable(s): list
More informationSimple Linear Regression, Scatterplots, and Bivariate Correlation
1 Simple Linear Regression, Scatterplots, and Bivariate Correlation This section covers procedures for testing the association between two continuous variables using the SPSS Regression and Correlate analyses.
More informationDirections for using SPSS
Directions for using SPSS Table of Contents Connecting and Working with Files 1. Accessing SPSS... 2 2. Transferring Files to N:\drive or your computer... 3 3. Importing Data from Another File Format...
More informationModule 9: Nonparametric Tests. The Applied Research Center
Module 9: Nonparametric Tests The Applied Research Center Module 9 Overview } Nonparametric Tests } Parametric vs. Nonparametric Tests } Restrictions of Nonparametric Tests } OneSample ChiSquare Test
More informationChapter 5 Analysis of variance SPSS Analysis of variance
Chapter 5 Analysis of variance SPSS Analysis of variance Data file used: gss.sav How to get there: Analyze Compare Means Oneway ANOVA To test the null hypothesis that several population means are equal,
More informationSimple Linear Regression in SPSS STAT 314
Simple Linear Regression in SPSS STAT 314 1. Ten Corvettes between 1 and 6 years old were randomly selected from last year s sales records in Virginia Beach, Virginia. The following data were obtained,
More informationRegression III: Dummy Variable Regression
Regression III: Dummy Variable Regression Tom Ilvento FREC 408 Linear Regression Assumptions about the error term Mean of Probability Distribution of the Error term is zero Probability Distribution of
More informationThe scatterplot indicates a positive linear relationship between waist size and body fat percentage:
STAT E150 Statistical Methods Multiple Regression Three percent of a man's body is essential fat, which is necessary for a healthy body. However, too much body fat can be dangerous. For men between the
More informationSPSS Guide: Tests of Differences
SPSS Guide: Tests of Differences I put this together to give you a stepbystep guide for replicating what we did in the computer lab. It should help you run the tests we covered. The best way to get familiar
More information6 Comparison of differences between 2 groups: Student s Ttest, MannWhitney UTest, Paired Samples Ttest and Wilcoxon Test
6 Comparison of differences between 2 groups: Student s Ttest, MannWhitney UTest, Paired Samples Ttest and Wilcoxon Test Having finally arrived at the bottom of our decision tree, we are now going
More informationAnalysis of Covariance
Analysis of Covariance 1. Introduction The Analysis of Covariance (generally known as ANCOVA) is a technique that sits between analysis of variance and regression analysis. It has a number of purposes
More informationBasic Data Analysis Using JMP in Windows Table of Contents:
Basic Data Analysis Using JMP in Windows Table of Contents: I. Getting Started with JMP II. Entering Data in JMP III. Saving JMP Data file IV. Opening an Existing Data File V. Transforming and Manipulating
More informationSPSS Explore procedure
SPSS Explore procedure One useful function in SPSS is the Explore procedure, which will produce histograms, boxplots, stemandleaf plots and extensive descriptive statistics. To run the Explore procedure,
More informationData analysis. Data analysis in Excel using Windows 7/Office 2010
Data analysis Data analysis in Excel using Windows 7/Office 2010 Open the Data tab in Excel If Data Analysis is not visible along the top toolbar then do the following: o Right click anywhere on the toolbar
More informationSPSS: Descriptive and Inferential Statistics. For Windows
For Windows August 2012 Table of Contents Section 1: Summarizing Data...3 1.1 Descriptive Statistics...3 Section 2: Inferential Statistics... 10 2.1 ChiSquare Test... 10 2.2 T tests... 11 2.3 Correlation...
More informationAllelopathic Effects on Root and Shoot Growth: OneWay Analysis of Variance (ANOVA) in SPSS. Dan Flynn
Allelopathic Effects on Root and Shoot Growth: OneWay Analysis of Variance (ANOVA) in SPSS Dan Flynn Just as ttests are useful for asking whether the means of two groups are different, analysis of variance
More informationData Analysis Tools. Tools for Summarizing Data
Data Analysis Tools This section of the notes is meant to introduce you to many of the tools that are provided by Excel under the Tools/Data Analysis menu item. If your computer does not have that tool
More informationID X Y
Dale Berger SPSS StepbyStep Regression Introduction: MRC01 This stepbystep example shows how to enter data into SPSS and conduct a simple regression analysis to develop an equation to predict from.
More informationSPSS for Exploratory Data Analysis Data used in this guide: studentp.sav (http://people.ysu.edu/~gchang/stat/studentp.sav)
Data used in this guide: studentp.sav (http://people.ysu.edu/~gchang/stat/studentp.sav) Organize and Display One Quantitative Variable (Descriptive Statistics, Boxplot & Histogram) 1. Move the mouse pointer
More informationProjects Involving Statistics (& SPSS)
Projects Involving Statistics (& SPSS) Academic Skills Advice Starting a project which involves using statistics can feel confusing as there seems to be many different things you can do (charts, graphs,
More informationAn introduction to IBM SPSS Statistics
An introduction to IBM SPSS Statistics Contents 1 Introduction... 1 2 Entering your data... 2 3 Preparing your data for analysis... 10 4 Exploring your data: univariate analysis... 14 5 Generating descriptive
More informationCHAPTER 2 AND 10: Least Squares Regression
CHAPTER 2 AND 0: Least Squares Regression In chapter 2 and 0 we will be looking at the relationship between two quantitative variables measured on the same individual. General Procedure:. Make a scatterplot
More informationThe Dummy s Guide to Data Analysis Using SPSS
The Dummy s Guide to Data Analysis Using SPSS Mathematics 57 Scripps College Amy Gamble April, 2001 Amy Gamble 4/30/01 All Rights Rerserved TABLE OF CONTENTS PAGE Helpful Hints for All Tests...1 Tests
More informationIntroduction to Regression. Dr. Tom Pierce Radford University
Introduction to Regression Dr. Tom Pierce Radford University In the chapter on correlational techniques we focused on the Pearson R as a tool for learning about the relationship between two variables.
More informationLinear Models in STATA and ANOVA
Session 4 Linear Models in STATA and ANOVA Page Strengths of Linear Relationships 42 A Note on NonLinear Relationships 44 Multiple Linear Regression 45 Removal of Variables 48 Independent Samples
More informationOneWay ANOVA using SPSS 11.0. SPSS ANOVA procedures found in the Compare Means analyses. Specifically, we demonstrate
1 OneWay ANOVA using SPSS 11.0 This section covers steps for testing the difference between three or more group means using the SPSS ANOVA procedures found in the Compare Means analyses. Specifically,
More informationUsing CrunchIt (http://bcs.whfreeman.com/crunchit/bps4e) or StatCrunch (www.calvin.edu/go/statcrunch)
Using CrunchIt (http://bcs.whfreeman.com/crunchit/bps4e) or StatCrunch (www.calvin.edu/go/statcrunch) 1. In general, this package is far easier to use than many statistical packages. Every so often, however,
More informationAn SPSS companion book. Basic Practice of Statistics
An SPSS companion book to Basic Practice of Statistics SPSS is owned by IBM. 6 th Edition. Basic Practice of Statistics 6 th Edition by David S. Moore, William I. Notz, Michael A. Flinger. Published by
More informationPractice 3 SPSS. Partially based on Notes from the University of Reading:
Practice 3 SPSS Partially based on Notes from the University of Reading: http://www.reading.ac.uk Simple Linear Regression A simple linear regression model is fitted when you want to investigate whether
More informationStatistical Significance and Bivariate Tests
Statistical Significance and Bivariate Tests BUS 735: Business Decision Making and Research 1 1.1 Goals Goals Specific goals: Refamiliarize ourselves with basic statistics ideas: sampling distributions,
More informationSimple Linear Regression
1 Excel Manual Simple Linear Regression Chapter 13 This chapter discusses statistics involving the linear regression. Excel has numerous features that work well for comparing quantitative variables both
More informationMinitab Guide. This packet contains: A Friendly Guide to Minitab. Minitab StepByStep
Minitab Guide This packet contains: A Friendly Guide to Minitab An introduction to Minitab; including basic Minitab functions, how to create sets of data, and how to create and edit graphs of different
More informationANNOTATED OUTPUTSPSS Simple Linear (OLS) Regression
Simple Linear (OLS) Regression Regression is a method for studying the relationship of a dependent variable and one or more independent variables. Simple Linear Regression tells you the amount of variance
More informationDoing Multiple Regression with SPSS. In this case, we are interested in the Analyze options so we choose that menu. If gives us a number of choices:
Doing Multiple Regression with SPSS Multiple Regression for Data Already in Data Editor Next we want to specify a multiple regression analysis for these data. The menu bar for SPSS offers several options:
More informationStatistics and research
Statistics and research Usaneya Perngparn Chitlada Areesantichai Drug Dependence Research Center (WHOCC for Research and Training in Drug Dependence) College of Public Health Sciences Chulolongkorn University,
More information496 STATISTICAL ANALYSIS OF CAUSE AND EFFECT
496 STATISTICAL ANALYSIS OF CAUSE AND EFFECT * Use a nonparametric technique. There are statistical methods, called nonparametric methods, that don t make any assumptions about the underlying distribution
More informationTechnology StepbyStep Using StatCrunch
Technology StepbyStep Using StatCrunch Section 1.3 Simple Random Sampling 1. Select Data, highlight Simulate Data, then highlight Discrete Uniform. 2. Fill in the following window with the appropriate
More informationHow To Run Statistical Tests in Excel
How To Run Statistical Tests in Excel Microsoft Excel is your best tool for storing and manipulating data, calculating basic descriptive statistics such as means and standard deviations, and conducting
More informationMain Effects and Interactions
Main Effects & Interactions page 1 Main Effects and Interactions So far, we ve talked about studies in which there is just one independent variable, such as violence of television program. You might randomly
More informationWe are often interested in the relationship between two variables. Do people with more years of fulltime education earn higher salaries?
Statistics: Correlation Richard Buxton. 2008. 1 Introduction We are often interested in the relationship between two variables. Do people with more years of fulltime education earn higher salaries? Do
More informationMultiple Regression Analysis in Minitab 1
Multiple Regression Analysis in Minitab 1 Suppose we are interested in how the exercise and body mass index affect the blood pressure. A random sample of 10 males 50 years of age is selected and their
More informationTesting Hypotheses using SPSS
Is the mean hourly rate of male workers $2.00? TTest OneSample Statistics Std. Error N Mean Std. Deviation Mean 2997 2.0522 6.6282.2 OneSample Test Test Value = 2 95% Confidence Interval Mean of the
More informationJanuary 26, 2009 The Faculty Center for Teaching and Learning
THE BASICS OF DATA MANAGEMENT AND ANALYSIS A USER GUIDE January 26, 2009 The Faculty Center for Teaching and Learning THE BASICS OF DATA MANAGEMENT AND ANALYSIS Table of Contents Table of Contents... i
More informationLesson Lesson Outline Outline
Lesson 15 Linear Regression Lesson 15 Outline Review correlation analysis Dependent and Independent variables Least Squares Regression line Calculating l the slope Calculating the Intercept Residuals and
More informationMultivariate analysis of variance
21 Multivariate analysis of variance In previous chapters, we explored the use of analysis of variance to compare groups on a single dependent variable. In many research situations, however, we are interested
More informationEXCEL Analysis TookPak [Statistical Analysis] 1. First of all, check to make sure that the Analysis ToolPak is installed. Here is how you do it:
EXCEL Analysis TookPak [Statistical Analysis] 1 First of all, check to make sure that the Analysis ToolPak is installed. Here is how you do it: a. From the Tools menu, choose AddIns b. Make sure Analysis
More informationGuide for SPSS for Windows
Guide for SPSS for Windows Index Table Open an existing data file Open a new data sheet Enter or change data value Name a variable Label variables and data values Enter a categorical data Delete a record
More informationSPSS Guide: Regression Analysis
SPSS Guide: Regression Analysis I put this together to give you a stepbystep guide for replicating what we did in the computer lab. It should help you run the tests we covered. The best way to get familiar
More informationLecture  32 Regression Modelling Using SPSS
Applied Multivariate Statistical Modelling Prof. J. Maiti Department of Industrial Engineering and Management Indian Institute of Technology, Kharagpur Lecture  32 Regression Modelling Using SPSS (Refer
More information0.1 Multiple Regression Models
0.1 Multiple Regression Models We will introduce the multiple Regression model as a mean of relating one numerical response variable y to two or more independent (or predictor variables. We will see different
More informationOnce saved, if the file was zipped you will need to unzip it.
1 Commands in SPSS 1.1 Dowloading data from the web The data I post on my webpage will be either in a zipped directory containing a few files or just in one file containing data. Please learn how to unzip
More informationCalibration and Linear Regression Analysis: A SelfGuided Tutorial
Calibration and Linear Regression Analysis: A SelfGuided Tutorial Part 1 Instrumental Analysis with Excel: The Basics CHM314 Instrumental Analysis Department of Chemistry, University of Toronto Dr. D.
More informationSCHOOL OF HEALTH AND HUMAN SCIENCES DON T FORGET TO RECODE YOUR MISSING VALUES
SCHOOL OF HEALTH AND HUMAN SCIENCES Using SPSS Topics addressed today: 1. Differences between groups 2. Graphing Use the s4data.sav file for the first part of this session. DON T FORGET TO RECODE YOUR
More informationSimple Linear Regression One Binary Categorical Independent Variable
Simple Linear Regression Does sex influence mean GCSE score? In order to answer the question posed above, we want to run a linear regression of sgcseptsnew against sgender, which is a binary categorical
More informationChapter 23. Inferences for Regression
Chapter 23. Inferences for Regression Topics covered in this chapter: Simple Linear Regression Simple Linear Regression Example 23.1: Crying and IQ The Problem: Infants who cry easily may be more easily
More informationChapter 16 Appendix. Nonparametric Tests with Excel, JMP, Minitab, SPSS, CrunchIt!, R, and TI83/84 Calculators
The Wilcoxon Rank Sum Test Chapter 16 Appendix Nonparametric Tests with Excel, JMP, Minitab, SPSS, CrunchIt!, R, and TI83/84 Calculators These nonparametric tests make no assumption about Normality.
More informationSydney Roberts Predicting Age Group Swimmers 50 Freestyle Time 1. 1. Introduction p. 2. 2. Statistical Methods Used p. 5. 3. 10 and under Males p.
Sydney Roberts Predicting Age Group Swimmers 50 Freestyle Time 1 Table of Contents 1. Introduction p. 2 2. Statistical Methods Used p. 5 3. 10 and under Males p. 8 4. 11 and up Males p. 10 5. 10 and under
More informationFormula for linear models. Prediction, extrapolation, significance test against zero slope.
Formula for linear models. Prediction, extrapolation, significance test against zero slope. Last time, we looked the linear regression formula. It s the line that fits the data best. The Pearson correlation
More information7. Tests of association and Linear Regression
7. Tests of association and Linear Regression In this chapter we consider 1. Tests of Association for 2 qualitative variables. 2. Measures of the strength of linear association between 2 quantitative variables.
More informationHYPOTHESIS TESTING: CONFIDENCE INTERVALS, TTESTS, ANOVAS, AND REGRESSION
HYPOTHESIS TESTING: CONFIDENCE INTERVALS, TTESTS, ANOVAS, AND REGRESSION HOD 2990 10 November 2010 Lecture Background This is a lightning speed summary of introductory statistical methods for senior undergraduate
More informationUsing Excel for inferential statistics
FACT SHEET Using Excel for inferential statistics Introduction When you collect data, you expect a certain amount of variation, just caused by chance. A wide variety of statistical tests can be applied
More informationSimple Linear Regression Chapter 11
Simple Linear Regression Chapter 11 Rationale Frequently decisionmaking situations require modeling of relationships among business variables. For instance, the amount of sale of a product may be related
More informationBill Burton Albert Einstein College of Medicine william.burton@einstein.yu.edu April 28, 2014 EERS: Managing the Tension Between Rigor and Resources 1
Bill Burton Albert Einstein College of Medicine william.burton@einstein.yu.edu April 28, 2014 EERS: Managing the Tension Between Rigor and Resources 1 Calculate counts, means, and standard deviations Produce
More informationUCLA STAT 13 Statistical Methods  Final Exam Review Solutions Chapter 7 Sampling Distributions of Estimates
UCLA STAT 13 Statistical Methods  Final Exam Review Solutions Chapter 7 Sampling Distributions of Estimates 1. (a) (i) µ µ (ii) σ σ n is exactly Normally distributed. (c) (i) is approximately Normally
More informationABSORBENCY OF PAPER TOWELS
ABSORBENCY OF PAPER TOWELS 15. Brief Version of the Case Study 15.1 Problem Formulation 15.2 Selection of Factors 15.3 Obtaining Random Samples of Paper Towels 15.4 How will the Absorbency be measured?
More informationSimple linear regression
Simple linear regression Introduction Simple linear regression is a statistical method for obtaining a formula to predict values of one variable from another where there is a causal relationship between
More informationWhen to use Excel. When NOT to use Excel 9/24/2014
Analyzing Quantitative Assessment Data with Excel October 2, 2014 Jeremy Penn, Ph.D. Director When to use Excel You want to quickly summarize or analyze your assessment data You want to create basic visual
More informationEXPERIMENT 6: HERITABILITY AND REGRESSION
BIO 184 Laboratory Manual Page 74 EXPERIMENT 6: HERITABILITY AND REGRESSION DAY ONE: INTRODUCTION TO HERITABILITY AND REGRESSION OBJECTIVES: Today you will be learning about some of the basic ideas and
More informationUsing SPSS, Chapter 2: Descriptive Statistics
1 Using SPSS, Chapter 2: Descriptive Statistics Chapters 2.1 & 2.2 Descriptive Statistics 2 Mean, Standard Deviation, Variance, Range, Minimum, Maximum 2 Mean, Median, Mode, Standard Deviation, Variance,
More informationEXCEL EXERCISE AND ACCELERATION DUE TO GRAVITY
EXCEL EXERCISE AND ACCELERATION DUE TO GRAVITY Objective: To learn how to use the Excel spreadsheet to record your data, calculate values and make graphs. To analyze the data from the Acceleration Due
More information1. What is the critical value for this 95% confidence interval? CV = z.025 = invnorm(0.025) = 1.96
1 Final Review 2 Review 2.1 CI 1propZint Scenario 1 A TV manufacturer claims in its warranty brochure that in the past not more than 10 percent of its TV sets needed any repair during the first two years
More informationDEPARTMENT OF HEALTH AND HUMAN SCIENCES HS900 RESEARCH METHODS
DEPARTMENT OF HEALTH AND HUMAN SCIENCES HS900 RESEARCH METHODS Using SPSS Session 2 Topics addressed today: 1. Recoding data missing values, collapsing categories 2. Making a simple scale 3. Standardisation
More informationKenyon College ECON 375 Introduction to Econometrics Spring 2011, Keeler. Introduction to Excel January 27, 2011
Kenyon College ECON 375 Introduction to Econometrics Spring 2011, Keeler Introduction to Excel January 27, 2011 1. Open Excel From the START button, go to PROGRAMS and then MICROSOFT OFFICE, and then EXCEL.
More informationNCSS Statistical Software Principal Components Regression. In ordinary least squares, the regression coefficients are estimated using the formula ( )
Chapter 340 Principal Components Regression Introduction is a technique for analyzing multiple regression data that suffer from multicollinearity. When multicollinearity occurs, least squares estimates
More informationDescriptive Statistics
Descriptive Statistics Primer Descriptive statistics Central tendency Variation Relative position Relationships Calculating descriptive statistics Descriptive Statistics Purpose to describe or summarize
More informationChapter 21 Section D
Chapter 21 Section D Statistical Tests for Ordinal Data The ranksum test. You can perform the ranksum test in SPSS by selecting 2 Independent Samples from the Analyze/ Nonparametric Tests menu. The first
More informationAn analysis method for a quantitative outcome and two categorical explanatory variables.
Chapter 11 TwoWay ANOVA An analysis method for a quantitative outcome and two categorical explanatory variables. If an experiment has a quantitative outcome and two categorical explanatory variables that
More informationSimple Linear Regression Excel 2010 Tutorial
Simple Linear Regression Excel 2010 Tutorial This tutorial combines information on how to obtain regression output for Simple Linear Regression from Excel and some aspects of understanding what the output
More informationTable of Contents. Preface
Table of Contents Preface Chapter 1: Introduction 11 Opening an SPSS Data File... 2 12 Viewing the SPSS Screens... 3 o Data View o Variable View o Output View 13 Reading NonSPSS Files... 6 o Convert
More informationc. The factor is the type of TV program that was watched. The treatment is the embedded commercials in the TV programs.
STAT E150  Statistical Methods Assignment 9 Solutions Exercises 12.8, 12.13, 12.75 For each test: Include appropriate graphs to see that the conditions are met. Use Tukey's Honestly Significant Difference
More informationHow to choose a statistical test. Francisco J. Candido dos Reis DGOFMRP University of São Paulo
How to choose a statistical test Francisco J. Candido dos Reis DGOFMRP University of São Paulo Choosing the right test One of the most common queries in stats support is Which analysis should I use There
More informationSimple Predictive Analytics Curtis Seare
Using Excel to Solve Business Problems: Simple Predictive Analytics Curtis Seare Copyright: Vault Analytics July 2010 Contents Section I: Background Information Why use Predictive Analytics? How to use
More informationSummary of important mathematical operations and formulas (from first tutorial):
EXCEL Intermediate Tutorial Summary of important mathematical operations and formulas (from first tutorial): Operation Key Addition + Subtraction  Multiplication * Division / Exponential ^ To enter a
More informationUsing Minitab for Regression Analysis: An extended example
Using Minitab for Regression Analysis: An extended example The following example uses data from another text on fertilizer application and crop yield, and is intended to show how Minitab can be used to
More informationPointBiserial and Biserial Correlations
Chapter 302 PointBiserial and Biserial Correlations Introduction This procedure calculates estimates, confidence intervals, and hypothesis tests for both the pointbiserial and the biserial correlations.
More informationPerform hypothesis testing
Multivariate hypothesis tests for fixed effects Testing homogeneity of level1 variances In the following sections, we use the model displayed in the figure below to illustrate the hypothesis tests. Partial
More informationThere are six different windows that can be opened when using SPSS. The following will give a description of each of them.
SPSS Basics Tutorial 1: SPSS Windows There are six different windows that can be opened when using SPSS. The following will give a description of each of them. The Data Editor The Data Editor is a spreadsheet
More informationIn Chapter 2, we used linear regression to describe linear relationships. The setting for this is a
Math 143 Inference on Regression 1 Review of Linear Regression In Chapter 2, we used linear regression to describe linear relationships. The setting for this is a bivariate data set (i.e., a list of cases/subjects
More informationComparing Two or more than Two Groups
CRJ 716 Using Computers in Social Research Comparing Two or more than Two Groups Comparing Means, Conducting TTests and ANOVA Agron Kaci John Jay College Chapter 9/1: Comparing Two or more than Two Groups
More informationEPS 625 ANALYSIS OF COVARIANCE (ANCOVA) EXAMPLE USING THE GENERAL LINEAR MODEL PROGRAM
EPS 6 ANALYSIS OF COVARIANCE (ANCOVA) EXAMPLE USING THE GENERAL LINEAR MODEL PROGRAM ANCOVA One Continuous Dependent Variable (DVD Rating) Interest Rating in DVD One Categorical/Discrete Independent Variable
More informationINTERPRETING THE REPEATEDMEASURES ANOVA
INTERPRETING THE REPEATEDMEASURES ANOVA USING THE SPSS GENERAL LINEAR MODEL PROGRAM RM ANOVA In this scenario (based on a RM ANOVA example from Leech, Barrett, and Morgan, 2005) each of 12 participants
More informationE205 Final: Version B
Name: Class: Date: E205 Final: Version B Multiple Choice Identify the choice that best completes the statement or answers the question. 1. The owner of a local nightclub has recently surveyed a random
More informationFactor B: Curriculum New Math Control Curriculum (B (B 1 ) Overall Mean (marginal) Females (A 1 ) Factor A: Gender Males (A 2) X 21
1 Factorial ANOVA The ANOVA designs we have dealt with up to this point, known as simple ANOVA or oneway ANOVA, had only one independent grouping variable or factor. However, oftentimes a researcher has
More informationwhere b is the slope of the line and a is the intercept i.e. where the line cuts the y axis.
Least Squares Introduction We have mentioned that one should not always conclude that because two variables are correlated that one variable is causing the other to behave a certain way. However, sometimes
More informationData Analysis. Lecture Empirical Model Building and Methods (Empirische Modellbildung und Methoden) SS Analysis of Experiments  Introduction
Data Analysis Lecture Empirical Model Building and Methods (Empirische Modellbildung und Methoden) Prof. Dr. Dr. h.c. Dieter Rombach Dr. Andreas Jedlitschka SS 2014 Analysis of Experiments  Introduction
More informationIBM SPSS Statistics 23 Part 4: ChiSquare and ANOVA
IBM SPSS Statistics 23 Part 4: ChiSquare and ANOVA Winter 2016, Version 1 Table of Contents Introduction... 2 Downloading the Data Files... 2 ChiSquare... 2 ChiSquare Test for GoodnessofFit... 2 With
More informationANOVA Analysis of Variance
ANOVA Analysis of Variance What is ANOVA and why do we use it? Can test hypotheses about mean differences between more than 2 samples. Can also make inferences about the effects of several different IVs,
More informationCHARTS AND GRAPHS INTRODUCTION USING SPSS TO DRAW GRAPHS SPSS GRAPH OPTIONS CAG08
CHARTS AND GRAPHS INTRODUCTION SPSS and Excel each contain a number of options for producing what are sometimes known as business graphics  i.e. statistical charts and diagrams. This handout explores
More information