SPSS Workshop: Utilizing and Implementing SPSS in Our OC-Math Statistics Classes
SPSS Workshop: Utilizing and Implementing SPSS in Our OC-Math Statistics Classes
My Name Is:
Starting SPSS, entering and modifying data will be studied in this section.
Create a folder on the desktop under your name. After this go to Start, Programs, Spss, Pasw Statistics 18 .
(wait one minute)
Name variables in SPSS before entering data. Click the Variable View tab (bottom).
Enter name for variable 1, then for variable 2,... Advices regarding Variable View
In Variable View, you can NOT enter spaces for the name of variables. Preferably, enter quantitative data. When you enter integer values, select 0 in Decimals You can also enter labels. Here you can type spaces etc. These would be the text that will be displayed when the mouse is on that variable. Example: If you enter AgeMarr for the name of the variable, you can type Age When Got Married under labels. 5. You can assign values to a variable, ex. 1=female, 2=male 1. 2. 3. 4.
Start Typing or copying data from excel or from another document Copying from excel: mark data and do control c then go to SPSS and do control v Saving: Click on File, Save, then browse to your folder and type example1.sav
My Name Is:
Select your variables and them into the Variable(s): pane on the right. Click on Statistics and on the new window (Frequencies Statistics) select
Mean, Median, Mode and Standard Deviation, then continue Click on Charts, Select None.
Ok In the Output Window: go to Left pane: and Click inside Frequency Table, observe.
Click on Charts , Bar Charts and click continue then click ok Now, do Analyze, Descriptive Statistics, Frequencies
Select your variables and them into the Variable(s): pane on the right.
Histogram,
Finally do Analyze, Descriptive Statistics, Frequencies Charts, Pie Charts, Percentages, Continue, Ok observe.
In the Output Window: go to Left pane: and Click inside Frequencies, find your charts and
My Name Is:
Relationship: Correlation
Analyze, Correlate, Bivariate In the Bivariate Correlations dialog window select 2 to 5 variables Make sure Pearson, Two Tailed and Flag Signif... are checked, Ok Output Window: Left pane: click on Correlations then Correlations
Analyze, Descriptive Statistics, Crosstabs Move variables from left to right, one in columns and one in rows Select Statistics, checkbox Chi-square then continue, then click ok Output Window: Left pane almost at bottom : click on Cross Tabs/Chi-Square
Analyze, Compare Means, Independent-Samples T... In the Test Variable select one variable, in the Group Variable select another.
You are trying to split into two groups to find the differences between these two groups.
My Name Is:
. then Ok
Linear Regression
(I want to know how salary is related to GPA)
Analyze, Regression, Linear Select dependent and independent variable Continue Then Ok. Look at Coefficients in the Model box. B column has a top and bottom coefficient.
Coeeficient a=bottom in column B, Coefficient b=top in Column B.
Linear Regression:
GRAPH
Graph, (Legacy Dialogs) Scater/Dot, Simple, Define Select dependent and independent variable Ok, Double click on the plot, be patient, then Options, then Reference line From Coefficients obtained in the Linear Regression, under Custom Equation insert Checkbox: Attach label to. Close. Also close the Chart Editor by closing the window on the top right corner.
y=(coefficient a you got before)*x + (coefficient b you got before),
My Name Is:
Click Plots then choose Zresid for the Y axis and ZPred for the X axis. Continue Then Ok. Double click on the plot, then Options, then Reference line From Equation, Enter
y=2.5, then Apply , close. Also close the Chart Editor.
Testing Normality
Analyze, Descriptive Statistics , Q-Q plots, choose a variable, Select test distribution to be normal, then Ok
Kolmogorov-Smirnov Test
Test if Data are Normal or Test Randomness
Analyze, Nonoparametric Test , One-Sample , (may choose a variable) Check Automatically Compare Observed data to hypothesized
(You could have also chosen the TEST SEQUENCE FOR RANDOMNESS to see if the sequence of values defined by some cutting point is random) If you dont see this AUTOMATICALLY COMP option, select Descriptive and also Exclude cases test-by-test.
Run Output Window: Left pane: click on NonParametric Tests, then Model Viewer
Binomial Test
When Normality fails, Test if Data are Binomial
Binomial ,
My Name Is:
Select a variable Cut Point Enter a value (about the mean) Test Proportion Enter a value (about 50%) Options, Descriptive and also Exclude cases test-by-test) Ok Output Window: Left pane: click on NPar Tests, then Binomial Test
Advanced statistics will be studied in this section.
Hypothesis Testing
Running a T-Test: Is the GPA=3.5?
Analyze, Compare Means, One-Sample T-Test In the Test Variable select one variable, Enter a value to test in Test Value (typically a mean) then OK Output Window: Left pane: click on T-Test, then Independent Samples
Chi Square
Expecting 10%Hispanic, 10%Asian, 70%White, 10%AfriAmer
Analyze, Nonoparametric Test , Chi-Square , Select a variable Expected Values Enter a value, then Add. Repeat. Options, Descriptive and also Exclude cases test-by-test) Ok Output Window: Left pane: click on NPar Tests, then Chi-Square Test
My Name Is:
One-Way ANOVA
(In AgeGender: 1=MaleOld, 2=MaleYong,3=FemaleOld, 4=Female Young) (for example Salary is the Factor, AgeGender is in the Dependent List)
Analyze, Compare Means , One-Way Anova , Select a independent variable (Factor), in the given example AgeGender. Drag a variable
into the Dependent List, for example Salary.
Post Hoc Select LSD then Continue. Options, Descriptive, Means Plot, Continue, Output Window: Left pane: click on Oneway, then ANOVA
OK
A large standard deviation indicates that the data points are far from the mean and a small standard deviation indicates that they are clustered closely around the mean. It is a measure of dispersion. For the cross tabs, for the Chi-Square value A) If Sig < 0.05, then the variables are significantly related. B) If Sig > 0.05, then the variables are NOT significantly related.
For the correlation, the important value is the Pearson (r) value, 1) if the r values greater than .50 indicate a strong correlation 2) if the r values around .30 indicatea moderate correlation 3) if the r values less than .20 indicate a weak correlation
For the t-test, independent t-test , 1)If Sig.< 0.05, use the output in the Equal variances NOT assumed rows. 2)If Sig.> 0.05, use the output in the Equal variances assumed rows. Then we move to the Sig (2-tailed), A) If Sig< 0.05 we conclude that there was significant difference between the two means, i.e. that the difference is real and not due to randomness. B) If Sig> 0.05 we conclude that there was no significant difference between the two means.
My Name Is: A) If Sig < 0.05 we conclude that there was significant difference between the two means, i.e. that the difference is real and not due to randomness. B) If Sig> 0.05 we conclude that there was no significant difference between the two means.
Kolmogorov-Smirnov Test
We look at the Sig (2-tailed) A) If Sig < 0.05 we conclude the data are NOT normal. B) If Sig) > 0.05 we conclude there is insufficient evidence to reject normality (we then proceed with the assumption the data are normal)
P - Value
We look at the Sig (2-tailed) The p-value is the same as the Asymp. Sig. A) If p-value (or Sig) < 0.05 we reject the null hypothesis B) If p-value (or Sig) >0.05 we conclude there is insufficient evidence to reject the null hypothesis (we then proceed with the assumption the null hypothesis is true, i.e. we accept it)
Zero P Value
If mean1=hypothesized mean (the test value we input in our t-test) and mean2=the mean from our sample of size N, then a p-value of 0 indicates that there is a zero probability that in a random sample of size N from a population with mean1 we, by purely chance, obtain mean 2. If this is the case, we say that it is unlikely that the mean1 has occurred by chance and it is more likely that the mean1 is not as hypothesized.
Hypothesis Testing is all about gathering evidence to suggest the null is not
true, the lack of such evidence warrants a Do not reject decision. Remember, not guilty doesnt mean the person is innocent, it means, the jury did not find enough evidence to condemn the person.
My Name Is:
Binomial Test
When normality fails, we use a weaker test: The Binomial test. We have a cutting point and a proposed proportion or percentage. We look at the Sig (2-tailed) A) If Sig < 0.05 we reject the null hypothesis, hence the proposed percentage of our variable is NOT above our cutting point B) If Sig) > 0.05 There is insufficient evidence to conclude that the proposed percentage of our variable is not above our cutting point.
For the Q-Q Plot, if the dots look pretty much like the straight line, we conclude the data look close to a normal distribution; otherwise we conclude that the data are far from a normal distribution.
One-Way Anova If p (Sig) <0.05 we can conclude as follows: Oneway ANOVA revealed that there were significant differences in the variable (of interest) between these groups. Whereas if p (Sig) >0.05 we can conclude as follows: According to the Anova Table, there is insufficient evidence to reject the null, in this case, we found not enough evidence to conclude that the there are no significant differences in the variable of interest among these groups.
The linear regression straight line, tells you that it is the best straight line that fits the data (it minimizes the sum of the squares of the distances from the points to the line). You can also save the output by going to the output window and click Save then browse to your folder and type example1.SPO
http://wwwstage.valpo.edu/other/dabook/ch8/c8-3.htm DATAPLOT (Visual), R (Fav), OPENSTAT (Soc. Sci.), INSTAT (Climat. Data)
My Name Is:
Linear Regression
Note: It is often very useful to look at the standardized residual versus standardized predicted plot in order
to look for outliers and to check for homogeneity of variance. The ideal situation is to see no observations beyond the reference lines, which means that there are no outliers. Also, we would like the points on the plot to be distributed randomly, which means that all the systematic variance has been explained by the model.
Anova
As with the t-test, ANOVA also tests for significant differences between groups. But while the t-test is limited to the comparison of only two groups, one-way ANOVA can be used to test differences in three or more groups. The ANOVA procedure produces an F statistic, a value whose probability enables the researcher to reject or retain the null hypothesis, i.e., to conclude whether or not the differences in the scores on the dependent variable are statistically significant or due to chance.
Two-Ways Anova Factorial ANOVA improves on one-way ANOVA in that the researcher can simultaneously assess the effects of two (or more) independent variables on a single dependent variable within the same analysis. Thus, factorial ANOVA yields the same information that two one-way ANOVA's would, but it does so in one analysis. But that's not all. Factorial ANOVA also allows the investigator to determine the possible combined effects of the independent variables. That is, it also assesses the ways in which these variables interact with one another to influence scores on the dependent variable.