Statistics Assignment
Statistics Assignment
Population: population is totality or collection of all objects, items, or individuals on which observation
are taken on the basis of the some characteristic of the objects in any field of inquiry.
1. Finite population.
2. Infinite population.
Finite population: when it contains finite number of experimental unit. For example: Age of all students
in department of clinical Psychology at University of Rajshahi.
Infinite population: when it contains infinite number of experimental unit. For example: Dots in a line.
What is sample?
Sample: A sample is a part of population that is taken and considered for study. Example; if the
population is the total students of the University of Rajshahi the sample would
be a small number of subset of the students of the University of Rajshahi.
b) Large Sample: In this type of sample the sample size exceeds the number 30.
parameter statistic
Any numerical value that describing the Any numerical value that describing the
characteristic of population is called parameter. characteristic of sample is called parameter.
It is denoted by the Greek letter (mu) It is denoted by small letter of the English alphabet
Population parameter is more accurate than sample Sample statistic is less accurate than population
statistic. parameter.
Example: population mean, population variance Sample mean. Sample variance
Population parameter is more reliable than sample Sample statistic is less than population parameter.
statistics
What is sampling?
On the other hand sampling defined as it is the process of selecting a sample from population.
Sampling is Classified into two broad categories:
1. Probability Sampling.
2. Non-probability sampling.
It’s also called random sampling. It’s also called non-random sampling.
Sampling error: Statistical error are sample error. The sampling error are influenced by sample size and
sampling scheme. It’s occurred by the act of taking sample and also occurs when the result from sample is
very different from result of population.
Sampling bias: According to Gillian Fournier it is also known as selection bias an error in choosing
participants for a scientific study.
On the other hand sampling bias is a bias in which a sample collected in such a way that some members
of the intend population are less likely to be included than others.
Test of significance: it is a procedure which enable us to decide whether to accept or reject the hypothesis
or to determine whether observed samples differ significantly from expected results is called test of
significance.
According to C.R Kothari the procedure which enable us to decide on the basis of sample if the deviation
between the observed sample statistic and hypothetical parametric value or two independent sample
statistics is significant or might be attributed to the calculation. It is also known as a test of hypothesis,
rules of decision.
What is variance?
Variance: variance is a statistical measure of how much a set of observation differ from each other. In
other words the arithmetic mean of the squares of the deviation of the observation from their arithmetic
mean is known as a variance. It is often represented by Var(x) or 𝜎2.
What is co-efficient of correlation?
Co- efficient of co-relation: co- relation coefficients are used in statistics to measure how strong a
relationship is between two variables. The values range between -1 to1. There are several types
of correlation coefficient Pearson’s correlation is a correlation coefficient commonly used in linear
regression.
T Test of hypothesis: A testable statement which may be true or not true is called test of hypothesis.
1. Null hypothesis.
2. Alternative hypothesis.
Null hypothesis: There are no significant differences between parameter and statistic.
Alternative hypothesis: which differ from null hypothesis is called alternative hypothesis.
Level of significance: In testing a given hypothesis, the maximum probability with which we
would be willing to risk a type-I error is called level of significance of the test.
The level of significance is defined as the probability of rejecting a null hypothesis by the test when it is
really true.
The probability of type-I error is called level of significance. The standard level of significance α=0.05
Degrees of freedom: In statistics, the number of degrees of freedom is the number of values in
the final calculation of a statistic that are free to vary. The formula of degrees of freedom (df) is,
df=N-1
Type-I error: In hypothesis testing if we reject hypothesis when it should be acceptable then it is
called type-I error also known as an error of the first kind. Example: rejection of null hypothesis when it
is true.
Type-II error: In hypothesis testing if we accept a hypothesis when it should be rejected then it is
called type-II error also known as an error of the second type. Example: accepting a null hypothesis
when it is not true.
Differences between one-tailed & two-tailed test.
Power of a test: Ability to correctly rejecting a null hypothesis is called power of a test.
On the other hand the power of a statistical test gives the likelihood of rejecting the null hypothesis when
the null hypothesis is false. Power of a test =1-β
Parametric test: parametric test are those test that are stated in terms of making assumption about
population parameter.
Non-parametric test: Non-parametric test are those that do not compare population parameter and make
fewer assumption than parametric test.
Assumption of non parametric test: There are some assumptions regarding a non-parametric test; these
are;
a) The data re not normally distributed
b) The data show in homogeneity of variance
c) The data are measured on an ordinal scale
d) Distribution is not continuous
t-test F-test
In this test sample should be relatively small. In this test the sample size can be 30 or more than
30.
t-test used to justify differences between two F-test used to justify differences between more than
groups. two groups.
Data obtained from two groups in this test. Data obtained from more than two groups.
Co-related & uncorrelated types of t-test. One way and two way analysis of variance is a type
of F-test.
Main effect: when an independent variable individually effect on dependent variable is called main effect.
Interaction effect: when an independent variable combinedly effect on dependent variable then its called
interaction effect.
Goodness of fit: Goodness of fit is the extent to which observed data matches the values expected by
theory. Chi square test and coefficient of determination can be used to determine goodness of fit.
In other words goodness of a fit of statistic model describes how well it fits a set of observation.
Factor analysis: factor analysis is a class of procedures that allow the researchers to observe a group of
variables that tend to be correlated to each other and identify the underlying dimensions that explain these
correlations.
According to C.R Kothari factor analysis is a technique applicable when there is a systematic
interdependence among a set of observed or manifest variables and the researchers is interested in finding
out something more fundamental or latent which creates this commonality.
What is uniqueness?
Uniqueness: Uniqueness is a state or condition where in someone or something is unlike anything else in
comparison.
What is communality?
Communality: The concept of communality in factor analysis shows how much of each variable is
accounted for the underlying factor taken together.
A high value of communality means not a much of the variable is left over after whether the factors
represent is taken into consideration.
Factor loading: factor loading is the correlation between the original variables and the factors.
According to C.R Kothari factor loading are those values which explain how closely the variables are
related to each one of the factors discovered. They are also known as a factor variable correlations.
Tied score: The score obtained from tied observation is known as tied score.Tied observations are
observations having the same value which prohibits the assignment of unique rank number .
Significant difference: The difference which arises due to some other reasons except sampling fluctuation
is known as a significant difference.
Insignificant difference: the difference which arises due to sampling fluctuation is known as insignificant
difference.
ANOVA stands for analysis of variance. ANOVA is a technique in which we compare three or
more population means to determine whether they could be equal.
Yet’s correction: It is a procedure used in certain situation when testing for certain independence in a
contingency table.