Estimation of Population Parameters: 1) Estimating The Population Mean
Estimation of Population Parameters: 1) Estimating The Population Mean
The overall objective of descriptive statistics is to give you a detailed description of the data you have on
hand. As we have only limited data or sample data on hand, we are mostly required to estimate the
population parameters from the sample data. The parameter calculated from the sample data is not
100% accurate and might result in small errors while estimating the parameters for the population.
a) POINT Estimate – is a statistic taken from a sample that is used to estimate a population
parameter, which is as good as the representativeness of its sample. If other random samples
are taken from the population, the point estimates derived from those samples are likely to
vary. The variations or errors likely to rise out of different samples are called STANDARD
ERRORS.
Definition: A confidence interval for a parameter is an interval of numbers within which we expect the
true value of the population parameter to be contained. The endpoints of the interval are computed
based on sample information
Certain factors may affect the confidence interval size including size of sample, level of
confidence, and population variability. A larger sample size normally will lead to a better
estimate of the population parameter.
The CI yields a range within which we feel with some confidence, the population mean is
located. The interpretation is like this, it is not certain that the population mean is in the interval
unless we have a 100% confidence interval that is infinitely wide. If we want to construct a 95%
CI, the level of confidence of the analyst is 95% or 0.95 that the intervals would include the
population mean and five would not include.
In reality, a CI with 100% confidence would be meaningless. So researches go with 90%, 95%,
98% or 99% max. The reason is that there is a trade-off between sample sizes, interval width,
level of confidence etc. For Ex: as the level of confidence is increased, the interval gets
narrower. Which means data coverage range / data distribution (z level) and CI have an inverse
relationship. As the analyst takes wider data coverage (z value), his chances of increasing
estimate accuracy increases and his Confidence Interval will be more precise and narrow.
How confident are we that the true population average is in the shaded area? We are 95% confident.
This is the level of confidence. How many standard errors away from the mean must we go to be 95%
confident? From -z to z there is 95% of the normal curve.
There are 4 typical levels of confidence: 99%, 98%, 95% and 90%. Each of the levels of confidence has a
different number of standard errors associated with it. We denote this by
where a is the total amount of area in the tails of the normal curve. Thus, for a 95% level of confidence,
the z values from the table are:
Level of confidence /2
90% 5% 1.645
98% 1% 2.33
example: A sample of 100 observations is collected and yields mean xbar=75 and s=8. Find a 95%
confidence interval for the true population average.
=(73.432,76.568). So the Population Mean is estimated to be between 73.432 and 76.568 for a CI of
95%.
𝝁 ± 𝒛√𝒑 ∗ 𝒒/𝑵
The sample mean is 0.39, the z value of error for a 95% CI is 1.96, the sample size is 87,
but the proportion p of sample is 0.39, proportion q is 1-p=0.61 (since we deal with
proportions, we talk about a term called q value, which 1-p, in this case 1-0.39=0.61)
0.39±𝟏. 𝟗𝟔√𝟎. 𝟑𝟗 ∗ 𝟎. 𝟔𝟏/𝟖𝟕
0.39±0.102 = 0.39+0.102=0.492 or 0.39-0.102=0.29
Hence the proportion of the population is estimated to be between 0.29 and 0.49 or
between 29% and 49%, at a confidence level of 95% with a range of 0.102 from the
point estimate.
The relationship of the sample variance to the population variance is captured by the CHI-
SQUARE Distribution (x2). That is , the ratio of the sample variance (s2) multiplied by N-1 to
the population variance (σ2) is approximately chi-square distributed, if the population is
normally distributed. This does not suit conditions where the data is not normally
distributed.
Degrees of Freedom:
Degrees of Freedom refers to the number of independent observations for a source of
variation minus the number of independent parameters estimated, in computing the
variation. So, if you have 50 observations and you calculate 2 parameters, your DF will be N-
parameters, 50-2=48.
HYPOTHESIS TESTING
Introduction
Business researchers often need to provide insight and information to decision makers to assist them in
answering questions like,
What container shape is most economical and reliable for shipping a product?
Which management approach best motivates employees in the retail industry?
What is the most effective means of advertising in a business setting?
How can the company’s retirement investment financial portfolio be diversified for optimum
performance?
For these purposes, researches develop “Hypotheses” to be studied and explored.
We explore all types of hypotheses, how to test them, how to interpret the results of such tests to
help decision making. Research hypothesis is a statement of what the researcher believes will be the
outcome of an experiment or a study. Business researchers have some idea or theory based on
previous experience and data as to how the study will turn out. These are typically concerning
relationships, approaches and techniques in business.
Statistical hypothesis is required in order to scientifically test the research hypothesis. All Statistical
hypothesis consist of 2 parts, NULL HYPOTHESIS & ALTERNATIVE HYPOTHESIS.
NULL HYPOTHESIS – usually states that “Null” condition exists, that there is nothing new
happening, the old theory is still true, the old standards / quality are correct, and the system is
under control. It is represented by H0.
ALTERNATIVE HYPOTHESIS - on the other hand usually states that the new theory is true, there
are new standards, the system is out of control or something different is happening all the time.
Represented as H1.
Ex 1 : Suppose a baking flour manufacturer has a package size of 40 ounces and wants to test whether
their packaging process is correct, the NULL hypothesis for this test would be that the average weight of
the pack is 40 ounces (no problem). The ALTERNATIVE hypothesis is that the average is not 40 ounces
(process has differences).
Ex2: Suppose a company held an 18% market share earlier and because of increased marketing effort,
company officials believe the market share is more than 18%, the market researches would like to prove
it.
H0: p = 0.18
H1: p > 0.18
Note: Though the statistical calculations gives results are T- Value of F Values for the above tests, in
SPSS,Excel.R,Python,SAS we look mainly at the Significance value(P-value). If this p value is less than
0.05, we reject Null Hypothesis, else accept it. This is because, the T Tables and F tables are not
handy.
STEPS for HYPOTHESIS TESTING – Most of the hypothesis testing is based on Mean comparisons.
1) State the Hypotheses, (both NULL and Alternative) clearly. H0 and H1. The purpose of the test
should be clearly understood as per the applications. The researcher needs to be doubly sure of
the requirement and purpose, to negate the null hypothesis.
2) Specify Level of significance – can also be said as allowable non-confidence limits. It is also the
probability of committing Type 1 Error. Common values of alpha are 0.05, 0.03, 0.01 etc,
depending on the criticality of the business errors. Ex: A Retail industry test may accept a level
of 5%, but the aeronautics industry would want only a 0.01 level. The drugs industry would
want even precise levels of testing as close as 99.2%.
3) Use appropriate statistical test, based on requirement and based on hypothesis in Step 1.
Hypothesis
Testing
4) Decision Rule- The researcher should be 1-α confident to prove his hypothesis. General rule is
P ≤ 0.05 (alpha level) you reject H0 and prove your theory.
P≥ 0.05 (alpha level is more) you accept H0 to prove the older theory.
OR
If T.Value or F.Value (from step 3) is more than table value, then reject H0, else accept H0.
Ex: if alpha is 0.05, and P value is 0.03, then the researcher is 97% confident of his
theory and can reject Null Hypothesis. If he is less confident or p value is more than alpha
level, his confidence levels are going down and is forced to accept the null hypothesis and
that his tests have failed.
Actual Claim H0
(TP)Correct
(TN) TYPE 2 Error
Accept Decision
(β=0.95)(1-alpha)
(1-alpha)=0.95
(FN) Correct
(FP) TYPE 1 Error Decision
Reject
Alpha=0.05 (power of test
Beta)
t-Test
A t-test is an analysis of two populations means through the use of statistical examination; a t-test with
two samples is commonly used with small sample sizes, testing the difference between the samples
when the variances of two normal distributions are not known.
A t-test looks at the t-statistic, the t-distribution and degrees of freedom to determine the probability of
difference between populations; the test statistic in the test is known as the t-statistic.
Example 1) Philips Bulb Co. states that the average lifetime of EchoStar Bulb is 10 years.
Now WIPRO doesn’t accept this claim and tests the average life of 15 Philips Bulbs.
14 -2.27 0.0392
1) The only variable being checked for Mean value here is LIFETIME.
2) N is the number of observations
3) Mean of LIFETIME variable is 9.3000 (here it says average life of the bulb is 9.30 yrs)
4) STD Deviation of all the 15 values from the mean is 1.1922
5) Minimum and maximum are the values in the data range in the LIFETIME variable.
6) 95% CL Mean – is the Mean at 95% Confidence Limit of data being picked up for analysis.
7) The Confidence Interval for the Mean is 8.6398 to 9.9602
8) 95% Confidence Interval for SD is 0.8729 to 1.8803
9) DF – Degrees of Freedom (always N-1)
10) T.Value is actually not considered here as we cannot compare it with the table values.
11) Probability to T.Value is 0.0392.
Here the P value is the one to be considered. Since it is much lesser than alpha 0.05, we REJECT Null
Hypothesis and say that the average life time of Philips Bulb is NOT 10 yrs. Here is the P value was higher
than 0.05, we would accept the H0 and say the lifetime is 10 yrs.
Example 2: 15 Customers each in Mumbai and Delhi were asked to rate Brand X on a 7 point scale. The
responses of all 30 customers are presented. Test whether the responses to Brand X is the same in both
cities.
Example 3:- We have recorded the ratings of Tamarind brand garments from 18 respondents before and
after an advertisement campaign were released for this brand. The ratings are on a 10 point scale. Test
whether the campaign had an effect on sales of Tamarind brand garments.
Before After
3 4 Step 1) H0: No effect of the advt campaign (Mean before and after are same)
4 5 H1: Advt Campaign had an effect on sales (Mean sales before / after are diff)
2 3 Step 2) Alpha 0.05
5 4 Step 3) Go for Paired T.Test since the same sample obs are tested twice
3 5
4 5
5 6
3 4
4 5
2 4
2 4
4 5
1 3
3 5
6 8
3 4
2 4
3 5
When we say, more than 2 variables, it means, influence of 1 or more factors on a variable.
Ex: Sales affected by location of item displayed. (Either window or near counter or shelf)-
Though there are 3 options, they are categorised as a single factor influencing sales. Here
ANOVA is mainly concerned with the analysis of “which one is better among the 3” options
within the same variable called storage area.
2 Factors can also influence a variable. Ex: Sales because of storage location and price. In such
cases, ANOVA gives the combination of the factors which will fetch max results.
ANOVA mainly deals with a detailed analysis of variances in the variables as follows:-
Between the variables, (ex. Variance of Variable 1 and variance of variable 2 )
Within the variable (variances in variable 1 from its mean)
In totality.
Analyse if the sales of kit Kat is affected by the area of display (storage location) in a store. Sales for a
week are observed when kit Kat is placed at the window side, counter side or on the shelf.
F- TESTS
F-Test to Compare Two Variances. However, the f-statistic is used in a variety of tests
including regression analysis, the Chow test and the Scheffe Test (a post-hoc ANOVA test).
An F-test (Snedecor and Cochran, 1983) is used to test if the variances of two populations are equal.
This test can be a two-tailed test or a one-tailed test. The two-tailed version tests against the alternative
that the variances are not equal. The one-tailed version only tests in one direction that is the variance
from the first population is either greater than or less than (but not both) the second population
variance. The choice is determined by the problem. For example, if we are testing a new process, we
may only be interested in knowing if the new process is less variable than the old process.
All the hypothesis tests like Simple T.Test, Independent T.Tests or even ANOVA deal with numeric or
continuous data, which can be quantified for further calculations. If your data is categorical or nominal
and when the data is qualitative, statistical calculations get limited.
Categorical data are nonnumerical data that are frequency counts of categories in one or more
variables.
Hence the question of comparison of Means or Variances doesn’t arise. A categorical comparison needs
to be done based on possible category values and their frequency occurrences, in both qualitative
variables.
In such cases also, business requires you to analyse the categorical data and give inferences based on
past history and forecast the future. Thus, we use CHI-SQUARE Tests for Categorical data. Chi-Square
tests can be used even if one of the variables is qualitative or all your variables are qualitative.
Chi-Square is simply the sum of the squares of the difference between Observed minus predicted value
divided by expected values. Chi-square records the differences between observed and expected, takes
their squares and compare it by the expected values.
(Observed−Expected)2
Chi-Square is represented by 𝑥2 = ∑
Expected
Applications of Chi-Square:-
Your categorical data can be analysed by using Chi-Square in the following situations:-
1) CHI-SQUARE GOODNESS OF FIT Test - To check goodness of fit and compare observed vs.
predicted values For Ex: if the variable is Economic class with 3 possible outcomes of lower
income class, middle income class and upper income class, the single dimension is economic
class the 3 possible outcomes are the 3 classes. On each trial, only one class can occur. Or the
family can be categorised only under one class at a time.
Here, the Chi-Square goodness of fit test compares the expected, or calculated frequencies of
categories with the observed / actual frequencies from a dataset to determine whether there is a
difference between what was expected and what was actually observed.
Ex: Financial investors decisions are based on region or not. The researcher gets answers to
two questions which have categorical answers.
a) In which region the investor resides (North, East, West, South)
b) Which type of investment he prefers (stocks, bonds, treasury bills)
In such cases, the researcher would tally the frequencies of responses to these two questions
into two categorical variables and record them in a two dimensional CONTINGENCY TABLE. Hence
this Chi-Square test of independence can also be called CONTINGENCY ANALYSIS.
(F0−fe)2
𝑥2 = ∑ ∑ fe
Note: Here the frequencies of both variables are taken into account, and the DF lost will be 2.
The relationship of the sample variance to the population variance is also captured by the CHI-SQUARE
Distribution (x2). That is , the ratio of the sample variance (s2) multiplied by N-1 to the population
variance (σ2) is approximately chi-square distributed, if the population is normally distributed. This does
not suit conditions where the data is not normally distributed.
PRACTICAL PROBLEMS
Example
Is gender independent of education level? A random sample of 395 people were surveyed and each
person was asked to report the highest education level they obtained. The data that resulted from the
survey is summarized in the following table:
BINOMIAL DISTRIBUTION:
Notation
EX: Consider the following statistical experiment. You flip a coin 2 times and count the
number of times the coin lands on heads. This is a binomial experiment because:
POISSION DISTRIBUTION:
where
e is the base of natural logarithms (2.7183)
μ is the mean number of "successes"
x is the number of "successes" in question
EX: Suppose you knew that the mean number of calls to a fire station on a
weekday is 8. What is the probability that on a given weekday there would be 11
calls? This problem can be solved using the following formula based on the
Poisson distribution
Normal Distribution
The random variable X in the normal equation is called the normal random variable. The
normal equation is the probability density function for the normal distribution.
The graph of the normal distribution depends on two factors - the mean and the
standard deviation.
The mean of the distribution determines the location of the center of the graph,
The standard deviation determines the height and width of the graph.
When the standard deviation is large, the curve is short and wide;
When the standard deviation is small, the curve is tall and narrow.
All normal distributions look like a symmetric, bell-shaped curve, as shown below.
The curve on the left is shorter and wider than the curve on the right, because the curve
on the left has a bigger standard deviation.
Additionally, every normal curve (regardless of its mean or standard deviation) conforms
to the following "rule".
About 68% of the area under the curve falls within 1 standard deviation of the
mean.
About 95% of the area under the curve falls within 2 standard deviations of the
mean.
About 99.7% of the area under the curve falls within 3 standard deviations of the
mean.
Collectively, these points are known as the empirical rule or the 68-95-99.7 rule. Clearly,
given a normal distribution, most outcomes will be within 3 standard deviations of the
mean.
Example 1
An average light bulb manufactured by the Acme Corporation lasts 300 days with a
standard deviation of 50 days. Assuming that bulb life is normally distributed, what is
the probability that an Acme light bulb will last at most 365 days?
Solution: Given a mean score of 300 days and a standard deviation of 50 days, we want
to find the cumulative probability that bulb life is less than or equal to 365 days. Thus,
we know the following:
We enter these values into the Normal Distribution Calculator and compute the
cumulative probability. The answer is: P( X < 365) = 0.90. Hence, there is a 90% chance
that a light bulb will burn out within 365 days.