Statistics Assignment Full
Statistics Assignment Full
Statistics Assignment Full
KAVYA SHRI
PROGRAMME MASTER OF BUSINESS ADMINISTRATION (MBA)
SEMESTER I
COURSE CODE & NAME DMBA103-STATISTICS FOR MANAGEMENT
ROLL NO. 2314514198
Assign 1
Q1:
Define statistics? Explain various functions of statistics and also the key limitations of
statistics?
Statistics:
Data collection, analysis, interpretation, presentation, and organization are all part of the
mathematical field of statistics. It offers techniques for drawing conclusions about
populations from the analysis of a representative sample. In many disciplines, such as public
health, commerce, economics, social sciences, and science, statistics is essential for
condensing data and producing insightful findings.
Q2 :
define measurement scales? discuss qualitative and quantitative data in detail with
examples?
The various classification and measuring schemes for variables are referred to as
measurement scales, sometimes called data scales or level of measurement. Measurement
scales can be classified into four categories: nominal, ordinal, interval, and ratio. These
scales offer a framework for comprehending the type and properties of the information
gathered.
The nominal scale
The most basic type of measurement scale is this one.
It entails naming or labeling many categories without regard to any sort of natural hierarchy
or order.
Gender (Female, Male), for example Divorce, single, married, and color-wise (red, blue,
green)
Ordinal Scale:
Although the data in this scale are sorted and categorized, there are irregular gaps between
the categories.
It displays the variables' relative rankings or order.
Examples: Level of Education (High School, Undergraduate, Graduate, and Doctorate),
position in a competition (first, second, or third).
Interval scale:
The variables are kept in the same order and have consistent intervals between them on the
interval scale.
It does not, however, have a real zero point.
Examples include IQ scores and temperature in degrees Celsius or Fahrenheit.
ratio scale:
The ratio scale has a constant interval between variables with the interval scale.
Additionally, it has a real zero point, meaning that zero denotes the lack of the measured
property entirely.
Examples are age, income, weight, and height.
Let's now talk about both quantitative and qualitative data:
Definition of Qualitative Data: Qualitative data is non-numerical information that
characterizes attributes or traits. It is frequently categorized and can be separated into many
groups according to certain characteristics.
Features:
Non-Numerical: Rather than using numerical values, qualitative data is expressed using
words, labels, or categories.
Subjective: Because qualitative data interpretation frequently incorporates opinions, feelings,
or perceptions, it may be subjective.
Labels or groups: Typically, data is arranged into classes or groups according to shared
characteristics.
Qualitative Data Examples:
Gender: Classifications such as male and female.
Color: Characteristics like green, blue, or red.
Categories for marital status include single, married, and divorced.
Level of Education: Terms like high school, bachelor's, master's, and doctorate.
Definition of Quantitative Data: Information that can be measured and expressed numerically
is referred to as quantitative data. It is unbiased and amenable to mathematical interpretation.
Features:
Numerical: Mathematical procedures are possible because quantitative data is expressed in
numerical values.
Objective: Since quantitative data interpretation involves precise measurements, it is
typically more objective.
Countable and unique values (e.g., number of automobiles) or an unlimited number of
alternative values (e.g., height) are the two conceivable states for quantitative data.
Quantitative Data Examples:
Height: expressed either in centimeters or inches.
Weight: Expressed in kilos or pounds.
Temperature: Stated either in Fahrenheit or Celsius.
Income: Measurable in terms of money (dollars, for example).
A discrete count of the number of students in a class.
Q3:
Discuss the basic laws of sampling theory. Define stratified sampling technique with the
help of examples
A subfield of statistics known as sampling theory examines how to choose a sample—a
subset of a larger population—from which to draw conclusions about the population as a
whole. Among the fundamental rules of sampling theory are:
The Law of Statistical Regularity postulates that a random sample taken from a population
will have the same features as the population it is taken from.
The Law of Inertia of Large Numbers asserts that the features of the sample will
progressively resemble those of the population as sample size grows.
The Law of Unintended Consequences postulates that biases or unanticipated events may
prevent random samples from adequately representing the population.
The Law of Sampling mistakes is a paradigm for estimating and measuring sampling
mistakes, and it recognizes that errors of some degree will always exist in each sample.
Stratified Sampling: This sampling technique involves randomly selecting samples from each
stratum once the population is split up into subgroups or strata according to specific features.
When there are notable variances or differences within the population that must be
represented in the sample, this approach is employed.
As an example, let's say you wish to survey students on their academic achievement in a
school. Rather than choosing pupils at random from the entire school, you opt to employ
grade-based stratified sampling. Here is how the strata are defined:
Cluster sampling.
Using the cluster sampling technique, a random sample of the population's clusters is chosen
after the population has been divided into groups or clusters. Every person in the selected
clusters is part of the sample. When the population is organically organized or when
compiling a comprehensive list of the population is challenging or costly, this approach is
especially helpful.
As an illustration, let's look at a study that aims to evaluate the academic achievement of
schools in a big city. Rather than selecting students for sampling, the researchers choose to
employ cluster sampling. Here's one way they could go about it:
Explain clusters:
In a metropolis, schools naturally cluster. Every school is made up of a collection of kids
with comparable resources, environments, and educational programs.
Choose Clusters at Random:
Select a certain number of schools at random from the city's list of all the schools. Assume
they select ten colleges from a pool of fifty.
Incorporate Every Student in the Chosen Schools:
After the schools have been chosen, include all of the pupils enrolled there in the sample.
This implies that each pupil in the designated schools joins the study.
Typical Situation:
Let's say you are in the midst of a nationwide schooling survey.
States could serve as Primary Sampling Units (PSUs).
SSUs, or secondary sampling units, can be any number of cities in a state.
Within each city, there may be particular school districts or communities that serve as
Tertiary Sampling Units (TSUs).
Finally, the units picked for data gathering could be specific children or schools within those
areas or districts.
Through the use of multi-stage sampling, researchers can effectively examine sizable and
heterogeneous populations while upholding a reasonable and practical survey size at each
stage.
Assignment 2
Que 1:
Define business forecasting? Explain various methods of business forecasting?
The technique of projecting future trends or results in a company environment is known as
business forecasting. To develop well-informed predictions regarding company activities, it
entails studying past data, present circumstances, and several variables that could affect the
future. Business forecasting's main objective is to lessen uncertainty and help decision-
makers make better decisions.
There are several approaches used in business forecasting, and the method selected is
determined by a number of factors, including the type of data being forecasted, the degree of
precision needed, and the forecast's time horizon. These are a few typical techniques for
business forecasting:
3.Models of Causation:
Regression analysis is the process of identifying correlations between variables in order to
forecast values in the future.
Econometric Models: Predicting business variables through statistical techniques and
economic theories.
Technology-Oriented Forecasting
Machine learning and artificial intelligence include the use of sophisticated algorithms to
evaluate and forecast vast amounts of data.
Big Data analytics is the process of using vast amounts of data to find trends and patterns.
Market Analysis:
Questionnaires and surveys are used to get feedback and expectations from suppliers,
consumers, and other stakeholders.
Focus groups: assembling a varied group of people to talk about and offer predictions about
upcoming trends.
Leading Measures:
Leading Indicator Analysis: Tracking particular economic indicators that typically shift
before the state of the economy as a whole, offering perceptions into potential future
developments.
Combinatorial Forecasting:
Combining Methods: To increase accuracy and dependability, combine several forecasting
techniques.
2nd que:
What is index number? Discuss the utility of index numbers?
A statistical metric known as an index number is used to show how a variable or set of
related variables has changed relative to a base period or base value. It is employed to
express and quantify changes over time in a group of variables. Index numbers are a helpful
tool for data comparison and analysis, particularly when working with big datasets with
several variables.
Data Simplification:
Big data sets may be difficult to understand and intricate. Index numbers streamline the
information display and facilitate more succinct understanding and analysis of patterns.
Financial Measures:
Index numbers, which show shifts in economic activity including industrial production,
employment, and commerce, are frequently employed as economic indicators. They offer
perceptions into the general state and trajectory of an economy.
Predicting:
In order to predict future trends based on historical data, index numbers are frequently
utilized in forecasting. Businesses and policymakers can make well-informed judgments
regarding future plans by examining index patterns.
3rd que:
Discuss various types of estimators?Also explain the criteria of a good estimator?
Utilizing sample data, estimators are statistical instruments that determine parameter values
in a population. A parameter is a population's numerical feature, like the variance or mean.
Estimators yield a population parameter approximation and are computed using sample data.
A number of criteria can be used to evaluate the properties of different types of estimators.
Point Calculators:
The population parameter estimate produced by a point estimator is a single number. As an
illustration, the population mean (⁽μ) can be estimated using the sample mean (⁽̉ x ˉ).
Estimators of intervals:
A range that the true parameter is expected to fall inside is provided by an interval estimator.
Typical interval estimators are confidence intervals.
Bayesian Appraisers:
Bayes Estimators: These estimators produce a posterior distribution for the parameter of
interest by combining the likelihood function with past knowledge about the parameters.
Sturdy Estimators:
M-Estimators: These robust estimators are less susceptible to data outliers and minimize a
certain objective function.
Shrinkage Appraisers:
The Ridge Regression Estimator is a tool used in regression issues that reduces the parameter
estimates by adding a penalty term to the ordinary least squares (OLS) estimation.
Lasso Estimator: The lasso estimator use the absolute values of the coefficients in addition to
adding a penalty term, much like ridge regression.
Efficiency: An estimator that is efficient delivers estimates that are not overly dispersed since
it has a minimal sample variation. The unbiased estimator with the lowest variance is seen as
being more effective.
Sufficiency: Every piece of data in the sample required to draw conclusions about a
parameter is included in a sufficient statistic.
Robustness: An estimator that is robust is not unduly impacted by anomalies or departures
from underlying hypotheses. Being robust is especially crucial when working with real-world
data that could have anomalies in it.
Minimal Mean Squared Error (MMSE): The least mean squared error (MMSE) unbiased
estimator is deemed ideal. The average of the squared discrepancies between the estimator
and the actual parameter value is known as the mean squared error.