Inferential Statistics
Inferential Statistics
Inferential Statistics
INTRODUCTION
There are basically two branches of statistics in the business environment, and these are
the Descriptive Statistics and Inferential Statistics.
Where by the Descriptive Statistics is concerned with describing the basic features of the
data in study and provide simple summaries about the sample and the measures.
And at the same time simple graphics analysis together they form the basis of virtually
every quantitative analysis of data.
Statistical information seen in News Papers, magazines, reports, and other publications
of data that are summarized and presented in a form that is easy for the reader to
understand. Summaries of such data can be presented in tabular form, graphical form,
numerical form or test form all this is referred to as descriptive statistics. All descriptive
statistics are used to present quantitative descriptions in a well and manageable form.
That way one finds that this type of statistics reduces lots of data into a simpler summary.
Descriptive statistics is distinguished from inferential statistics where by with it, one is
concerned with trying to reach conclusions that extend beyond the immediate data done.
INFERENTIAL STATISTICS
WIkipedia, defines it as an inference about population from a random sample drawn from
it, or more generally about random process from its observed behavior during a finite
period of time which includes point estimation, interval estimation, hypothesis testing
(statistical significance testing) and production.
Lucey (2002), defines statistical inference as the process by which behaviors are drawn
about some measure or attribute of a population. For instance (mean or standard
deviation) based on analysis of sample data. Samples are taken in order to draw
conclusions about the whole population and one finds that the testing process is
destructive so that is why sampling is preferred.
The basis of statistical inference is to take a sample of a given population in study which
is analyzed so that the properties of that population is estimated from the properties of the
sample.
That is why we find that statistical inferences uses sampling distributions which show the
distribution of values expected in samples.
The purpose of sampling in statistical inference:
• Is to get reliable data by looking at a few observations rather than all possible
observations. The properties of a population are then estimated by looking at the
properties of the sample. Then later on, conclusions are made.
• Sampling helps in reducing costs especially when more information is needed,
collected, and analyzed. And when many people are involved in the surveying
process.
• Sampling reduces the time needed to collect and analyze data. And also the
amount of effort put in data collection.
• Sampling in statistical inference gives better or accurate results than when all the
population is taken. This is because it would be impossible for all the population
to give constant response.
• It is again impossible to test all the population.
• Statistical Inference uses values from samples to estimate values for the
population and this is based on simple random samples.
• Statistical Inference are used to draw conclusions that extend beyond the
immediate data alone. For instance this is used to infer from the sample data
what the population might think.
• Inferential Statistics is used to make judgments of the probability that an
observed difference between two groups is a dependable one or one that might
have happened by chance.
• Thus it is used to make inferences from our data to more general conditions.
• Statistical Inference is also useful in experimental and quasi experimental
research design or in program outcome evaluation. For example, when you
want to compare the average performance of two groups on a single measure
to see if there is a difference. One might want to know whether eight grade
boys and girls differ in maths test scores or whether a program group differs
on the outcome measure from a control group.
Whether one wishes to compare the average performance between two groups one
considers the t – test for differences between groups.
Statistical Inference about the population from a random sample drawn from it or
more generally about random process from where it is observed during a finite
period of time includes the following:
1. Point Estimation
2. Interval Estimation
3. Hypothesis Testing or Significance Testing
4. Prediction / forecasting
Below are some of the justifications of Statistical Inferences, and these are based
on the idea of what real world of phenomena can be modeled as a probability:
1. Frequency Probability
2. Bayesian Probability
3. Fiducial Probability
4. Eclectic Probability
TYPES OF INFERENCES
ESTIMATION
This one deals with the estimation of population characteristics such as population
mean and standard deviation from sample characteristics such as sample mean
and standard deviation.
The population characteristics are known as Population Parameters
And sample characteristics are known as Sample Statistics
This type of statistical inference has four properties of which are classified as
properties of good estimation:
a) Unbiased
An estimate is said to be un biased if the mean of the sample means X of all
possible random samples of size n drawn from a population of size N, equals
the population parameter U. Thus the mean of distribution of sample means
would equal the population mean.
b) Consistency
An estimate is said to be consistent if as the sample size increases, the
precision of the estimate of the population parameter also increases.
c) Efficiency
An estimate is said to be more efficient than another if in a repeated sampling
its variance is smaller.
d) Sufficiency
An estimate is said to be sufficient if it uses all the information in the sample
in estimating the required population parameter.
These are the symbols used in sample statistics and population parameters
Lucey (2002), identifies the use of the sample mean X, to make inferences about
the population mean in common and if series of samples size n (n>= 30) is taken
from a population, it will be found that:
The proportion in this aspect one finds that it represents the ratio of defective to
good production; the proportion of consumers who plan to buy a given product or
other similar piece of information.
But when n is large and np & nq are over 5, then the Binomial distribution can be
approximated by the normal distribution.
This simplifies the analysis and concepts outlined for the mean and can be applied
directly to the proportion.
Sps = √ (pq/ n)
For example a random sample of 400 passengers is taken and 55% are in favor of
the proposed new timetables.
With 95% confidence, what proportion of all rail passengers are in favor of the
timetable.
Solution
Assuming that all samples have been large, (n > 30). In this
aspect S, the sample standard deviation, is used as an estimate
of σ, the population standard deviation.
One also finds the that the distribution of sample means is approximately normal
so that the properties of the normal distribution can be used to calculate
confidence limits using the standard error of the mean.
If the properties and relationship are not true, and the sample size is small (n<30),
because the arithmetic means of small samples are not normally distributed, in
this scenario student’s t distribution must be used.
The t distribution
Characteristics of t Distribution
To develop interval estimates, for the two population small sample case, two
assumptions must be made about the population and the samples selected from the
two populations and these are:
Wherever the sample sizes are equal, the procedure here provides acceptable
results even if the population variances are not equal.
Thus a research with control of the sample sizes would consider equal sample
sizes where; n1 = n2
Fig 1.
Summary:
Statistical Inference is the process of drawing conclusions about the population from
samples.
Estimation is concerned with population parameters from sample statistics.
Where the sample size is large the sampling distribution of means is a normal
distribution.
Finding population proportions from sample information follows the usual estimation
process that np and nq are over 5 so that normal approximation can be used.
Where n < 30 the sample is small, the student’s t distribution must be used instead of the
normal distribution.
HYPTHESIS TESTING
This is yet another type of statistical inference. The hypothesis testing requires decision
makers to formulate a position or make a claim regarding decision environment they are
dealing with.
After which is sample is selected, and basing on its contents, either affirm that this
position is correct or conclude it is wrong.
In dealing with this hypothesis testing one can know how the predetermined position or
claims are formulated and how data is used to substantiate or refute the position.
Two types of errors are going to be discussed and in this very aspect devotion making
rules will be established in light of chances of making each type of error.
Lacey (2002) Defines hypothesis testing as significance testing and similar to the process
of estimation .Random sampling is involved and the properties of distribution of samples
means proportions are used.
Hypothesis testing is a belief or opinion or is the process by which the belief is tested by
statistical means. For example, from a large batch of components a random sample may
be diameter of the population of component is 50mm.Based on the results from the
sample the hypothesis would be either accepted or reflected. And the Hypothesis to be
tested is the NULL Hypothesis denoted as Ho.
Hypothesis testing helps decision makers to make decisions in which way that the
chances of decision errors can easily be controlled or at least identified.
It again provides techniques that help managers to identify and control the level of
certainly and this is due to the fact hypothesis testing does not eliminate the uncertainly
in the managerial environment.
There are quite number of questions like what happens to a certain behavior. If...? What
causes people to? Can a person s behavior be influenced? Possible explanations for the
behavior can be studied based on the previous gathered facts and theories expressed as
prediction and scientific hypothesis not always true but it is stated in much way that it can
be proved false if it is indeed false.
`
For example Victo and Viane noticed that there was a lot of violence on television and in
movies, and wondered, television and film actors known as models for children from this
question the hypothesis that was generated was : Children who view film depicting
aggressive adult models will exhibit more aggressive acts than children who see a film
depicting passive models . This hypothesis makes a production that is should be easy to
verify.
In experiments, there are two mutually exclusive hypotheses meaning that if one is true,
the other can’t be.
A researcher hypothesis researcher is one that the researcher wishes to support and on the
other hand, the Null hypothesis is the one the researcher wants to reject because it
proposes that there will be no charge in behavior no difference between groups being
measured.
And it is this Null hypothesis that is tested in research study .If the Null hypothesis is
known to be false, then the research hypothesis is supported.
Null hypothesis depends on the decision problem to be either supported or refuted
in order to reach conclusions will depend on the hypothesis being tested.
Lacey (2002) Puts down four possible results when a hypothesis test is carried out and
these are:
1. We accept a true hypothesis – a correct decision.
2. We reject a false hypothesis – a correct decision.
3. We reject a true hypothesis – an incorrect decision and this known as Type 1 error
4. We accept a false hypothesis – an incorrect decision and this is the Type 2
error.---
Groebner (1981) puts down the possible action and possible states of nature associated
with all the hypothesis – testing problem.
And these are three possible out comes.
(a) – no error.
(b) – Type 1 error
(c) – Type 2 error
Only one of the three outcomes will occur for every test of a Null hypothesis.
Though every one would wish to eliminate all chances of error ,the decision maker may
make either Type 1 error or Type 2 statistical error depending upon which decision is
selected .
If the Null hypothesis is true and an error is made , it must be a Type 1 error on the other
hand ,if the Null hypothesis is false and an error is made, it must be aType2 error.
Ho True Ho false
Reject Ho Type 1 error No error
The objective of a hypothesis testis to use sample information to decide whether to accept
or reject the Null hypothesis about the population value. How do decision makers
determine whether the sample information supports or rejects the Null hypothesis?
The answer is to compare the sample results with a predetermined decision rule.
Decision Rule:
Based on the sample:
For example Wabwire has been hired as the head of production for crepes bottler’s
company. Some soft derricks bottlers have been under pressure from consumer groups,
which claim that bottlers have been increasing the price of crepes and filling the bottlers
with less than what has been advertised. Although Wabwire feels no manufacturer would
purposely short fill the bottle , he knows that filling machines some times fail to operate
properly and fill the bottle less than full .barbwire is responsible for making sure the
filling machine at the company operate correctly , ha samples every hour an decoded on
the sample results , decodes whether to adjust the machines. If is not interested in
whether the bottles are filled with too much soft drink, can identify two possible tastes of
nature for 35- ounce bottles.
State 1 the bottles are filled with 35 or more ounces of self drink on the average.
State 2 the bottles are filled with less than 35 ounces on the average.
In the above scenario , if the all hypothesis is rejected , barbwire will half production and
have a maintenance crew adjust the filling machine to increase the average fill on the
Reject Ho A ?x = 35 Accept Ho
of probability committing a Type I error
other hand , if X is greater than or equal to A he will accept the null hypothesis , and
conclude that the filling machine e are working properly . There is a need to determine a
critical value because a decision involves determining the critical value, A.
Selecting the critical value;
Fig : 2
X , possible X values.
μx = 35
Here the distribution of possible sample means will be approximately normal with center
at the population means .
And the null hypothesis in the crepes bottler company is μx >= 35 but even if its true,
we may get a sample mean less than 35 (sampling error) .
selecting a critical value ,” A hypothesis test require a decision maker to answer
questions like, what values X of will tend to reject the null hypothesis values much
smaller than μx, values much larger than μx or values both much smaller
and much larger than μx?
The chances of committing a type 1 error can be reduced if the critical value A , is more
further to the left of (μx >= 35) as above .
In order to determine the appropriate value for A , decision makers must determine how
large an they want .this decision makers must select the value of in light of the costs
thawed in committing a type 1error .
For example if Wabwire rejects the null hypothesis when it’s true, he will shut down the
production and incur the costs of machine adjustments. This can even affect the future
production, so calculate these costs and determine the probability of incurring them in am
management decision
These stapes must be followed in order to test any null hypothesis: presses:
Time series use statistical analysis on past data arranged in time series for example
– sales by month for the last ten years
– Annual production of agriculture products over twelve month.
a) Are the past data representatives? For example do they contain the results of
recession boom shift of taste?
b) Time series methods are appropriate where short term forecasts are required.
c) Time series methods best limited to stable situations. Where
fluctuations are common and are expected to change, then may give poor results.
- Helps to understand the past behavior which gives a production for the future
for which is statistically important for business planning.
- Gives information which lets as a base for comparison in values of different
phenomenon at different times.
- Helps in evaluation of current achievements.
- It helps in interpreting and evaluating changes in economic phenomenon in the
hope of more correctly anticipating the course of future events.
The trend components in the long term increase or decrease in a variable being measured
over time
Today, organization is faced with increased planning problems caused by changing
technology, government regulations and uncertain foreign competition. A combination of
these forces most organizations with increasing the time span of their planning cycle.
And due to fact that long term forecasting is increasing, the trend component in time
series analysis is important to all organizations.
Seasonal component
Some organizations or industries are affected by seasonal variations and not only long
term trends.
The seasonal component represents these changes in time every year.
Organization affected by seasonal variation need to identify and measure the seasonality
to help with planning for temporally increases or decreases in labor requirements,
inventory , training, periodic maintenance and many others.
Organization need to know if seasonal variations they experience occur at more or less
than the average.
Cyclical component
Cyclical effects in time series are represented by wave like fluctuations around long term
trend. These fluctuations are caused by factors such as interest rates, supply, consumer
demand, inventory levels, national and international market conditions and government
policies.
The cyclical fluctuations repeat themselves but occur with differing frequencies and
intensities. So each though one knows what happened to the firm during the last cycle, he
or she has no guarantee the effect will be the same as the next time.
These are some of the methods considered moving average and Exponential smoothing.
In this aspect, if forecast for next months sales say for December, was the actual sales for
November, then the forecasts obtained would fluctuation. If forecasts for the next months
sales for several preceding months then random fluctuations would cancel each other.
This is the principle of moving average method.
Illustration:
Past Sales
Month Actual sales 3 monthly 6 monthly 12 monthly
units moving average average moving average
January 500
February 650
March 420
April 600 523.33
May 350 556.66
June 840 456.66
July 920 596.66 560
August 950 703.33 630
September 800 903.33 680
October 750 890 743
November 850 833.33 768.33
December 930 800 851.66
January 990 843.33 866.66
Any month’s forecast is the average of the proceeding al in actual sales. For example the
3 months moving average forecasts were prepared as follows:
= 1570
3
= 523.33
=1670
3
=556.66
And for a six months moving average forecasts were prepared as follows:
July forecasts = Jan sales + Feb sales + Mar sales + April sales +May sales + June sales
6
= 3360
6
= 560
- Equal weighing is given to each of the values used in the moving average
calculation, where as it is reasonable to impose that the most recent data is
more relevant to current conditions.
- Moving average calculation takes no account of data out side the period of
average, no full use is not made of all the data available.
- The use of unadjusted moving average as forecasts can cause misleading results
when there is underlying variation.
- An n period moving average requires the storage of n – 1 value to which is
added the latest observation.
Exponential Smoothing: Method
Lucey: (2002). Asserts that this is a frequently forecasting technique which largely over
comes the limitations of the moving average method. It involves the automatic weighing
of past data with weights that decreases exponentially with time. Meaning that the most
current values receive the greatest weighing and the older observations receive a
decreasing weighing. Exponential smoothing method is a weighed moving average
system having the principle that;
Note:
1. Because no previous forecasts were available Jan. Sales were used as
February forecasts.
2. Formula used when α = 0.1, March forecasts = Feb forecasts + 0.1
(Feb sales – February Forecasts)
= March forecasts = 450 + 0.1 (440 -450)
= March forecasts = 449
Summary
- Time series is based on data arranged in regular time periods e.g. sales per month
- Moving average system is based on any number of periods. Say 3 months, 6 month or
storage.
-The key factor in exponential smoothing is the choice of smoothing constant α. The
higher the value, the more responsive is the system to the current condition.
PROBABILITY:
Definition
Salem: (1997: 253). Asserts that probability is the ratio of the number of favorable cases
to the total number of equally likely cases
That is if there are severally equally likely events that may happen, the probability that
any one of these events will happen, is the ratio of the number of cases favorable to it’s
happening to the total number of possible cases.
Lucey: (2002:8).Asserts that probability can be considered as the qualification of
uncertainty where by an uncertainty is expressed as like “hood “chance or “risk “
Probability is expressed by P and takes values ranging from zero to and one meaning that
zero is an impossibility and one is certain for example P (crossing the ocean unaided ) = 0
band P(dieing ) = I
Approaches of probability
There are compulsions about probability due to the reason that, it means different
things to different people. there are four basic methods or approaches to probability
and these are:
i. Relative frequency of occurrence or empirical approach
ii. Subjective probability assessment or personal statistic approach
iii. Classical probability assessment.
iv. Axiomatic approach.
Relative frequency of occurrence
This approach is based on actual observation and it borrows the concept of relative
frequency which implies the probability of that the probability of an event is given by the
frequency of that event relative to the total number of trials.
For example if one is the probability of ten customers or fewer customers actually did
arrive before 8.00 am. The probability assessment would be the ratio of days when ten or
few customers arrived to the total number of days observed,
This approach of probability is not directly applicable to business decision making as the
subjection and relation frequency methods if there is A possible outcome favorable to the
occurrence of an event E and B possible out come unfavorable to the occurrence of E
,and all these possible outcomes one equally likely and mutually exclusive then this
probability that the event E will occur, denoted by
P (E) =A / A+B
Limitation of classical approach:
The definition is not applicable when the assumption of equally likely does not
hold.
The definition becomes vague when the number of possible outcomes may be
infinite.
It may be difficult to determine the values or the numerator and denominator
Axiomatic approach:
Before these rules are considered mutually exclusive events and independent event must
cross the minds of the reader.
Independent event:
Two or more events are independent if the occurrence or none. Occurrence of any one
event does not affect the occurrence or non occurrence of the others. E.g. the outcome of
any throw of a die is independent of the outcome of any proceeding or succeeding
outcome.
If the probability of an event takes the highest value of 1 expressed P(A)= 1 it means
that the event is certain.
Expressed as follows;
P(AnB)=P(A/B)xP(B)
P(BnA)=P(B/A)x P(A).
This rule is used when there is a string of independent events for which each
individual probability is known and it is required to know the overall probability.
This rule is concerned with calculating the probability of two or more mutually
exclusive events. Such that the probabilities of the separate events are added.
For example the probability of showing a 3 or a 6 die would be expressed.
P(throwing 3) = 1/6 and P(throwing a 6) = 1/6 .
P(throwing a 3 or a 6) 1/6 + 1/6 = 1/3 .
Other Probability Rules:
This rule requires that conditional probability be had since the result on the second
draw depends on the card selected on the first draw.
Expressed as:
P(E1 and E2) = P(E1) P(E2).
Baye’s Rule:
Summary:
In this topic the consideration is put on identifying the processes that are represented by
discrete distributions in general and by binomial and Poisson distributions in particular.
Again the probabilities associated with particular outcomes in discrete distribution are
considered.
Determine the mean and standard deviation for general discrete distributions and for the
binomial and Poisson distribution.
These are the major concepts used in probability distributions and one needs to fully
understand them when dealing with this topic.
• Random Variable
• Probability function
• Expected value and variance
Random Variable
Saleem (1997), defines a random variable that takes specified values with specified
probabilities. The probabilities specified in such a way that the random experiment is
conducted and the way in which the variable is defined and observed on the random
experiment.
Capital letters are used to denote random variables and corresponding small letters to
represent any specified value of the random variable.
Groebner (1981:123), asserts that a random variable is a variable whose numerical value
is determined by the outcome of a random experiment or trial.
A discrete random variable is a random variable that assumes only distinct values for
example: if a manager examines 10 accounts the number of inaccurate balance can be
represented by the value X. then x is a random variable with values {0,1,2,3,…,10}
Continuous random variables are ones which assume any value on a continuum. For
example time is continuous.
It is noted that discrete probability distribution is an extension of relative frequency
distribution for example DELL computers Limited, each week offers specials on to 5
specific computers as part of sales. For a period of 40 weeks, the sales manager recorded
how many of the 5 computers were sold each week, as shown in the table.
P(x) X. (number of computer sold)
Computers Sold
In this aspect the probability of an outcome or (value of a random variable) occurring can
be assessed by the relative frequency of that outcome.
The probability distribution must add to one (1) and can be shown in a graphical form as
below.
Probability Distribution
Decision makers need to calculate the distribution’s mean, and standard deviation and
these values measure the central location and spread respectively.
This is also called the expected value of the discrete random variable. And the expected
value is the weighted average of the random variable values where the weights are the
probabilities assigned to the values.
Formulae
E (x) = ∑ x p(x)
Where E (x) = Expected value of x
X = value of the random variable
P(x) = probability of each value of x
This measures the spread, or dispersion, in a set of data. It again measures the spread in
the values of a random variable calculation of standard deviation for a discrete
probability distribution.
1. Binomial Distribution
2. Poisson distribution
3. Normal Distribution
BINOMIAL DISTRIBUTION
Expand the binomial (q+ p) n. the power n is equal to one less than the number of terms in
the expanded binomial. Thus when n =2, there will be three terms in the binomial and if n
= 4, there will be five terms.
Multiply each term of the expanded binomial by N (the total frequency) in order to obtain
the expected frequency in each category.
Comments about the Binomial Distribution
When the value of p differs from 0.5 in either decision, the Binomial Distribution is
skewed.
In using binomial distribution, one is able to count the number of successes and the
number of failures. But in other applications, the number of successes may be counted
and the number of failures may be difficult to count.
If the total number of possible outcomes cannot be determined, then the binomial
distribution cannot be applied in decision making aid.
This calls for the application of Poisson distribution which can be applied in situations
without knowing the total possible outcomes. In order to apply the Poisson distribution
one needs to know the average number of success for a given segment.
These are the conditions that must be satisfied before applying the Poisson distribution.
A physical situation must have certain characteristics before it can be described byth
Poisson distribution.
7. Poisson distribution has either one or two modes (like the binomial distribution)
when m is not an integer, mode is the largest integer contained in m. when m is an
integer, there are two modes that is m and m – 1.
8. The distribution is positively skewed to the left. With an increase in value of the
mean m, the distribution shifts to the right and the skewness diminishes. Poisson
distribution differs from binomial distribution in two ways:
a. Poisson distribution operates continuously over a given period of time,
distance, area.
b. Poisson distribution produces success which occurs at random points in
the specified time, distance, and area. And these successes are called
occurrences.
1. Number of customers arriving at service facility in unit time for instance per hour
2. Number of telephone calls arriving at telephone switchboard per unit time for
instance per minute
3. Number of defects along a tape
4. Dimensional errors in engineering drawings
5. Number of radioactive particles decaying in a given interval of time
6. Number of printing mistakes per page in a book
7. Number of accidents on a particular road per day
8. Hospital emergencies per day
9. Number of defective materials of products, say pins
10. Number of goals in a football match
Example
i) P(x = 0) and
ii) P(x >2)
Given that e-6 = 0.00248
Solution
Here one needs to obtain the values of m and calculate the probability of zero
occurrences. The other probabilities will be calculated by recurrence relation as follows:
F(x) e-mm x / x!
N.B. The variance for the Poisson distribution will always equal to the mean
σx2 = λt
Standard deviation of the Poisson distribution is the square root of the mean
σx = √ λt
Where a Poisson distribution applies, the uncertainty can be controlled by controlling the
mean, which must be within the decision maker’s control.
Summary:
In this chapter, discrete random variable concepts have been introduced and showed how
probability distribution is developed for a discrete random variable. The computation of
mean, and standard deviation for discrete distribution have been considered.
Binomial and Poisson distributions represent two of the most commonly used theoretical
distributions. These distributions are used in a number of managerial applications.
Some concepts connected with discrete distributions from a managerial perspective have
been dealt with such as random variable, probability function and expected value and
variance.
NORMAL DISTRIBUTION
This is the mother type of discrete probability distribution which is associated with the
names of Laplace and Gauss and is called Gaussian distribution.
In this aspect, one finds that whether or not p is equal to q, the binomial distribution
forms a continuous curve when n becomes large. This correspondence between binomial
and the normal curve is close even for low values of n provided p and q are fairly near
equality. The limiting curve, obtained as n becomes large and is called the normal
frequency or the normal curve.
The y = f(x) is a bell shaped curve. And the top of the bell is directly
above the mean μ. For large values of σ, the curve tends to flatten out
and for small values of σ it has a sharp peak.
?
Since there is only one point of maximum frequency (at the mean) the normal distribution
is uni-modal.
The mean deviation is 4/5 SD
Quatiles are given by
Q1 = μ – 0.6745 σ
Q3 = μ + 0.6745 σ
The points of inflexion (the points at which the curve changes its direction) are each at a
distance of one standard deviation from the mean.
The curve is asymmetric to the base time i.e. continues to approach but never touches the
base line. No portion of the curve lies below the base line.
The percentage distribution of area under standard normal curve is broadly shown below.
Here we have:
• Useful in statistical quality control where the control limit are set by using this
distribution
• By virtue of the central limit Theorem the distribution of the means of samples
taken from any population which need not be normal tends towards the normal
distribution if the sample is large.