SigmaXL Version 8 Workbook
SigmaXL Version 8 Workbook
SigmaXL Version 8 Workbook
0 ®
Workbook
Contact Information:
Technical Support: 1-866-475-2124 (Toll Free in North America) or
1-519-579-5877
Sales: 1-888-SigmaXL (888-744-6295)
E-mail: [email protected] Web: www.SigmaXL.com
iii
SigmaXL: Table of Contents
iv
SigmaXL: Table of Contents
Basic Statistical Templates – 1 Sample Chi-Square Test and CI for Standard Deviation ....... 70
Basic Statistical Templates – 2 Sample F-Test and CI (Compare 2 Standard Deviations)...... 71
Basic Statistical Templates – 1 Proportion Test and Confidence Interval ............................. 72
Basic Statistical Templates – 2 Proportions Test and Confidence Interval ........................... 74
Basic Statistical Templates – 2 Proportions Equivalence Test............................................... 76
Basic Statistical Templates – 1 Poisson Rate Test and Confidence Interval .......................... 78
Basic Statistical Templates – 2 Poisson Rates Test and Confidence Interval ........................ 80
Basic Statistical Templates – 2 Poisson Rates Equivalence Test............................................ 82
Basic Statistical Templates – One-Way Chi-Square Goodness-of-Fit Test............................. 84
Basic Statistical Templates – One-Way Chi-Square Goodness-of-Fit Test - Exact ................. 85
Probability Distribution Calculators – Normal ....................................................................... 88
Basic MSA Templates – Type 1 Gage Study ........................................................................... 89
Basic MSA Templates – Gage Bias and Linearity Study ......................................................... 91
Basic MSA Templates – Gage R&R Study (MSA) .................................................................... 93
Basic MSA Templates - Gage R&R: Multi-Vari & X-bar R Charts............................................ 95
Basic MSA Templates – Attribute Gage R&R (MSA) .............................................................. 96
Basic Process Capability Templates – Process Sigma Level – Discrete Data ......................... 98
Basic Process Capability Templates – Process Sigma Level – Continuous Data .................... 99
Basic Process Capability Templates – Process Capability Indices........................................ 100
Basic Process Capability Templates – Process Capability & Confidence Intervals .............. 101
Basic Process Capability Templates – Tolerance Interval Calculator (Normal Exact).......... 102
Basic Control Chart Templates – Individuals Chart.............................................................. 103
Part C – Descriptive/Summary Statistics ................................................................................... 106
Descriptive Statistics ............................................................................................................ 106
Descriptive Statistics - Options ............................................................................................ 108
Part D – Histograms ................................................................................................................... 115
Basic Histogram Template ................................................................................................... 115
Single (Basic) Histogram....................................................................................................... 115
Multiple Histograms............................................................................................................. 116
Part E – Dotplots ........................................................................................................................ 118
Dotplots................................................................................................................................ 118
Part F – Boxplots ........................................................................................................................ 119
v
SigmaXL: Table of Contents
vi
SigmaXL: Table of Contents
vii
SigmaXL: Table of Contents
viii
SigmaXL: Table of Contents
Computing Odds Ratio and Confidence Interval for 2x2 Table ........................................... 304
Chi-Square Test – Two-Way Table Data: Advanced Tests and Measures of Association – Ordinal
Categories ............................................................................................................................ 305
Chi-Square Test (Stacked Column Format Data) ................................................................. 310
Chi-Square Test – Fisher’s Exact .......................................................................................... 315
Chi-Square Test – Two Way Table Data – Fisher’s Exact ..................................................... 319
References for Fisher’s Exact ............................................................................................... 323
Power & Sample Size for One Proportion Test .................................................................... 324
Power & Sample Size for Two Proportions Test .................................................................. 326
Part P – Analysis of Means (ANOM) Charts ............................................................................... 328
Introduction – What is ANOM ............................................................................................. 328
ANOM Normal One-Way ..................................................................................................... 332
ANOM Normal Two-Way (with Main Effects and Slice Charts) ........................................... 334
ANOM Binomial Proportions One-Way ............................................................................... 342
ANOM Binomial Proportions Two-Way ............................................................................... 344
ANOM Poisson Rates One-Way ........................................................................................... 346
ANOM Poisson Rates Two-Way ........................................................................................... 348
ANOM Nonparametric Transformed Ranks......................................................................... 350
ANOM Variances .................................................................................................................. 354
ANOM Levene Robust Variances ......................................................................................... 356
Part Q – Multi-Vari Charts .......................................................................................................... 358
Multi-Vari Charts .................................................................................................................. 358
Part R – Scatter Plots.................................................................................................................. 362
Scatter Plots ......................................................................................................................... 362
Scatter Plot Matrix ............................................................................................................... 364
Part S – Correlation Matrix ........................................................................................................ 365
Correlation Matrix................................................................................................................ 365
Reference ............................................................................................................................. 366
Part T – Multiple Regression ...................................................................................................... 367
Multiple Regression ............................................................................................................. 367
Multiple Regression with Continuous and Categorical Predictors ...................................... 371
Part U – Logistic Regression ....................................................................................................... 373
ix
SigmaXL: Table of Contents
x
SigmaXL: Table of Contents
xi
SigmaXL: Table of Contents
xii
SigmaXL: Table of Contents
xiii
SigmaXL® What’s New in Version 8.0, System
Requirements and Getting Help
See Appendix: SigmaXL Version 8.0 Feature List Summary for a complete feature list summary.
New features in SigmaXL Version 8.0 include:
• Descriptive Statistics
o New features are available by clicking the dialog Options button
o Percentile Report
27 values from 0.135 to 99.865
o Percentile Ranges
75 - 25 (50%, Interquartile Range IQR)
90 - 10 (80%, Interdecile Range IDR)
95 - 5 (90%, Span)
97.5 - 2.5 (95%, +/- 1.96 Sigma Equivalent)
99 - 1 (98%)
99.5 - 0.5 (99%)
99.865 - 0.135 (99.73%, +/- 3 Sigma Equivalent)
o Percentile Confidence and Tolerance Intervals
Interpolated or Exact
Minimum sample size reported if unable to compute CI or TI
Quartile Confidence Intervals (25, 50, 75)
Percentile Confidence Intervals (27 values from 0.135 to 99.865)
Percentile Tolerance Intervals (50%, 80%, 90%, 95%, 98%, 99%, 99.73%)
o Additional Descriptive Statistics
5% Trimmed Mean
Standard Error of Mean
Variance
Coefficient of Variation
Short Term StDev (MR-bar/d2)
2
SigmaXL: What’s New, Installation Notes, Getting Help and Product Registration
3
SigmaXL: What’s New, Installation Notes, Getting Help and Product Registration
o Binomial Proportions Two-Way with Main Effects and Wludyka & Noguera Slice
Charts
In collaboration, Peter Wludyka and John Noguera of SigmaXL extended the
Slice Charts to Binomial and Poisson.
Interaction P-Value automatically determined from Logistic regression
o Poisson Rates One-Way
Warning is given if normal approximation to Poisson is not applicable
(nu <= 5)
o Poisson Rates Two-Way with Main Effects and Wludyka & Noguera Slice Charts
Interaction P-Value automatically determined from Poisson regression
o Nonparametric Transformed Ranks
o Variances
o Levene Robust Variances
• One-Way ANOVA
o New features are available by clicking the dialog Options button
o Multiple Comparison of Means Probability Methods (a.k.a. Post-Hoc)
Fisher
Tukey
Dunnett with Control
o Display ANOM Normal One-Way Chart
o Confidence Level
o Residual Plots
• Two-Way ANOVA
o Display ANOM Normal Two-Way Chart
4
SigmaXL: What’s New, Installation Notes, Getting Help and Product Registration
5
SigmaXL: What’s New, Installation Notes, Getting Help and Product Registration
Activating SigmaXL
Please proceed with the following steps if you have a valid serial number and your computer is
connected to the Internet. If you do not have an Internet connection, please email
[email protected] from a different system or call 1-888-SigmaXL (1-888-744-6295) or 1-
519-579-5877. If you purchased SigmaXL as a download, you received the serial number by e-
mail. If you purchased a CD, the serial number is on the label of the CD.
If your trial has timed out and you do not have a serial number but wish to purchase a SigmaXL
license, please click Purchase SigmaXL in the Activation Wizard Box and this will take you to
SigmaXL’s order page http://www.sigmaxl.com/Order%20SigmaXL.shtml. You can also call 1-
888-SigmaXL (1-888-744-6295) or 1-519-579-5877 to place an order.
1. In the Activation Wizard box select Activate SigmaXL (Enter a serial number.)
2. Click Next. Enter your serial number. We recommend that you copy and paste the serial
number into the serial number field to avoid typos.
6
SigmaXL: What’s New, Installation Notes, Getting Help and Product Registration
4. Click OK. In Excel 2007 and newer, SigmaXL appears as a new Ribbon:
7
SigmaXL: What’s New, Installation Notes, Getting Help and Product Registration
Error Messages
1. Incorrect Serial Number:
This is due to an incorrect serial number entry in the previous registration window – click OK
and please reenter your serial number.
The license is currently activated on the maximum allowable number of computers. You will
need to deactivate SigmaXL from your old computer while connected to the Internet. To
deactivate a SigmaXL license go to SigmaXL > Help > Deactivate SigmaXL.
If you do not have an Internet connection, please email [email protected] from a different
system or call 1-888-SigmaXL (1-888-744-6295) or 1-519-579-5877.
8
SigmaXL: What’s New, Installation Notes, Getting Help and Product Registration
Click SigmaXL > Help > SigmaXL Defaults > Clear Saved Defaults. A warning message is given prior
to clearing saved defaults.
This can be changed to: Always use my pre-selected data range without prompting. ‘Data Labels’
will be used. This setting saves you from having to click Next at the start of every function, but the
user is responsible to ensure that the proper data selection is made prior to starting any menu
item.
Click SigmaXL > Help > SigmaXL Defaults > Data Selection Default to make this change. This will
apply permanently unless you revert back to the Prompt me setting or click Clear Saved Defaults
shown above.
9
SigmaXL: What’s New, Installation Notes, Getting Help and Product Registration
Click SigmaXL > Help > SigmaXL Defaults > Menu Options – Set SigmaXL’s Menu to Classical or
DMAIC. The Set Menu dialog allows you to choose between Classical (default) and DMAIC:
If you select the DMAIC Menu, the DMAIC Menu Ribbon appears as shown (Excel 2007 and newer):
All SigmaXL tools are available with this menu format, but they are categorized using the Six Sigma
DMAIC phase format. Note that some tools will appear in more than one phase.
This workbook uses the classical (default) menu format, but the chapters are organized as
Measure, Analyze, Improve and Control.
10
SigmaXL: What’s New, Installation Notes, Getting Help and Product Registration
Windows Operating System: Microsoft Windows 7 with current Service Packs, or later operating
system.
Microsoft Windows Excel version: Excel 2007 and newer with latest service packs installed.
Mac Excel Version: Excel 2011 or later with latest service packs installed.
11
SigmaXL: What’s New, Installation Notes, Getting Help and Product Registration
To access the help system, please click SigmaXL > Help > Help.
Please note that users obtain free technical support and upgrades for one year from date of
purchase. Optional maintenance is available for purchase prior to the anniversary date.
To register by web, simply click SigmaXL > Help > Register SigmaXL.
12
Introduction to SigmaXL®
Data Format and Tools Summary
Introduction
SigmaXL is a powerful but easy to use Excel Add-In that will enable you to Measure, Analyze,
Improve and Control your service, transactional, and manufacturing processes. This is the ideal
cost effective tool for Six Sigma Green Belts and Black Belts, Quality and Business Professionals,
Engineers, and Managers.
SigmaXL will help you in your problem solving and process improvement efforts by enabling you to
easily slice and dice your data, quickly separating the “vital few” factors from the “trivial many”.
This tool will also help you to identify and validate root causes and sources of variation, which then
helps to ensure that you develop permanent corrective actions and/or improvements.
X Process Y
The mathematical expression Y = f(X) denotes that the variable Y is a function of X. Y can also be
viewed as the effect of interest and X is the cause. For example, Y could be customer satisfaction
as measured on a survey and X could be location or responsiveness to calls (also measured on a
survey). The goal is to figure out which X’s from among many possible are the key X’s and to what
extent do they impact the Y’s of interest. Solutions and improvements then focus on those key X’s.
15
SigmaXL: Introduction to Data Format and Tools Summary
Note that in SigmaXL, a discrete X can be text or numeric, but a continuous X must be numeric. Y’s
must be numeric. If Y is discrete, count data will be required. If the data of interest is discrete text,
it should be referenced as X1 and SigmaXL will automatically search through the text data to obtain
a count (applicable for Pareto, Chi-Square and EZ-Pivot tools).
16
SigmaXL: Introduction to Data Format and Tools Summary
If the data is in unstacked multiple column format, each unique group of X corresponds to a
different column. The above data is now shown in unstacked format with customer satisfaction
scores for each customer type in separate columns:
17
Tool What Type of Data When to Use Location in SigmaXL
Pareto Chart Plots a bar chart of the response in Y=Discrete (e.g., To separate the vital few from the trivial SigmaXL > Templates & Calculators >
descending order with a cumulative Defect Count) or many, help specify a problem statement, and Basic Graphical Templates > Pareto Chart
sum line. Continuous (e.g., Cost; prioritize potential root causes. This chart is SigmaXL > Graphical Tools > Basic
must be additive) based on the Pareto principle, which states Pareto Chart (Single)
X=Discrete that typically 80% of the defects in a process SigmaXL > Graphical Tools > Advanced
(Category) or product are caused by only 20% of the Pareto Charts (Multiple)
possible causes.
Pivot Chart Plots a stacked (or clustered) bar Y=Discrete or To easily ‘slice and dice’ your data, quickly SigmaXL > Graphical Tools > EZ-
chart from an Excel Pivot Table. Continuous look at different X factors and their Pivot/Pivot Charts
X = Discrete contribution to the total. It is similar to the
(Category) Pareto Chart, but without the descending bar
order.
Histogram Visual display of one variable Y=Continuous 1. Summarize large amounts of data SigmaXL > Templates & Calculators >
showing data center, spread, shape X=Discrete 2. To get a ‘feel for the data’ Basic Graphical Templates > Histogram
and outliers. (Category) 3. To compare actual description to SigmaXL > Graphical Tools > Basic
customer specifications Histogram (Single)
SigmaXL > Graphical Tools > Histograms
& Descriptive Statistics
SigmaXL > Graphical Tools > Histograms
& Process Capability
Dotplots Visual display of one variable Y=Continuous 1. Small sample size (n < 30) SigmaXL > Graphical Tools > Dotplots
showing data center, spread, shape X=Discrete 2. To get a ‘feel for the data’
and outliers. (Category)
Boxplots Visual display of the summary of Y Y=Continuous Summary display to visualize differences in SigmaXL > Graphical Tools > Boxplots
data grouped by category of X. X=Discrete data center, spread and outliers across
18
(Category) categories.
Normal Plots data in a straight line if the Y=Continuous To check for Normality and Outliers. SigmaXL > Graphical Tools > Normal
Probability Plot data is normally distributed. X=Discrete Probability Plots
(Category)
SigmaXL: Introduction to Data Format and Tools Summary
Run Charts Plots observations in time sequence Y=Continuous or To view process performance over time for SigmaXL > Templates & Calculators >
Discrete trends, shifts or cycles. Basic Graphical Templates > Run Chart
SigmaXL > Graphical Tools > Run Chart
To test for Randomness using the SigmaXL > Graphical Tools > Overlay
Nonparametric Runs Test Run Chart
Multi-Vari Charts Plots vertical lines with dots to allow Y=Continuous To visually compare subgroups by individual SigmaXL > Graphical Tools > Multi-Vari
comparison of subgroups on one X=Discrete data points and the mean. Charts
variable. (Category) To identify major sources of variation (e.g.,
Summary of Graphical Tools
Analysis of Plots response Y mean for each Y=Continuous ANOM is a complement to ANOVA showing SigmaXL > Graphical Tools > Analysis of
Means (ANOM) level of X category with decision X=Discrete which group means are significantly different Means (ANOM)
limits. (Category) than the grand mean.
Control Charts Plots observations in time sequence Y=Continuous or To monitor the process over time for trends, SigmaXL > Control Charts >
against a mean and control limits. Discrete shifts or cycles in order to control and
improve process performance.
t-Test Determine if there is a significant Y=Continuous 1. Test if mean = specified value SigmaXL > Templates & Calculators > Basic Statistical Templates >
difference between two group X=Discrete (Category) 2. Test if 2 sample means are equal 1 Sample t-Test and Confidence Interval for Mean
means or if the true mean of the 3. Paired t: to reduce variation when comparing 2 Sample t-Test and Confidence Interval (Compare 2 Means)
data is equal to a standard value. two sample means
4. Multiple pairwise comparisons SigmaXL > Statistical Tools >
1 Sample t-Test & Confidence Intervals / Paired t-Test
2 Sample t-Test / 2 Sample Comparison Tests
One-Way ANOVA & Means Matrix
One-Way ANOVA (Analysis Determine if there is a difference Y=Continuous Determine if there is a statistically significant SigmaXL > Statistical Tools > One-Way ANOVA & Means Matrix (for
of Variance) in mean among many groups. X=Discrete (Category) difference in means among the groups. equal variance)
SigmaXL > Statistical Tools > Equal Variance Tests > Welch’s ANOVA
(for unequal variance)
Two-Way ANOVA (Analysis Determine if there is a difference Y=Continuous Determine if there is a statistically significant SigmaXL > Statistical Tools > Two-Way ANOVA
of Variance) in mean among many groups for X1=Discrete (Category) difference in means among the groups.
two factors plus their interaction. X2=Discrete (Category)
Nonparametric Tests Determine if there is a difference Y=Continuous 1. Test if median = specified value: SigmaXL > Statistical Tools > Nonparametric Tests
between two or more group X=Discrete (Category) 1 Sample Sign Test or Wilcoxon
medians or if the median of the 2. Test if 2 sample medians are equal: SigmaXL > Statistical Tools > Nonparametric Tests – Exact (for small
data is equal to a standard value. 2 Sample Mann-Whitney sample)
3. Test if there is a difference in medians among
the groups: Kruskal-Wallis or Mood’s Median
CI for Standard Deviation/ Determine if there is a difference Y=Continuous 1. Determine the confidence interval for a single SigmaXL > Templates & Calculators > Basic Statistical Templates >
F-test / between two or more group X=Discrete (Category) standard deviation. 1 Sample Chi-Square Test and CI for Standard Deviation
Bartlett’s Test/ Levene’s Test variances or determine the 2. Test if 2 sample variances (standard 2 Sample F-Test and CI (Compare 2 Standard Deviations)
confidence interval of a single deviations) are equal.
standard deviation. 3. Determine if there is a statistically significant SigmaXL > Statistical Tools > Two Sample Comparison Tests
difference for the variances among the
groups. Use Bartlett’s test for normal data. SigmaXL > Statistical Tools > Equal Variance Tests >Bartlett / Levene
19
Use Levene’s test for non-normal data.
Proportions Test Determine if there is a difference Y=Discrete (Proportion) 1. Determine the confidence interval for a single SigmaXL > Templates & Calculators > Basic Statistical Templates >
between two proportions or X=Discrete (Category) proportion. 1 Proportion Test & Confidence Interval
determine the confidence interval 2. Determine if there is a statistically significant 2 Proportions Test & Confidence Interval
of a single proportion. difference for two proportions.
Poisson Rate Test Determine if there is a difference Y=Discrete (Count) 1. Determine the confidence interval for a single SigmaXL > Templates & Calculators > Basic Statistical Templates >
between two rates or determine X=Discrete (Category) rate. 1 Poisson Rate Test and Confidence Interval
the confidence interval of a single 2. Determine if there is a statistically significant 2 Poisson Rates Test and Confidence Interval
rate. difference for two rates.
One-Way Chi-Square Determine if the observed Y=Discrete (Count) Test the distribution of observed frequency counts SigmaXL > Templates & Calculators > Basic Statistical Templates >
Goodness-of-Fit frequencies for one discrete against expected (typically uniform discrete). One-Way Chi-Square Goodness-of-Fit Test
variable are distributed equally One-Way Chi-Square Goodness-of-Fit Test Exact (for small sample)
Χ2 Determine if there is a difference Y=Discrete (Count) Determine if there is a relationship between two SigmaXL > Statistical Tools > Chi-Square Tests >
Chi Square for observed frequencies of two X=Discrete (Category) discrete variables. Chi-Square Test (for raw discrete data in stacked column format)
discrete variables. Chi-Square Test – Two Way Table Data (for pivot or contingency table)
Summary of Statistical Tools
SigmaXL > Statistical Tools > Chi-Square Tests – Exact (for small
sample)
Anderson-Darling Normality Determine if the data is normally Y=Continuous Test if the sample data is normally distributed. SigmaXL > Graphical Tools > Histograms & Descriptive Statistics
Test distributed.
SigmaXL > Statistical Tools > Descriptive Statistics
Correlation Quantify strength of relationships. Y=Continuous Determine if there is evidence of a relationship SigmaXL > Statistical Tools > Correlation Matrix
X=Continuous between Xs and Ys, quantify the relationship,
identify root causes.
Regression (Simple Linear & Summarizes, describes, predicts Y=Continuous 1. Determine if there is evidence of a relationship SigmaXL > Statistical Tools > Regression > Multiple Regression
Multiple) and quantifies relationships. X=Continuous between Xs and Ys.
2. Model data to develop a mathematical
Logistic Regression Summarizes, describes, predicts Y=Discrete (Binary or SigmaXL > Statistical Tools > Regression >
equation to quantify the relationship.
and quantifies relationships. Ordinal) Binary Logistic Regression
3. Identify root causes.
X=Continuous Ordinal Logistic Regression
4. Make predictions using the model.
Design of Experiments Systematic and efficient proactive Y=Continuous or Discrete To establish cause and effect relationship between SigmaXL > Design of Experiments
SigmaXL: Introduction to Data Format and Tools Summary
(DOE) approach to testing relationships. X=Continuous or Discrete Ys and Xs. To identify ‘vital few’ Xs.
SigmaXL: Measure Phase Tools
Note that Loyalty, Overall Satisfaction, Responsive to Calls, Ease of Communications, and Staff
Knowledge were obtained from surveys. A Likert scale of 1 to 5 was used, with 1 being very
dissatisfied, and 5 very satisfied. Survey results were averaged to obtain non-integer results.
Category Subset
1. Click SigmaXL > Data Manipulation > Category Subset.
2. If you are working with a portion of a dataset, specify the appropriate range, otherwise check
Use Entire Data Table.
3. Click Next.
23
SigmaXL: Measure Phase Tools
5. Click OK.
Random Subset
1. Click Sheet 1 Tab of Customer Data.xlsx.
3. Ensure that the entire data table is selected. If not, check Use Entire Data Table. Click Next.
4. Enter Number of Rows in Random Subset as 30. The default Sort Data selection is Original
Order.
24
SigmaXL: Measure Phase Tools
5. Click OK. A new worksheet is created that contains a random subset of 30 rows.
This feature is useful for data collection to ensure a random sample, e.g., given a list of
transaction numbers select a random sample of 30 transactions.
Numerical Subset
1. Click Sheet 1 Tab of Customer Data.xlsx.
3. Ensure that entire data table is selected. If not, check Use Entire Data Table. Click Next.
5. Click OK.
A new subset worksheet is created containing only those rows with Overall Satisfaction >= 4.
Date Subset
1. Click Sheet 1 Tab of Customer Data.xlsx.
3. Ensure that entire data table is selected. If not, check Use Entire Data Table. Click Next.
25
SigmaXL: Measure Phase Tools
4. Select Order Date, select 1/9/2016, click Start Date, select 1/12/2016, click End Date.
5. Click OK. A new subset worksheet is created containing only those rows with Order Date
between 1/9/2016 to 1/12/2016.
Transpose Data
1. Open Catapult Data Row Format.xlsx.
2. Manually select the entire data table if the data is not already selected.
4. This will transpose rows to columns or columns to rows. It is equivalent to Copy, Paste Special,
Transpose.
6. Click SigmaXL > Data Manipulation > Stack Subgroups Across Rows.
8. Click on Shot 1. Shift Click on Shot 3 to highlight the three columns of interest.
26
SigmaXL: Measure Phase Tools
9. Click Numeric Data Variables (Y) >> to select Shot 1 to Shot 3. Enter Distance as the Stacked
Data (Y) Column Heading and Shot No as the Category (X) Column Heading. Select Operator,
click Additional Category Columns (X) >>:
Note that any selected column may be removed by highlighting and double-clicking or clicking
the Remove button.
27
SigmaXL: Measure Phase Tools
Stack Columns
1. Open Customer Satisfaction Unstacked.xlsx.
4. Shift Click on Overall Satisfaction_3 to highlight all three column names. Click Select Columns
>>. Enter the Stacked Data (Y) Column Heading (Optional) as Overall Satisfaction. Enter the
Category (X) Column Heading (Optional) as Customer Type.
28
SigmaXL: Measure Phase Tools
6. Data that is in stacked column format can be unstacked using Data Manipulation > Unstack
Columns.
Random Data
The normal random data generator is used to produce normal random data. Column headings are
automatically created with Mean and Standard Deviation values (e.g. 1: Mean = 0; StDev = 1). This
utility works with Recall SigmaXL Dialog (F3) to append columns to the current Normal Random
Data worksheet. An example is shown in Measure Phase Tools, Part G – Normal Probability Plots.
Additional random number generators include Uniform (Continuous & Integer), Lognormal,
Exponential, Weibull and Triangular. The column headings show the specified parameter values.
Box-Cox Transformation
This tool is used to convert nonnormal data to normal by applying a power transformation.
Examples of use are given in Measure Phase Tools, Part J – Process Capability for Nonnormal Data
and Control Phase Tools, Part A – Individuals Charts for Nonnormal Data.
Standardize Data
This tool is used to Standardize ((Yi – Mean)/StDev) or Code (Ymax = +1, Ymin = -1) your data. This
is particularly useful when performing Multiple Regression. Standardized Predictors have better
statistical properties. For example, the importance of model coefficients can be determined by the
relative size because units are removed. Another statistical benefit is reduced multicollinearity
when investigating two-factor interactions.
29
SigmaXL: Measure Phase Tools
3. Ensure that the entire data table is selected. If not, check Use Entire Data Table. Click Next.
4. Select Responsive to Calls and Ease of Communications and click Numeric (Y) Columns to
Standardize >>
2. Insert a new column in B; Click Column B heading, Right-Click > Insert > Columns.
3. Insert a new row in row 2. Click Row 2 label, Right-Click > Insert > Rows as shown:
30
SigmaXL: Measure Phase Tools
4. This is now an example of a data set that requires deletion of empty rows and columns. Click
SigmaXL >Data Manipulation >Data Preparation >Remove Blank Rows and Columns.
31
SigmaXL: Measure Phase Tools
7. Click Yes. The empty rows and columns are deleted automatically.
32
SigmaXL: Measure Phase Tools
Alternatively, you can use the Hot Key F3 (not available in Excel Mac 2011). This feature can also
be accessed by clicking SigmaXL > Help > Hot Keys > Recall SigmaXL Dialog.
Note that Recall SigmaXL Dialog may not be available for all functions.
33
SigmaXL: Measure Phase Tools
Worksheet Manager
The Worksheet Manager is a powerful utility that allows you to list all open Excel workbooks,
display all worksheets and chart sheets in a selected workbook and quickly select a worksheet or
chart sheet of interest. To access, click SigmaXL > Worksheet Manager.
34
SigmaXL: Measure Phase Tools
Click SigmaXL > Templates & Calculators to access the templates and calculators. Basic Graphical
Templates can also be accessed by clicking SigmaXL > Graphical Tools > Basic Graphical Templates.
Basic Statistical Templates can be accessed by clicking SigmaXL > Statistical Tools > Basic Statistical
Templates. Basic MSA Templates can be accessed by clicking SigmaXL > Measurement Systems
Analysis > Basic MSA Templates. Basic Process Capability Templates can be accessed by clicking
SigmaXL > Process Capability > Basic Process Capability Templates. Basic DOE Templates can also
be accessed by clicking SigmaXL > Design of Experiments > Basic DOE Templates. Basic Control
Chart Templates can also be accessed by clicking SigmaXL > Control Charts > Basic Control Chart
Templates.
35
SigmaXL: Measure Phase Tools
o SIPOC Diagram
o Control Plan
• Lean Templates:
o Takt Time Calculator
36
SigmaXL: Measure Phase Tools
o C-Chart
37
SigmaXL: Measure Phase Tools
Team/Project Charter
Click SigmaXL > Templates & Calculators > DMAIC & DFSS Templates > Team/Project Charter to
access the Team/Project Charter template.
38
SigmaXL: Measure Phase Tools
SIPOC Diagram
Click SigmaXL > Templates & Calculators > DMAIC & DFSS Templates > SIPOC Diagram to access
the SIPOC Diagram template.
39
SigmaXL: Measure Phase Tools
Notes:
1. Overwrite the text in the Cause, Sub-Cause and Problem Statement/Effect as appropriate.
3. The arrows with text are grouped. To ungroup use Excel's Draw > Ungroup tool.
40
SigmaXL: Measure Phase Tools
Simply fill in the template table and click the Fishbone Diagram button to create the diagram.
41
SigmaXL: Measure Phase Tools
The following example is given in SigmaXL > Help > Template Examples > C&E Matrix. This is a
simple Cause and Effect Matrix example for a Call Center. If prompted, please ensure that macros
are enabled in order to allow Pareto Charts to be created.
Notes for use of the Cause and Effect Matrix template, also known as the XY Matrix:
1. Weight the Output Variables (Y's) on a scale of 1 to 10 with 10 indicating most important to the
Customer.
2. For Root Cause Analysis, assign the association/effect multiplier score for each X to Y using a
scale of 0, 1, 3, 9, where 0 = None, 1 = Weak, 3 = Moderate, and 9 = Strong. Initially this
assignment will likely be a team subjective assessment. Data should be collected and the
degree of association should be validated with Graphical and Statistical Tools.
3. For Project Selection or Solution Selection, assign the association multiplier score for each X to
Y using a scale of 1 to 10, with 10 indicating strong association.
42
SigmaXL: Measure Phase Tools
The following example is given in SigmaXL > Help > Template Examples > FMEA. This is a simple
Failure Mode and Effects example for Stocking Inventory. If prompted, please ensure that macros
are enabled in order to allow sort by RPN.
If you hover the mouse cursor over the Severity, Occurrence or Detection heading the
recommended scale will appear as a comment:
43
SigmaXL: Measure Phase Tools
The recommended scales for Severity, Occurrence, and Detection are shown below:
44
SigmaXL: Measure Phase Tools
Click Risk Priority Number (RPN) Sort to sort the rows by RPN in descending order.
Click Revised Risk Priority Number Sort to sort the rows by Revised RPN in descending order.
Click Row Number Sort to restore the FMEA worksheet to the original row order (ascending).
45
SigmaXL: Measure Phase Tools
46
SigmaXL: Measure Phase Tools
47
SigmaXL: Measure Phase Tools
Control Plan
Click SigmaXL > Templates & Calculators > DMAIC & DFSS Templates > Control Plan to access the
Control Plan template.
48
SigmaXL: Measure Phase Tools
49
SigmaXL: Measure Phase Tools
50
SigmaXL: Measure Phase Tools
51
SigmaXL: Measure Phase Tools
Notes:
Open the file Template & Calculator Examples.xlsx. Click on the worksheet named VSM Example.
This is a present state value stream map of a manufacturing operation. This example is courtesy
and copyright of Strategos, Inc. www.strategosinc.com.
52
SigmaXL: Measure Phase Tools
Pareto Chart
100 100%
90 90%
80 80%
70 70%
60 60%
Count
50 50%
40 40%
30 30%
20 20%
10 10%
0 0%
ls r ng r e
al de lo bl
-c or -lo co ila
rn t o- t oo g- va
e tu lt - s- r on t- a
R ff icu ke W No
Di r- ta
e
rd
O
Category
Notes:
1. This Pareto Chart template should be used with count data like number of errors. You can also
create Pareto charts with cost data.
2. You can replace the Category and Count column headings with any headings that you wish.
3. Enter the Pareto categories in the Category column. These can be Name, Location, Error Type
or other text information. Pareto categories are required and will appear on the horizontal X-
Axis of the Pareto Chart.
53
SigmaXL: Measure Phase Tools
54
SigmaXL: Measure Phase Tools
Notes:
1. This Histogram template should be used with continuous data like cycle time.
2. You can replace the Data column heading with any heading that you wish.
55
SigmaXL: Measure Phase Tools
56
SigmaXL: Measure Phase Tools
Run Chart
60
50
Mean: 37.95
40
Run Chart: Data
30
20
10
0
re
re
re
re
re
r
te
te
te
te
te
fo
fo
fo
fo
fo
Af
Af
Af
Af
Af
Be
Be
Be
Be
Be
Notes:
1. This Run Chart template can be used with a variety of data types: continuous data like cycle
time, count data like number of errors, or cost data. The data must be in chronological time-
sequence order.
2. You can replace the X-Axis Label and Data column headings with any headings that you wish.
4. Enter labels in X-Axis Label column. Labels can be Date, Time, Name, or other text information.
These labels are optional and will appear on the horizontal X-Axis of the Run Chart.
57
SigmaXL: Measure Phase Tools
2. Delta is desired proportion margin of error. Enter as the half-width, i.e. if the desired margin of
error is +/- 3%, enter 0.03.
5. Power (1 – Beta) is not considered in these calculations. Power and Sample Size may be
calculated using SigmaXL > Statistical Tools > Power & Sample Size Calculators.
This sample size calculator is based on a confidence interval approach, where you enter the desired
half-interval value (delta). Let’s say that you desire the proportion margin of error to be +/- 3%. As
noted above, you simply enter a delta (half-interval) value of .03.
The challenge with this sample size calculator is that we need an estimate of the population
proportion. This is a bit of a chicken and egg situation – which comes first? I want to calculate an
appropriate sample size to determine an outcome, but I am being asked to enter an estimate of the
population proportion. First, keep in mind that this tool is a planning tool – the true confidence
intervals will be determined after you collect your data. Estimating the population proportion can
be done if you have good historical data to draw from, for example, in historical customer surveys
the percentage of satisfied customers was 80%. In this case, you could use P=0.8. If you do not have
a priori knowledge, then leave P=0.5 which gives the most conservative value (i.e., largest estimate
58
SigmaXL: Measure Phase Tools
of sample size). If you enter a value other than 0.5 it will result in a smaller sample size
requirement.
59
SigmaXL: Measure Phase Tools
1. Delta uses the same units as the standard deviation. Enter as the half-width, i.e., if the desired
margin of error is +/- 0.25, enter 0.25.
2. Enter (optional) population size N to adjust for small populations (N < 1000).
3. Power (1 – Beta) is not considered in these calculations. Power and Sample Size may be
calculated using SigmaXL > Statistical Tools > Power & Sample Size Calculators.
Similar to the Sample Size – Discrete calculator, this sample size calculator is based on a confidence
interval approach, where you enter the desired half-interval value (delta). If a survey had responses
that were on a continuous scale of 1 to 5, and you desired a margin of error on the mean to be +/-
.25, then you would use .25 as the delta value. Note that in this continuous case the units are not
percentages but level of satisfaction.
Here we need an a priori estimate of the population standard deviation. If you have no idea what
the standard deviation will be, then you could take a small sample to get a rough estimate of the
standard deviation.
60
SigmaXL: Measure Phase Tools
It is well known that the central limit theorem enables the t-Test and ANOVA to be fairly robust to
the assumption of normality. A question that invariably arises is, “How large does the sample size
have to be?” A popular rule of thumb answer for the one sample t-Test is “n = 30.” While this rule
of thumb often does work well, the sample size may be too large or too small depending on the
degree of nonnormality as measured by the Skewness and Kurtosis. Furthermore it is not
applicable to a One Sided t-Test, 2 Sample t-Test or One Way ANOVA.
To address this issue, we have developed a unique template that gives a minimum sample size
needed for a hypothesis test to be robust.
Click SigmaXL > Templates & Calculators > Basic Statistical Templates > Minimum Sample Size for
Robust t-Tests and ANOVA to access this template. It includes minimum sample size for
robustness for the 1 Sample t-Test, 2 Sample t-Test and the One Way ANOVA.
The user may specify the alternative hypothesis as “Less Than” (one sided), “Not Equal To” (two
sided) or “Greater Than” (one sided). Confidence levels of 90% (α = 0.1), 95% (α = .05) or 99% (α =
.01) may also be specified:
To use the template, simply select the appropriate Hypothesis Test, Alternative Hypothesis and
Confidence Level using the drop down selection. Enter Skewness and Kurtosis values as shown in
the yellow highlighted cells.
Note that in the example shown, the rule of thumb for a 1 Sample t-Test “n = 30” is confirmed with
a moderate skew value of 1.
61
SigmaXL: Measure Phase Tools
Now change the Alternative Hypothesis to “Less Than” and Confidence Level to 99% as shown:
On the other hand, if you want to perform a standard One Way ANOVA, enter the values as shown:
Now the minimum sample size requirement is only 3. This value applies to each sample or group,
so for the 3 Sample ANOVA that would mean each sample has n = 3 for a total number of
observations = 9.
Note that this calculator is strictly addressing the question of alpha robustness to nonnormality.
Power is not considered here.
If the minimum sample size requirements cannot be met, you should use a nonparametric
equivalent to the parametric hypothesis test (i.e. One Sample Sign or Wilcoxon, Two Sample Mann-
Whitney, Kruskal-Wallis or Mood’s Median: SigmaXL > Statistical Tools > Nonparametric Tests).
62
SigmaXL: Measure Phase Tools
A slight Skew is +/- 0.5, moderate Skew is +/- 1, severe Skew is +/- 2 and extreme Skew is +/- 5.The
Skewness range used should be -5 to +5. Values beyond this range are extrapolated so may be
inaccurate.
Kurtosis for a normal distribution is 0. Kurtosis must be greater than or equal to:
(Skew ^ 2 - 1.48).
Kurtosis “delta” is Kurt - Skew2 and is used in the regression equation. Kurtosis delta range should
be -1.48 to +1.48. A kurtosis delta less than -1.48 denotes a bimodal distribution so this is a lower
boundary. Values above +1.48 are extrapolated so may be inaccurate.
The calculator assumes that all samples have the same Skewness and Kurtosis.
Minimum sample size for robustness occurs when the simulated observed alpha is within +/- 20%
of the specified alpha (Bradley 1980 and Rhiel 1996).
Regression Models
A separate regression model was constructed for each hypothesis test, alternative hypothesis, and
confidence level (total of 21 models), using coding as shown.
The model predictor terms are Skewness^2 and Kurtosis Delta. The response is n minimum. The
term coefficient values are stored in the template as shown.
63
SigmaXL: Measure Phase Tools
Model R-Square values are typically over 99%, with some exceptions (96%) due to small estimated
sample sizes.
The model used is the one which matches the selected hypothesis test, alternative hypothesis, and
confidence level. The selected n is highlighted in red and is also displayed in the template Results
cell. If n > 2000, a text display “> 2000” is shown in the Results cell.
References
Boos, D.D. and Hughes-Oliver, J.M. (2000), “How Large Does n Have to be for Z and t Intervals?”
The American Statistician, 54(2), 121-128.
Bradley, J. V. (1980), “Nonrobustness in Z, t, and F Tests at Large Sample Sizes,” Bulletin of the
Psychonomics Society, 16(5), 333-336.
64
SigmaXL: Measure Phase Tools
Notes:
1. Enter summarized Sample Data, Known Population Standard Deviation, Null Hypothesis and
Confidence Level in cells with yellow highlight. Do not modify any other part of this worksheet.
3. This calculator will automatically detect the version of Excel and use the appropriate statistical
functions. If Excel 2010 or higher, then the Excel 2010 statistical functions are used.
65
SigmaXL: Measure Phase Tools
Notes:
1. Enter summarized Sample Data, Null Hypothesis and Confidence Level in cells with yellow
highlight. Do not modify any other part of this worksheet.
3. This calculator will automatically detect the version of Excel and use the appropriate statistical
functions. If Excel 2010 or higher, then the Excel 2010 statistical functions are used.
66
SigmaXL: Measure Phase Tools
Notes:
1. Enter summarized Sample Data, Null Hypothesis and Confidence Level in cells with yellow
highlight. Do not modify any other part of this worksheet.
3. This calculator will automatically detect the version of Excel and use the appropriate statistical
functions. If Excel 2010 or higher, then the Excel 2010 statistical functions are used.
67
SigmaXL: Measure Phase Tools
Notes:
1. Enter summarized Sample Data, Target, Upper & Lower Equivalence Limits and Confidence
Level in cells with yellow highlight. Do not modify any other part of this worksheet.
2. Null hypothesis for P1: Difference >= UEL; Null hypothesis for P2: Difference <= LEL. Both null
hypothesis must be rejected to conclude that equivalence is true.
3. LEL and UEL establish the zone or region of equivalence and are determined by what size
difference is considered practically significant.
4. This calculator will automatically detect the version of Excel and use the appropriate statistical
functions. If Excel 2010 or higher, then the Excel 2010 statistical functions are used.
68
SigmaXL: Measure Phase Tools
Notes:
1. Enter summarized Sample Data, Upper & Lower Equivalence Limits and Confidence Level in
cells with yellow highlight. Do not modify any other part of this worksheet.
3. Null hypothesis for P1: Mean Difference >= UEL; Null hypothesis for P2: Mean Difference <= LEL.
Both null hypothesis must be rejected to conclude that equivalence is true.
4. LEL and UEL establish the zone or region of equivalence and are determined by what size mean
difference is considered practically significant.
5. This calculator will automatically detect the version of Excel and use the appropriate statistical
functions. If Excel 2010 or higher, then the Excel 2010 statistical functions are used.
69
SigmaXL: Measure Phase Tools
Notes:
1. Enter summarized Sample Data, Null Hypothesis and Confidence Level in cells with yellow
highlight. Do not modify any other part of this worksheet.
3. This calculator will automatically detect the version of Excel and use the appropriate statistical
functions. If Excel 2010 or higher, then the Excel 2010 statistical functions are used.
70
SigmaXL: Measure Phase Tools
Notes:
1. Enter summarized Sample Data, Null Hypothesis and Confidence Level in cells with yellow
highlight. Do not modify any other part of this worksheet.
3. This calculator will automatically detect the version of Excel and use the appropriate statistical
functions. If Excel 2010 or higher, then the Excel 2010 statistical functions are used.
4. The F-test for two sample standard deviations/variances assumes that samples are normally
distributed.
71
SigmaXL: Measure Phase Tools
Notes:
1. Enter summarized Sample Data, Null Hypothesis and Confidence Level in cells with yellow
highlight. Do not modify any other part of this worksheet.
72
SigmaXL: Measure Phase Tools
• Use Normal Approximation only when n is large and proportion is close to 0.5 (check that
np >= 5 and nq >= 5, or the more conservative npq, shown above, >= 5). See Rosner (2010).
• Confidence intervals for binomial proportions have an "oscillation" phenomenon where the
coverage probability varies with n and p.
• Exact (Clopper-Pearson) is strictly conservative and will guarantee the specified confidence
level as a minimum coverage probability, but results in wide intervals. Recommended only
for applications requiring strictly conservative intervals.
• Wilson Score and Jeffreys Beta have mean coverage probability matching the specified
confidence interval. We recommend Wilson for two-sided confidence intervals and Jeffreys
for one-sided.
• Normal Approximation (Wald) is not recommended, but included here to validate hand
calculations. Use only when n is large and proportion is close to 0.5 (check that npq, shown
above, >= 5).
5. This calculator will automatically detect the version of Excel and use the appropriate statistical
functions. If Excel 2010 or higher, then the Excel 2010 statistical functions are used.
REFERENCES
1. Agresti, A. and Coull, B.A. (1998), Approximate is Better than “Exact” for Interval Estimation of
Binomial Proportions. The American Statistician, 52,119–126.
2. Brown, L. D., Cai, T. T. and DasGupta, A. (2001), Interval estimation for a binomial proportion.
Statistical Science, 16, 101-133. With comments and a rejoinder by the authors.
3. Cai, T. T. (2005), One-sided confidence intervals in discrete distributions. J. Statist. Plann. Inference,
131, 63-88.
4. Clopper, C.J., and Pearson, E.S.(1934), The Use of Confidence or Fiducial Limits Illustrated in the
Case of the Binomial. Biometrika, 26, 404–413.
5. Newcombe, R. (1998a), Two-sided confidence intervals for the single proportion: Comparison of
seven methods, Statistics in Medicine, 17, 857-872.
6. Rosner, B. (2010), Fundamentals of Biostatistics. Seventh Edition. Equation 5.15, page 133.
Duxbury Press.
7. Vollset, S.E. (1993). Confidence intervals for a binomial proportion, Statistics in Medicine, 12, 809–
824.
73
SigmaXL: Measure Phase Tools
Notes:
1. Enter summarized Sample Data, Null Hypothesis and Confidence Level in cells with yellow
highlight. Do not modify any other part of this worksheet.
74
SigmaXL: Measure Phase Tools
• Use Normal Approximation only when minimum expected cell value, shown above, is >= 5.
Unpooled is used if H0: P1 - P2 <> 0. Use pooled if H0: P1 - P2 = 0. (See Peltier).
• Newcombe-Wilson Score is recommended and has a mean coverage probability that is close
to the specified confidence interval. (See Newcombe method 10).
• Newcombe-Wilson Score (CC = Continuity Corrected) is conservative and will typically meet
the specified confidence level as a minimum coverage probability, but results in wider
intervals. (See Newcombe method 11).
• Exact methods that are strictly conservative (like Clopper-Pearson for the one proportion
case) are not included in this template because they are computationally intensive and
slow.
5. Due to the complexity of calculations, this template uses vba macros rather than Excel
formulas. SigmaXL must be initialized and appear on the menu in order for this template to
function.
REFERENCES
1. Beal, S. L. (1987), Asymptotic confidence intervals for the difference between two binomial
parameters for use with small samples. Biometrics, 43, 941-950.
2. Newcombe, R. G. (1998b), Interval estimation for the difference between independent proportions:
Comparison of eleven methods. Statistics in Medicine, 17:873–890.
4. Radhakrishna S., Murthy B.N., Nair N.G., Jayabal P., Jayasri R. (1992), Confidence intervals in
medical research. Indian J Med Res., Jun;96:199-205.
75
SigmaXL: Measure Phase Tools
Notes:
1. Enter summarized Sample Data, Upper & Lower Equivalence Limits and Confidence Level in
cells with yellow highlight. Do not modify any other part of this worksheet.
76
SigmaXL: Measure Phase Tools
• Exact methods that are strictly conservative (like Clopper-Pearson for the one proportion case) are
not included in this template because they are computationally intensive and slow.
3. Null hypothesis for P1: Proportion Difference >= UEL; Null hypothesis for P2: Proportion
Difference <= LEL. Both null hypothesis must be rejected to conclude that equivalence is true.
The P-Values are based on the normal approximation unpooled method.
4. LEL and UEL establish the zone or region of equivalence and are determined by what size
proportion difference is considered practically significant.
5. Since the confidence interval options use different methods than the hypothesis test (except
for Normal option), it is possible that the conclusion from the Equivalence P-Value will be
different from that of the confidence interval. In this case, we recommend using just the
confidence interval method.
6. Due to the complexity of calculations, this template uses VBA macros rather than Excel
formulas. SigmaXL must be initialized and appear on the menu in order for this template to
function.
REFERENCES
1. Beal, S. L. (1987), Asymptotic confidence intervals for the difference between two binomial parameters for
use with small samples. Biometrics, 43, 941-950.
2. Newcombe, R. G. (1998b), Interval estimation for the difference between independent proportions:
Comparison of eleven methods. Statistics in Medicine, 17:873–890.
3. Radhakrishna S., Murthy B.N., Nair N.G., Jayabal P., Jayasri R. (1992), Confidence intervals in medical
research. Indian J Med Res., Jun;96:199-205.
77
SigmaXL: Measure Phase Tools
Notes:
1. Enter summarized Sample Data, Null Hypothesis and Confidence Level in cells with yellow
highlight. Do not modify any other part of this worksheet.
2. Sample Size (area of opportunity) can be a count, time, length or other unit that defines the
area of opportunity for an occurrence.
78
SigmaXL: Measure Phase Tools
• Use Normal Approximation only when the number of occurrences, x > 10.
• Confidence intervals for Poisson rates have an "oscillation" phenomenon where the
coverage probability varies with x and n.
• Exact (Garwood Chi-Square) is strictly conservative and will guarantee the specified
confidence level as a minimum coverage probability, but results in wide intervals.
Recommended only for applications requiring strictly conservative intervals.
• Jeffreys has mean coverage probability matching the specified confidence interval and is
recommended.
• Normal Approximation (Wald) is not recommended, but included here to validate hand
calculations. Use only when the number of occurrences, x > 10.
6. This calculator will automatically detect the version of Excel and use the appropriate statistical
functions. If Excel 2010 or higher, then the Excel 2010 statistical functions are used.
REFERENCES
1. Garwood, F. (1936). Fiducial limits for the Poisson distribution. Biometrika 28:437–442.
2. Swift, M.B. (2009). Comparison of confidence intervals for a Poisson Mean-Further considerations,
Communication in Statistics — Theory and Methods, 38, 748–759.
79
SigmaXL: Measure Phase Tools
Notes:
1. Enter summarized Sample Data, Null Hypothesis and Confidence Level in cells with yellow
highlight. Do not modify any other part of this worksheet.
2. Sample Size (area of opportunity) can be a count, time, length or other unit that defines the
area of opportunity for an occurrence.
80
SigmaXL: Measure Phase Tools
• Use Normal Approximation only when the number of occurrences for each sample is
greater than 10. Unpooled is used if H0: λ1 - λ2 <> 0. Use pooled if H0: λ1 - λ2 = 0.
• Jeffreys Hybrid is recommended and has a mean coverage probability that is close to the
specified confidence interval. (See Li, 2011).
6. Due to the complexity of calculations, this template uses vba macros rather than Excel
formulas. SigmaXL must be initialized and appear on the menu in order for this template to
function.
REFERENCES
1. Li, H.Q., Tang, M.L., Poon, W.Y. and Tang, N.S. (2011), Confidence intervals for difference between
two Poisson rates, Comm. Stat. Simul. Comput. 40, pp. 1478–1491.
2. Przyborowski, J., Wilenski, H., (1940), Homogeneity of results in testing samples from Poisson
series. Biometrika 31, 313–323.
81
SigmaXL: Measure Phase Tools
Notes:
1. Enter summarized Sample Data, Upper & Lower Equivalence Limits and Confidence Level in
cells with yellow highlight. Do not modify any other part of this worksheet.
2. Sample Size (area of opportunity) can be a count, time, length or other unit that defines the
area of opportunity for an occurrence.
• Confidence intervals for difference in Poisson rates have an "oscillation" phenomenon where the
coverage probability varies with x and n.
• Jeffreys Hybrid is recommended and has a mean coverage probability that is close to the specified
confidence interval. (See Li, 2011).
• Normal Approximation will match the P-Value calculations, but should only be used when the
number of occurrences for each sample is greater than 10.
82
SigmaXL: Measure Phase Tools
4. Null hypothesis for P1: Rate Difference >= UEL; Null hypothesis for P2: Rate Difference <= LEL.
Both null hypothesis must be rejected to conclude that equivalence is true. The P-Values are
based on the normal approximation unpooled method.
5. LEL and UEL establish the zone or region of equivalence and are determined by what size rate
difference is considered practically significant.
6. Since the Jeffreys Hybrid confidence interval option uses a different method than the normal
based hypothesis test, it is possible that the conclusion from the Equivalence P-Value will be
different from that of the confidence interval. In this case, we recommend using just the
confidence interval method.
7. Due to the complexity of calculations, this template uses VBA macros rather than Excel
formulas. SigmaXL must be initialized and appear on the menu in order for this template to
function.
REFERENCES
1. Li, H.Q., Tang, M.L., Poon, W.Y. and Tang, N.S. (2011), Confidence intervals for difference between two
Poisson rates, Comm. Stat. Simul. Comput. 40, pp. 1478–1491.
2. Przyborowski, J., Wilenski, H., (1940), Homogeneity of results in testing samples from Poisson series.
Biometrika 31, 313–323.
83
SigmaXL: Measure Phase Tools
Notes:
1. Enter Category ID, Observed Count values and (optional) Historical Counts. Do not modify any
other part of this worksheet.
2. If optional Historical Counts are not specified, chi-square is calculated using equal expected
proportions.
3. If optional Historical Counts are specified, a value must be entered for each observed count.
4. Chi-Square statistic requires that no more than 20% of cells have an expected count less than 5.
84
SigmaXL: Measure Phase Tools
Notes:
1. Enter Category ID, Observed Count values and (optional) Historical Counts. Do not modify any
other part of this worksheet.
2. If optional Historical Counts are not specified, chi-square is calculated using equal expected
proportions.
3. If optional Historical Counts are specified, a value must be entered for each observed count.
4. The Chi-Square statistic requires that no more than 20% of cells have an expected count less
than 5 (and none of the cells have an expected count less than 1). If this assumption is not
satisfied, the Chi-Square approximation may be invalid and Exact or Monte Carlo P-Values
should be used. In this example 40% of the cells have an expected count less than 5 so Exact
should be used.
5. This example shows that the Exact P-Value is 0.5013 and the “large sample” Chi-Square P-Value
is 0.4897.
6. Chi-Square Exact solves the permutation problem using enhanced enumeration. For further
details refer to the Appendix Exact and Monte Carlo P-Values for Nonparametric and
Contingency Test.
7. It is important to note that while exact P-Values are “correct,” they do not increase (or
decrease) the power of a small sample test, so they are not a solution to the problem of failure
to reject the null due to inadequate sample size.
8. For data that requires more computation time than specified, Monte Carlo P-Values provide an
approximate (but unbiased) P-Value that typically matches exact to two decimal places using
10,000 replications. One million replications give a P-Value that is typically accurate to three
decimal places. A confidence interval (99% default) is given for the Monte Carlo P-Values.
85
SigmaXL: Measure Phase Tools
1. The Exact P-Value for this example is solved very quickly, so Monte Carlo is not needed, but we
will run it for continuity in the example. Click Monte Carlo P-Value. Select Number of
Replications = 10000 and Confidence Level for P-Value = 99%.
Tip: The Monte Carlo 99% confidence interval for P-Value is not the same as a confidence
interval on the test statistic due to data sampling error. The confidence level for the hypothesis
test statistic is still 95%, so all reported P-Values less than .05 will be highlighted in red to
indicate significance. The 99% Monte Carlo P-Value confidence interval is due to the
uncertainty in Monte Carlo sampling, and it becomes smaller as the number of replications
increases (irrespective of the data sample size). The Exact P-Value will lie within the stated
Monte Carlo confidence interval 99% of the time.
The Monte Carlo P-Value here is 0.5012 with a 99% confidence interval of 0.4896 to 0.5128.
This will be slightly different every time it is run (the Monte Carlo seed value is derived from the
system clock). The true Exact P-Value = 0.5013 lies within this confidence interval.
3. Now we will consider a small sample problem. Enter the following values for sample data in the
yellow highlight region. Note that the displayed Monte Carlo (or Exact) P-Values are cleared
when new data is entered in the template:
86
SigmaXL: Measure Phase Tools
This example is adapted from Mehta, C.R. and Patel, N. R., IBM SPSS Exact Tests, IBM Corp.,
page 44.
ftp://public.dhe.ibm.com/software/analytics/spss/documentation/statistics/21.0/en/client/Ma
nuals/IBM_SPSS_Exact_Tests.pdf
4. Click Exact P-Value. Select Time Limit for Exact Computation = 60 seconds.
Note that the Exact P-Value is 0.0523 which is a “fail-to-reject” of the null hypothesis (H0), but
the “large sample” or “asymptotic” Chi-Square P-Value incorrectly rejected H0 with a P-Value of
0.046. The exact P-Value matches that given in the reference. The error can also go the other
way, where a large sample Chi-Square P-Value is a “fail-to-reject” of the null hypothesis and the
Exact P-Value is a rejection of H0.
In conclusion, always use the Exact (or Monte Carlo) P-Value when the Chi-Square large sample
assumptions are not met.
87
SigmaXL: Measure Phase Tools
88
SigmaXL: Measure Phase Tools
The following example is given in SigmaXL > Help > Template Examples > MSA > Type 1 Gage
Study.
89
SigmaXL: Measure Phase Tools
Notes:
2. Enter your data in the Measurement Data column. Minimum recommended sample size is 25,
but 50 is preferred. This template has a maximum sample size of 1000, suitable for automated
test systems.
3. Enter Tolerance (Upper Specification Limit USL - Lower Specification Limit LSL).
4. Enter Part Reference Value. If missing, Gage Capability Cgk is not computed. Bias will assume a
0 value.
6. The default StDev Multiplier is 6 (99.73% coverage). Change this to 5.15 if AIAG convention is
being used (99% coverage). Dietrich [1] recommends using 4 for approximate 95% coverage.
We recommend 6 for typical and critical measurement systems, and 4 for non-critical.
7. The default Percent of Tolerance for Cg/Cgk is 20. Stacey [2] suggests using 10 for critical
measurement systems.
8. If Cg and Cgk are >= 1.33, the measurement device is capable. If Cg or Cgk are < 1.33, the
measurement device is not capable and needs to be improved.
9. If Resolution % of Tolerance (RE%) is <= 5%, the resolution is acceptable. If RE% is > 5% the
resolution is not acceptable.
10. Click the Type 1 Gage Run Chart button to create a Run Chart with center line = Ref, upper limit
= Ref + 0.1*Tol and lower limit = Ref - 0.1*Tol (assuming 20% of Tolerance).
REFERENCES
90
SigmaXL: Measure Phase Tools
The following example is given in SigmaXL > Help > Template Examples > MSA > Gage Bias and
Linearity Study.
91
SigmaXL: Measure Phase Tools
Notes:
1. Enter Process Variation (SV from Gage R&R or 6 * Historical StDev). If neither are available, use
Process Tolerance. % Average Bias will be blank if not specified. Note that % Linearity does not
require Process Variation.
2. Enter part Reference Values. These should be relative to a traceable standard, but if not
available, see AIAG MSA Reference Manual [1], pp. 92-93 for suggestions on how to obtain
reference values. Reference values should cover the operating range of the gage.
3. Enter part measurement data in the yellow highlight region below the respective
part/reference value. AIAG recommends a minimum of 10 trials per part/reference. It is
important that the parts be selected at random for each trial in order to minimize appraiser
"recall" bias.
5. If a P-Value is < .05, it is highlighted in bold red to indicate significance. If the Average Bias P-
Value or Linearity Slope P-Value are < .05, the respective bar in Percent of Process Variation will
also be red to highlight significance.
6. Statistical significance is the best indicator of a problem with bias or linearity. Some
practitioners also use a rule of thumb that a good measurement system will have % Average
Bias < 5% and % Linearity < 5%.
7. The dashed lines on the Linearity scatter plot are the regression fit 95% confidence intervals. If
the zero reference line is not within the intervals, the linearity slope is significant, indicating
that the measurement system linearity is not acceptable.
8. This calculator requires Excel 2010 or higher. Linearity regression uses vba macros rather than
Excel formulas. SigmaXL must be initialized and appear on the menu in order for this template
to function.
9. Rows 34-39 and 42 are hidden to simplify the output report. These may be unhidden to view
how the calculations are performed.
REFERENCE
1. Automotive Industry Action Group AIAG (2010). Measurement Systems Analysis MSA Reference
Manual, 4th Edition
92
SigmaXL: Measure Phase Tools
The following example is given in SigmaXL > Help > Template Examples > MSA > Gage R&R. If
prompted, please ensure that macros are enabled.
1. The Automotive Industry Action Group (AIAG) recommended study includes 10 Parts, 3
Operators and 3 Replicates. The template calculations will work with a minimum of 2
Operators, 2 Parts and 2 Replicates. The data should be balanced with each operator
measuring the same number of parts and the same number of replicates. Use SigmaXL >
Measurement Systems Analysis to specify up to 30 Parts, 10 Operators and 10 Replicates.
2. Enter process Upper Specification Limit (USL) and Lower Specification Limit (LSL) in the Process
Tolerance window. This is used to determine the % Tolerance metrics. If the specification is
single-sided, leave both entries blank.
93
SigmaXL: Measure Phase Tools
3. The default StDev multiplier is 6. Change this to 5.15 if AIAG convention is being used.
4. The cells shaded in light blue highlight the critical metrics Gage R&R % Total Variation (also
known as %R&R) and %Tolerance: < 10% indicates a good measurement system;
> 30% indicates an unacceptable measurement system.
94
SigmaXL: Measure Phase Tools
Operator - Part 01 Operator - Part 02 Operator - Part 03 Operator - Part: Part 04 Operator - Part: Part 05
1.6890
1.5890
X-Bar - Reading1 - Reading 3
1.4890 1.46
1.42
1.3890 1.38
1.2890
1.1890
1.0890
-1
-2
-3
-4
-5
-6
-7
-8
-9
0
0
-1
-2
-3
-4
-5
-6
-7
-8
-9
-1
-2
-3
-4
-5
-6
-7
-8
-9
-1
-1
-1
C
A
C
A
0.06970
0.07
0.06
R - Reading1 - Reading3
0.05
0.04
0.03
0.02133
0.02
0.01
0.00000
0
-1
-2
-3
-4
-5
-6
-7
-8
-9
0
0
-1
-2
-3
-4
-5
-6
-7
-8
-9
-1
-2
-3
-4
-5
-6
-7
-8
-9
-1
-1
-1
C
A
C
A
Notes for use of the Gage R&R: Multi-Vari & X-bar R Charts:
1. The Gage R&R Multi-Vari and X-bar & R charts can only be generated if a Gage R&R template
has been completed and is the active worksheet. Select the Gage R&R worksheet and click
SigmaXL > Templates & Calculators > Gage R&R: Multi-Vari & X-Bar R Charts to create the
above charts.
2. The Multi-Vari chart shows each Part as a separate graph. Each Operator’s response readings
are denoted as a vertical line with the top tick corresponding to the Maximum value, bottom
tick is the Minimum, and the middle tick is the Mean. The horizontal line across each graph is
the overall average for each part.
3. When interpreting the X-bar and R chart for a Gage R&R study, it is desirable that the X-bar
chart be out-of-control, and the Range chart be in-control. The control limits are derived from
within Operator repeatability.
95
SigmaXL: Measure Phase Tools
96
SigmaXL: Measure Phase Tools
2. Recommend for study: 3 Appraisers, 2 to 3 Replicates, Minimum of 10 Good Parts and 10 Bad
Parts. The data should be balanced with each appraiser evaluating the same number of parts
and the same number of replicates.
3. Specify the Good Part or Unit as G or other appropriate text (P, Y, etc.). Specify the Bad Part or
Unit as NG or other appropriate text (F, N, etc.). Be careful to avoid typing or spelling errors
when entering the results. A space accidentally inserted after a character will be treated as a
different value leading to incorrect results.
97
SigmaXL: Measure Phase Tools
Notes for use of the Process Sigma Calculator for Discrete Data:
1. Total number of defects should include defects made and later fixed.
98
SigmaXL: Measure Phase Tools
Note: This calculator assumes that the Mean and Standard Deviation are computed from data that
are normally distributed.
99
SigmaXL: Measure Phase Tools
1. This calculator assumes that the Mean and Standard Deviation are computed from data that
are normally distributed.
2. Reports Cp, Cpk if entered S is Within or Short Term (using a control chart).
100
SigmaXL: Measure Phase Tools
Notes for use of the Process Capability & Confidence Intervals Calculator:
1. This calculator assumes that the Mean and Standard Deviation are computed from data that
are normally distributed.
2. Reports Cp, Cpk if entered S is Within or Short Term (using a control chart).
101
SigmaXL: Measure Phase Tools
Notes:
1. Enter summarized Sample Data, Population Coverage and Confidence Level in cells with yellow
highlight.
3. This calculator assumes that the summary statistics come from data that is normally
distributed. Tolerance intervals are exact, one-sided uses the noncentral t distribution, two-
sided uses the noncentral chi-square distribution. See Appendix Tolerance Interval Calculator
(Normal Exact) for statistical details.
4. Due to the complexity of calculations, this template uses vba macros rather than Excel
formulas. SigmaXL must be initialized and appear on the menu in order for this template to
function.
102
SigmaXL: Measure Phase Tools
Click the Individuals Control Chart button to produce the Individuals Chart (note that in actual
application, one should use a minimum of 20 to 30 data points to compute the initial control
limits).
70 64.34
60
40
30.06
30
20
10
0
Before Before Before Before Before Before Before Before Before Before
103
SigmaXL: Measure Phase Tools
Now enter the data as shown (or copy and paste the “After” data from the histogram data used
earlier):
Click the Add Data button to add the “After” data to the Individuals chart.
70
64.34
60
40
30.06
30
20
10
0
re
re
re
re
re
r
te
te
te
te
te
fo
fo
fo
fo
fo
Af
Af
Af
Af
Af
Be
Be
Be
Be
Be
104
SigmaXL: Measure Phase Tools
Notes:
1. This Individuals Control Chart template should be used with continuous data like cycle time.
The data must be in chronological time-sequence order.
2. You can replace the X-Axis Label and Data column headings with any headings that you wish.
4. Enter labels in X-Axis Label column. Labels can be Date, Time, Name, or other text information.
These labels are optional and will appear on the horizontal X-Axis of the Individuals Control
Chart.
5. Click the Individuals Control Chart button to create an Individuals Control Chart.
6. After the control chart has been created and additional new data entered into the Data
column, click the Add Data button to add the data to the existing chart.
105
SigmaXL: Measure Phase Tools
Descriptive Statistics
1. Open Customer Data.xlsx. (To access, click SigmaXL > Help > Sample Data or Start > Programs
> SigmaXL > Sample Data). Click Sheet 1 Tab.
4. Select Overall Satisfaction, click Numeric Data Variables (Y) >>. Select Customer Type, click
Group Category (X1) >> as shown:
5. Click OK.
6. Descriptive Statistics are given for Customer Satisfaction grouped by Customer Type:
Which Customer Type has the highest mean satisfaction score? Clearly Type 2. However, we have
to be careful concluding that there is a significant difference in satisfaction solely by looking at the
Means. In the Analyze Phase, we will run tests of hypothesis to validate that Type 2 Customers are,
in fact, significantly more satisfied.
106
SigmaXL: Measure Phase Tools
Tip: Click on Column B, click View | Window | Split, Freeze Panes. This freezes Column A and
allows you to scroll across the Descriptive Statistics for each level of the Group Category. This is
particularly beneficial when there are a large number of columns.
7. Click Recall SigmaXL Dialog menu or press F3 to recall last dialog. Change the format selected
to Column Format as shown:
8. Click OK. Descriptive Statistics are given for Customer Satisfaction broken out by Customer
Type in Column Format:
Tip: This column format is useful to create subsequent graphs on the summary statistics.
107
SigmaXL: Measure Phase Tools
Tip: Select only those options that are of interest in order to minimize the size of the report.
Here we are selecting all options for demonstration purposes. Note that when any option is
checked, Row Format is automatically selected, Column Format and Group Category (X2) are
greyed out. These display options are limited due to the amount of information displayed in
the extended report.
2. Click OK. Extended Descriptive Statistics are given for Customer Satisfaction grouped by
Customer Type:
108
SigmaXL: Measure Phase Tools
109
SigmaXL: Measure Phase Tools
110
SigmaXL: Measure Phase Tools
9. The Percentile Tolerance Intervals are 50%, 80%, 90%, 95%, 98%, 99%, and 99.73%.
10. Confidence Intervals and Tolerance Intervals can be exact or interpolated. If exact, the actual
exact confidence level will be a value greater than or equal to specified, due to percentile
values being discrete in nature. The actual exact level will also be reported in this case. If
interpolated, the result will be an interpolated estimate of the specified confidence level
(typically 95.0%) and is the recommended setting. See Appendix Percentile (Nonparametric)
Confidence and Tolerance Intervals for further details.
11. If the Confidence Interval or Tolerance Interval cannot be computed due to inadequate sample
size, a minimum sample size is reported.
111
SigmaXL: Measure Phase Tools
12. The Outlier (Boxplot Rules) Tests are: Potential 1.5(IQR), Likely 2.2(IQR), Extreme 3.0(IQR)
13. Grubbs Outlier Test is more powerful at detecting a single outlier as maximum or minimum but
assumes that the remainder of the data are normally distributed.
15. The Outlier and Randomness Tests use the same Green, Yellow, Red highlight given in the
automatic assumptions report that are included in t-tests and ANOVA.
112
SigmaXL: Measure Phase Tools
16. Click Recall SigmaXL Dialog menu or press F3 to recall last dialog. Click Options. Uncheck
Select All to clear the selections and check Percentile Confidence Intervals, select Exact and
Percentile. Check Percentile Tolerance Intervals, and select Exact as shown:
113
SigmaXL: Measure Phase Tools
17. Click OK. The Percentile Confidence Intervals and Tolerance Intervals are displayed:
18. The specified 95% is a guaranteed minimum. The exact confidence level is given with each
reported interval.
114
SigmaXL: Measure Phase Tools
Part D – Histograms
3. Ensure that entire data table is selected. If not, check Use Entire Data Table. Click Next.
4. Select Overall Satisfaction, click Numeric Data Variable (Y) >> as shown:
5. Click Next. Ensure that Normal Curve is checked. Set Start Point = 1. Change the Bin Width to
0.5, and the Number of Bins to 8. Click Update Chart to view the histogram. (If the survey
satisfaction data was pure integer format we would have checked the Integer Data option).
Note: For SigmaXL Mac version, click SigmaXL > Graphical Tools > Basic Histogram Options to
set options. The Histogram Basic Chart is displayed after you click Finish.
115
SigmaXL: Measure Phase Tools
7. Note that bin one is 1 to < 1.5, bin 2 is 1.5 to < 2, etc.
Tip: Any graph produced by SigmaXL can be Copied/Pasted into Word. It can also be enlarged
by clicking on the graph and dragging the corner. The number of decimal places displayed can
be modified by clicking on the Axis Label and selecting the Number tab to adjust. The text label
alignment can also be modified by: Select Axis, Right Click, Format Axis.
Multiple Histograms
1. Click Sheet 1 Tab of Customer Data.xlsx (or press F4 to activate last worksheet).
2. Click SigmaXL > Graphical Tools > Histograms & Descriptive Statistics.
3. Ensure that entire data table is selected. If not, check Use Entire Data Table. Click Next.
4. Select Overall Satisfaction, click Numeric Data Variables (Y) >>, select Customer Type, click
Group Category (X1) >> as shown:
116
SigmaXL: Measure Phase Tools
5. Click OK. Multiple Histograms and Descriptive Statistics of Customer Satisfaction By Customer
Type are produced:
Clearly Customer Type 2 shows a higher level of overall satisfaction, with the data skewed left.
Note that Customer Type 1 and 3 have data that is normally distributed, but this is not desirable
when the response is a satisfaction score!
6. Note that bin one is 1.72 to < 1.99, bin 2 is 1.99 to < 2.26, etc. The number of decimals
displayed can be changed by double-clicking on the X axis, click Number tab, and adjust decimal
places.
117
SigmaXL: Measure Phase Tools
Part E – Dotplots
Dotplots
1. Click Sheet 1 Tab of Customer Data.xlsx (or press F4 to activate last worksheet). Click SigmaXL
> Graphical Tools > Dotplots.
2. Ensure that entire data table is selected. If not, check Use Entire Data Table. Click Next.
3. Select Overall Satisfaction, click Numeric Data Variable (Y) >>, select Customer Type, click
Group Category (X1) >>:
4. Click OK. Multiple dotplots of Customer Satisfaction By Customer Type are produced:
5. Since all data points are shown, dotplots are a useful alternative to Histograms, particularly
when the group sample sizes are small (n < 30).
118
SigmaXL: Measure Phase Tools
Part F – Boxplots
Boxplots
1. Click Sheet 1 Tab of Customer Data.xlsx (or press F4 to activate last worksheet).
3. Ensure that entire data table is selected. If not, check Use Entire Data Table. Click Next.
4. Select Overall Satisfaction, click Numeric Data Variable (Y) >>, select Customer Type, click
Group Category (X1) >>, check Show Mean, check Show Legend:
6. The legend indicates that the solid center line is the median. The dashed red line shows the
sample mean. The top of the box is the 75th percentile (Q3). The bottom of the box is the 25th
percentile (Q1). The height of the box is called the Inter-Quartile Range (IQR) and is a robust
measure of spread or sample variability. The data point highlighted for Customer Type 2 is a
potential outlier (< Q1 – 1.5 * IQR or > Q3 + 1.5 * IQR). Note that extreme outliers are
highlighted with a solid dot (< Q1 – 3 * IQR or > Q3 + 3 * IQR).
119
SigmaXL: Measure Phase Tools
Tip: If you hover your mouse cursor in the middle of any of the Boxplot lines as shown, you will see
the balloon help indicating what the line is and its numeric Y value.
7. Now we will modify the Y axis scale, showing 1 as minimum and 5 as maximum (given that the
response data comes from a survey with 1-5 scale). To do this right click on the Y axis and
select “Format Axis”, modify the minimum value and maximum value. Change Horizontal axis
crosses > Axis value to 1 as shown:
120
SigmaXL: Measure Phase Tools
10. Select Overall Satisfaction, click Numeric Data Variable (Y) >>; select Customer Type, click
Group Category (X1) >>; select Size of Customer, click Group Category (X2) >>; check Show
Mean; uncheck Show Legend:
121
SigmaXL: Measure Phase Tools
11. Click OK. Boxplots of Customer Satisfaction By Customer Type and Size are produced:
12. In order to adjust the Y-axis scale for both charts, click SigmaXL Chart Tools > Set Chart Y-Axis
Max/Min.
13. Click OK. The Y-axis scale maximum and minimum are now modified for both charts.
122
SigmaXL: Measure Phase Tools
3. Create a Histogram & Descriptive Statistics for this data. Your data will be slightly different due
to the random number generation:
If the P-Value of the Anderson-Darling Normality test is greater than or equal to .05, the data is
considered to be normal (interpretation of P-Values will be discussed further in Analyze).
4. Create a normal probability plot of this data: Click SigmaXL Random Data (1) Sheet, Click
SigmaXL > Graphical Tools > Normal Probability Plots.
5. Ensure that entire data table is selected. If not, check Use Entire Data Table. Click Next.
6. Select Normal Data, click Numeric Data Variable (Y) >>. Check Add Title. Enter Example
Normal Prob Plot. Click OK.
123
SigmaXL: Measure Phase Tools
7. Click OK. A Normal Probability Plot of simulated random data is produced (again, your plot will
be slightly different due to the random number generation):
The data points follow the straight line fairly well, indicating that the data is normally distributed.
Note that the data will not likely fall in a perfectly straight line. The eminent statistician George
Box uses a “Fat Pencil” test where the data, if covered by a fat pencil, can be considered normal!
We can also see that the data is normal since the points fall within the normal probability plot 95%
confidence intervals (confidence intervals will be discussed further in Analyze).
10. Ensure that entire data table is selected. If not, check Use Entire Data Table. Click Next.
124
SigmaXL: Measure Phase Tools
11. Select Overall Satisfaction; click Numeric Data Variable (Y) >>. Click OK. A Normal Probability
Plot of Customer Satisfaction data is produced:
Is this data normally distributed? See earlier histogram and descriptive statistics of Customer
Satisfaction data.
12. Now we would like to stratify the customer satisfaction score by customer type and look at the
normal probability plots.
13. Click Sheet 1 of Customer Data.xlsx. Click SigmaXL > Graphical Tools > Normal Probability
Plots. Ensure that Entire Table is selected, click Next. (Alternatively, press F3 or click Recall
SigmaXL Dialog to recall last dialog).
14. Select Overall Satisfaction, click Numeric Data Variable (Y) >>; select Customer Type as Group
Category (X) >>. Click OK. Normal Probability Plots of Overall Satisfaction by Customer Type
are produced:
Reviewing these normal probability plots, along with the previously created histograms and
descriptive statistics, we see that the satisfaction data for customer type 2 is not normal, and
skewed left, which is desirable for satisfaction data! Note that although the customer type 2 data
falls within the 95% confidence intervals, the Anderson Darling test from descriptive statistics
125
SigmaXL: Measure Phase Tools
shows p < .05 indicating nonnormal data. Smaller sample sizes tend to result in wider confidence
intervals, but we still see that the curvature for customer type 2 is quite strong.
Tip: Use the Normal Probability Plot (NPP) to distinguish reasons for nonnormality. If the data fails
the Anderson-Darling (AD) test (with p < 0.05) and forms a curve on the NPP, it is inherently
nonnormal or skewed. Calculations such as Sigma Level, Pp, Cp, Ppk and Cpk assume normality and
will therefore be affected. Consider transforming the data using LN(Y) or SQRT(Y) or using the Box-
Cox Transformation tool (SigmaXL > Data Manipulation > Box-Cox Transformation) to make the
data normal. Of course, whatever transformation you apply to your data, you must also apply to
your specification limits. See also the Process Capability for nonnormal data tools.
If the data fails the AD normality test, but the bulk of the data forms a straight line and there are
some outliers, the outliers are driving the nonnormality. Do not attempt to transform this data!
Determine the root cause for the outliers and take corrective action on those root causes.
126
SigmaXL: Measure Phase Tools
Run charts, also known as trend charts and time series plots, add the dimension of time to the
graphical tools. They allow us to see trends and outliers in the data. Run Charts are a precursor to
control charts, which add calculated control limits. Note that Run Charts should be used only on
unsorted data, in its original chronological sequence.
Run Charts
1. Click Sheet 1 Tab of Customer Data.xlsx (or press F4 to activate last worksheet). Click SigmaXL
> Graphical Tools > Run Chart. Ensure that entire data table is selected. If not, check Use Entire
Data Table. Click Next.
2. Select Overall Satisfaction, click Numeric Data Variable (Y) >>. Select Show Mean. Uncheck
Nonparametric Runs Test (to be discussed later in Part N of Analyze Phase).
3. Click OK. A Run Chart of Overall Satisfaction with Mean center line is produced.
127
SigmaXL: Measure Phase Tools
4. Select the Y axis, Right Click, Format Axis, to activate the Format Axis dialog. Change Minimum
to 1, Maximum to 5, Horizontal axis crosses > Axis value to 1:
6. Are there any obvious trends? Some possible cycling, but nothing clearly stands out. It may be
interesting to look more closely at a specific data point. Any data point value can be identified
by simply moving the cursor over it:
128
SigmaXL: Measure Phase Tools
7. A label can be added to a data point by one single-click on the data point, followed by a Right
mouse click, and select Add Data Label. See also SigmaXL Chart Tools > Add Data Label in
Control Phase Tools, Part B - X-Bar & Range Charts.
8. Click OK. Resulting Run Chart with label attached to data point:
9. This label can be changed to a text comment. Single-click three times on the label and type in a
comment as shown:
129
SigmaXL: Measure Phase Tools
2. Ensure that entire data table is selected. If not, check Use Entire Data Table. Click Next.
3. Select Overall Satisfaction, Responsive to Calls and Ease of Communications. Click Numeric Data
Variable (Y) >>.
4. Click OK. An Overlay Run Chart of Overall Satisfaction, Responsive to Calls and Ease of
Communications is produced.
Overall Satisfaction
3 Responsive to Calls
Ease of Communications
0
1 5 9 13 17 21 25 29 33 37 41 45 49 53 57 61 65 69 73 77 81 85 89 93 97
130
SigmaXL: Measure Phase Tools
See Basic MSA Templates – Type 1 Gage Study for an example of the Type 1 Gage Study template.
See Basic MSA Templates – Gage Bias and Linearity Study for an example of the Gage Bias and
Linearity Study template.
See Basic MSA Templates – Gage R&R Study (MSA) for an example of the Gage R&R Study (MSA)
template. If you wish to analyze the template data with the Gage R&R Analysis Tool, click Create
Stacked Column Format for “Analyze Gage R&R” >>.
After entering the data into the Gage R&R Study template, you can create Gage Multi-Vari & X-bar
R charts. Click SigmaXL > Templates and Calculators > Basic MSA Templates > Gage R&R: Multi-
Vari & X-bar R Charts or SigmaXL > Measurement Systems Analysis > Basic MSA Templates >
Gage R&R: Multi-Vari & X-bar R Charts. See Basic MSA Templates - Gage R&R: Multi-Vari & X-bar
R Charts for an example.
131
SigmaXL: Measure Phase Tools
See Basic MSA Templates – Attribute Gage R&R (MSA) for an example of the Attribute Gage R&R
(MSA) template.
2. The default settings are 10 Parts, 3 Operators, 3 Replicates with Parts and Operators
randomized. Clicking the Reset button will restore the dialog to these default settings.
132
SigmaXL: Measure Phase Tools
5. Your worksheet order will be different due to the randomization. The results of the Gage R&R
study are entered in the yellow highlighted Measurement cells.
133
SigmaXL: Measure Phase Tools
2. Click SigmaXL > Measurement Systems Analysis > Analyze Gage R&R (Crossed). The data
worksheet is recognized by SigmaXL and highlighted automatically. Click Next.
3. Select Part, Operator and Measurement as shown. Check Display Multi-Vari & X-Bar R Charts.
Check Tolerance/Historical StDev (Optional) and enter 8 for the Process Tolerance Upper-
Lower Spec.
134
SigmaXL: Measure Phase Tools
135
SigmaXL: Measure Phase Tools
5. The first ANOVA table shows that the Operator * Part Interaction is not significant with the P-
Value = 0.9741 (greater than the alpha-to-remove value of 0.1). The second ANOVA table
shows the results excluding the Part * Operator Interaction.
6. The cells shaded in light blue highlight the critical metrics: Gage R&R % Total Variation (also
known as %R&R) and %Tolerance: < 10% indicates a good measurement system;
> 30% indicates an unacceptable measurement system. The Number of Distinct Categories
should be at least 5. Initial analysis shows that this is a marginal measurement system.
(Traditionally NDC is truncated to an integer value, but SigmaXL reports a more informative one
decimal place).
7. SigmaXL also reports the confidence intervals for the Gage R&R metrics. Note that when the
confidence intervals are taken into account, it is possible that this measurement system is
totally inadequate. This strongly suggests that the measurement system needs to be improved
but it also points out the weakness of the traditional Gage R&R study with 10 Parts, 3
Operators, and 3 Replicates. These sample sizes will typically yield wide confidence intervals.
For statistical details of the Gage R&R confidence intervals see Burdick, R. K., Borror, C. M., and
Montgomery, D. C., “Design and Analysis of Gauge R&R Studies: Making Decisions with
Confidence Intervals in Random and Mixed ANOVA Models”, ASA-SIAM Series on Statistics and
Applied Probability, 2005.
8. Click on the Gage R&R – Multi-Vari Sheet tab to view the Multi-Vari Chart:
The Multi-Vari chart shows each Part as a separate graph. Each Operator’s response readings
are denoted as a vertical line with the top tick corresponding to the Maximum value, bottom
tick is the Minimum, and the middle tick is the Mean. The horizontal line across each graph is
the overall average for each part.
136
SigmaXL: Measure Phase Tools
9. Click on the Gage R&R – X-Bar R Sheet tab to view the X-Bar & R Chart:
When interpreting the X-bar and R chart for a Gage R&R study, it is desirable that the X-bar
chart be out-of-control, and the Range chart be in-control. The control limits are derived from
within Operator repeatability. The Range chart indicates a problem with Operator B being
inconsistent.
137
SigmaXL: Measure Phase Tools
1. Open the file Attribute MSA – AIAG.xlsx. This is an example from the Automotive Industry
Action Group (AIAG) MSA Reference Manual, 3rd edition, page 127 (4th Edition, page 134).
There are 50 samples, 3 appraisers and 3 trials with a 0/1 response. A “good” sample is denoted
as a 1. A “bad” sample is denoted as a 0. Note that the worksheet data must be in stacked
column format and the known reference values must be consistent for each sample.
2. Click SigmaXL > Measurement Systems Analysis > Attribute MSA (Binary). Ensure that the
entire data table is selected. Click Next.
3. Select Part, Appraiser, Assessed Result and Reference as shown. Check Report Information and
enter AIAG Attribute MSA Binary for Product/Unit Name. Select Percent Confidence Interval
Type – Exact. The default Good Level of “1”will be used as specified in the AIAG manual :
Tip: The Good Level definition is used to determine Type I and Type II error rates. It is
applicable only when a True Standard is selected.
Tip: Percent Confidence Interval Type applies to the Percent Agreement and Percent
Effectiveness Confidence Intervals. These are binomial proportions that have an "oscillation"
phenomenon where the coverage probability varies with the sample size and proportion value.
Exact is strictly conservative and will guarantee the specified confidence level as a minimum
coverage probability, but results in wider intervals. Wilson Score has a mean coverage
138
SigmaXL: Measure Phase Tools
probability that matches the specified confidence interval. Since the intervals are narrower and
thereby more powerful, Wilson Score is recommended for use in attribute MSA studies due to
the small sample sizes typically used. Exact is selected in this example for continuity with the
results from SigmaXL Version 6.
4. Click OK. The Attribute MSA Binary Analysis Report is produced. The tables and associated
graphs are described separately by section for clarity.
Tip: While this report is quite extensive, a quick assessment of the attribute measurement
system can be made by viewing the Kappa color highlights: Green - very good agreement
(Kappa >= 0.9); Yellow - marginally acceptable, improvement should be considered (Kappa 0.7
to < 0.9); Red - unacceptable (Kappa < 0.7). Further details on Kappa are given below.
139
SigmaXL: Measure Phase Tools
Within Appraiser Percent Agreement will decrease as the number of trials increase because a
match occurs only if an appraiser is consistent across all trials. Use the Kappa/CI: Within Appraiser
Agreement Graph to determine adequacy of the Within Appraiser agreement. See below for
additional interpretation guidelines.
Tip: Hover the mouse pointer over the heading cells to view the following report comments.
140
SigmaXL: Measure Phase Tools
LC = Percent Lower Confidence Limit. UC = Percent Upper Confidence Limit. Confidence intervals
(CI) for binomial proportions have an "oscillation" phenomenon where the coverage probability
varies with n and p. Exact is strictly conservative and will guarantee the specified confidence level
as a minimum coverage probability, but results in wide intervals. Wilson Score has a mean
coverage probability that matches the specified confidence interval. Since the intervals are
narrower and thereby more powerful, they are recommended for use in attribute MSA studies due
to the small sample sizes typically used. See Appendix Percent Confidence Intervals (Exact Versus
Wilson Score) for references.
Fleiss’ Kappa statistic is a measure of agreement that is analogous to a “correlation coefficient” for
discrete data. Kappa ranges from -1 to +1: A Kappa value of +1 indicates perfect agreement. If
Kappa = 0, then agreement is the same as would be expected by chance. If Kappa = -1, then there
is perfect disagreement. “Rule-of-thumb” interpretation guidelines: >= 0.9 very good agreement
(green); 0.7 to < 0.9 marginally acceptable, improvement should be considered (yellow); < 0.7
unacceptable (red). See Appendix Kappa for further details on the Kappa calculations and “rule-of-
thumb” interpretation guidelines.
Fleiss’ Kappa P-Value: H0: Kappa = 0. If P-Value < alpha (.05 for specified 95% confidence level),
reject H0 and conclude that agreement is not the same as would be expected by chance. Significant
P-Values are highlighted in red.
Fleiss' Kappa LC (Lower Confidence) and Fleiss' Kappa UC (Upper Confidence) limits use a kappa
normal approximation. Interpretation Guidelines: Kappa lower confidence limit >= 0.9: very good
agreement. Kappa upper confidence limit < 0.7: the attribute agreement is unacceptable. Wide
confidence intervals indicate that the sample size is inadequate.
In this example, we have marginal Within Appraiser Agreement for each of the appraisers.
141
SigmaXL: Measure Phase Tools
Each Appraiser vs. Standard Agreement is an assessment of each appraiser’s ratings across trials
compared to a known reference standard. This is analogous to Gage R&R Accuracy.
Tip: The Percent/CI Each Appraiser vs. Standard Agreement Graph can be used to compare
agreement to standard across the appraisers, but should not be used as an absolute measure of
agreement. Each Appraiser vs. Standard Agreement will decrease as the number of trials increase
because a match occurs only if an appraiser agrees with the standard consistently across all trials.
Use Fleiss’ Kappa in the Each Appraiser vs. Standard Agreement Table to determine the adequacy
of Each Appraiser versus Standard agreement.
# Matched: A match occurs only if an appraiser agrees with the standard consistently across all
trials.
Kappa is interpreted as above: >= 0.9 very good agreement (green); 0.7 to < 0.9 marginally
acceptable, improvement should be considered (yellow); < 0.7 unacceptable (red).
Appraisers A and C have marginal agreement versus the standard values. Appraiser B has very good
agreement to the standard.
142
SigmaXL: Measure Phase Tools
A Type I Error occurs when the appraiser consistently assesses a good part/sample as bad. "Good"
is defined by the user in the Attribute MSA analysis dialog.
A Mixed Error occurs when the assessments across trials are not identical.
All Appraisers vs. Standard Agreement is an assessment of all appraisers’ ratings across trials
compared to a known reference standard. This is analogous to Gage R&R Accuracy.
Kappa is interpreted as above: >= 0.9 very good agreement (green); 0.7 to < 0.9 marginally
acceptable (yellow); < 0.7 unacceptable (red).
Since the Between Appraiser Agreement and All Appraisers vs. Standard Agreement are
marginally acceptable, improvements to the attribute measurement should be considered. Look
for unclear or confusing operational definitions, inadequate training, operator distractions or poor
lighting. Consider the use of pictures to clearly define a defect.
143
SigmaXL: Measure Phase Tools
The Attribute Effectiveness Report is similar to the Attribute Agreement Report, but treats each
trial as an opportunity. Consistency across trials or appraisers is not considered. This has the
benefit of providing a Percent measure that is unaffected by the number of trials or appraisers.
Also, the increased sample size for # Inspected results in a reduction of the width of the Percent
confidence interval. The Misclassification report shows all errors classified as Type I or Type II.
Mixed errors are not relevant here. This report requires a known reference standard and includes:
Each Appraiser vs. Standard Effectiveness, All Appraisers vs. Standard Effectiveness, and
Effectiveness and Misclassification Summary.
Each Appraiser vs. Standard Effectiveness is an assessment of each appraiser’s ratings compared
to a known reference standard. This is analogous to Gage R&R Accuracy. Unlike the Each Appraiser
vs. Standard Agreement table above, consistency across trials is not considered here - each trial is
considered as an opportunity. This has the benefit of providing a Percent measure that is
unaffected by the number of trials. Also, the increased sample size for # Inspected results in a
reduction of the width of the Percent confidence interval.
144
SigmaXL: Measure Phase Tools
# Matched: A match occurs if an appraiser agrees with the standard (consistency across trials is not
considered here).
Kappa is interpreted as above: >= 0.9 very good agreement (green); 0.7 to < 0.9 marginally
acceptable (yellow); < 0.7 unacceptable (red).
Tip: The Kappa values in the Effectiveness tables are very similar to those in the Agreement tables
(the slight difference is due to average Kappa for unstacked versus Kappa for stacked data). This is
why the Kappa/CI Each Appraiser vs. Standard Agreement graph is not shown. It would essentially
be a duplicate of the Kappa/CI Each Appraiser vs. Standard Effectiveness graph.
Appraisers A and C have marginal agreement versus the standard values, with less than 95%
Effectiveness and Kappa < 0.9. Appraiser B has very good agreement to the standard.
145
SigmaXL: Measure Phase Tools
All Appraisers vs. Standard Effectiveness is an assessment of all appraisers’ ratings compared to a
known reference standard. This is analogous to Gage R&R Accuracy. Unlike the All Appraiser vs.
Standard Agreement table above, consistency across trials and appraisers is not considered here -
each trial is considered as an opportunity. This has the benefit of providing a Percent measure that
is unaffected by the number of trials or appraisers. Also, the increased sample size for # Inspected
results in a reduction of the width of the Percent confidence interval.
# Matched: A match occurs if an appraiser agrees with the standard (consistency across trials and
appraisers is not considered here).
The interpretation guidelines for Kappa and Percent Effectiveness are the same as noted above.
This measurement system is marginally acceptable.
146
SigmaXL: Measure Phase Tools
A Type I error occurs when the appraiser assesses a good part/sample as bad (consistency across
trials is not considered here). "Good" is defined by the user in the Attribute MSA analysis dialog.
See Misclassification Legend for specific definition of Type I and Type II Errors.
A Type II error occurs when a bad part/sample is assessed as good. See Misclassification Legend
for specific definition of Type I and Type II Errors.
A Type I error occurs when the appraiser assesses a good part/sample as bad (consistency across
trials is not considered here). "Good" is defined by the user in the Attribute MSA analysis dialog.
See Misclassification Legend for specific definition of Type I and Type II Errors.
A Type II error occurs when a bad part/sample is assessed as good. See Misclassification Legend
for specific definition of Type I and Type II Errors.
147
SigmaXL: Measure Phase Tools
Effectiveness and Misclassification Summary is a summary table of all appraisers’ correct rating
counts and misclassification counts compared to the known reference standard values.
Attribute MSA Data is a summary showing the original data in unstacked format. This makes it
easy to visually compare appraiser results by part. If a reference standard is provided, the cells are
color highlighted as follows: agrees with reference (green); Type I error (yellow); Type II error (red):
In conclusion, with the Kappa scores in the “yellow zone” (< 0.9) and Percent Effectiveness less
than 95% this measurement system is marginal and should be improved. Appraiser B is the
exception and does well against the standard. Look for unclear or confusing operational definitions,
inadequate training, operator distractions or poor lighting. Consider the use of pictures to clearly
define a defect. Use Attribute MSA as a way to “put your stake in the ground” and track the
effectiveness of improvements to the measurement system.
148
SigmaXL: Measure Phase Tools
• Customer Loyalty –Likely to Recommend score which contains ordinal integer values from 1
to 5, where a 1 indicates that the customer is very unlikely to recommend and a 5 indicates
that the customer is very likely to recommend. This example is used in Ordinal Logistic
Regression.
• Taste Score on a scale of 1-7 where 1 is "awful" and 7 is "delicious." This is used in the cake
bake taste test Design of Experiments: Part B – Three Factor Full Factorial Example Using
DOE Template.
An Ordinal Attribute MSA study should be done prior to formal ordinal data collection for use in
hypothesis testing, regression or design of experiments.
1. Open the file Attribute MSA – Ordinal.xlsx. This is an Ordinal MSA example with 50 samples, 3
appraisers and 3 trials. The response is 1 to 5, grading product quality. One denotes “Very Poor
Quality,” 2 is “Poor,” 3 is “Fair,” 4 is “Good” and a 5 is “Very Good Quality.” The Expert
Reference column is the reference standard from an expert appraisal. Note that the worksheet
data must be in stacked column format and the reference values must be consistent for each
sample.
2. Click SigmaXL > Measurement Systems Analysis > Attribute MSA (Ordinal). Ensure that the
entire data table is selected. Click Next.
149
SigmaXL: Measure Phase Tools
3. Select Sample No., Appraiser, Assessed Result and Expert Reference as shown. Check Report
Information and enter Attribute MSA Ordinal for Product/Unit Name. Select Percent
Confidence Interval Type – Wilson Score:
Tip: While this report is quite extensive, a quick assessment of the attribute measurement
system can be made by viewing the Kendall Concordance and Kendall Correlation color
highlights: Green - very good agreement; Yellow - marginally acceptable, improvement should
be considered; Red - unacceptable. Further details on the Kendall Coefficients are given below.
Tip: Fleiss’ Kappa and Percent Agreement are included in the report for completeness but not
recommended for use with Ordinal response data because they treat each response level as
nominal. Kendall’s Concordance and Correlation take the order of the data into account, so a
deviation of 1 is not as bad as a deviation of 2 or more. See Attribute MSA – Nominal for a
discussion of the Fleiss’ Kappa report.
150
SigmaXL: Measure Phase Tools
Kendall's Concordance P-Value: H0: Kendall's Coefficient of Concordance = 0. If P-Value < alpha
(.05 for specified 95% confidence level), reject H0 and conclude that agreement is not the same as
would be expected by chance. Significant P-Values are highlighted in red.
The Within Appraiser Agreement for Joe is marginal, Moe is unacceptable and Sue is very good.
151
SigmaXL: Measure Phase Tools
Kendall's Correlation Coefficient (Kendall's tau-b) is a measure of association for discrete ordinal
data, used for assessments that include a known reference standard. Kendall’s correlation
coefficient ranges from -1 to 1: A coefficient value of 1 indicates perfect agreement. If the
coefficient = 0, then the agreement is random, i.e., the same as would be expected by chance. A
coefficient value of -1 indicates perfect disagreement. “Rule-of-thumb” interpretation guidelines:
>= 0.8 very good agreement (green); 0.6 to < 0.8 marginally acceptable, improvement should be
considered (yellow); < 0.6 unacceptable (red). See Appendix Kendall’s Correlation Coefficient for
further details on the Kendall Correlation calculations and “rule-of-thumb” interpretation
guidelines.
152
SigmaXL: Measure Phase Tools
Kendall's Correlation P-Value: H0: Kendall's Correlation Coefficient = 0. If P-Value < alpha (.05 for
specified 95% confidence level), reject H0 and conclude that agreement is not the same as would
be expected by chance. Significant P-Values are highlighted in red.
Kendall's Correlation LC (Lower Confidence) and Kendall's Correlation UC (Upper Confidence) limit
use a normal approximation. Interpretation Guidelines: Correlation lower confidence limit >= 0.8:
very good agreement. Correlation upper confidence limit < 0.6: the attribute agreement is
unacceptable. Wide confidence intervals indicate that the sample size is inadequate.
Tip: Kendall’s Correlation values in the Effectiveness tables are very similar to those in the
Agreement tables (the slight difference is due to average Kendall for unstacked versus Kendall for
stacked data). This is why the Kendall’s Correlation/CI Each Appraiser vs. Standard Agreement
graph is not shown. It would essentially be a duplicate of the Kendall’s Correlation/CI Each
Appraiser vs. Standard Effectiveness graph.
Appraiser Joe has marginal agreement versus the standard values. Appraiser Moe has
unacceptable agreement to the standard. Sue has very good agreement to the standard.
Note that the Percent Agreement results in All Appraisers vs. Standard Agreement Table show
only 2% agreement! This is due to the requirement that all appraisers agree with the standard
across all trials for a 5 level response, which is very unlikely to occur. This highlights the problem
with using Percent Agreement in an Ordinal MSA. Kendall’s coefficients are the key metric to assess
an Ordinal MSA.
153
SigmaXL: Measure Phase Tools
Effectiveness and Misclassification Summary is a summary table of all appraisers’ correct rating
counts and misclassification counts compared to the known reference standard values.
Attribute MSA Data is a summary showing the original data in unstacked format. This makes it
easy to visually compare appraiser results by part. If a reference standard is provided, the cells are
color highlighted as follows: absolute deviation = 0 (green); absolute deviation = 1 (yellow);
absolute deviation >=2 (red):
In conclusion, this measurement system is marginal and should be improved. Appraiser Moe needs
training and Appraiser Joe needs a refresher. Sue has very good agreement based on Kendall’s
Concordance and Correlation, but would have been considered marginal based on Kappa (< .9) and
Percent Effectiveness (< 95%). As discussed above, Kappa and Percent Effectiveness do not take
the order of the response data into account, so are not as useful as Kendall’s coefficients in an
Ordinal MSA study.
154
SigmaXL: Measure Phase Tools
1. Open the file Attribute MSA – Nominal.xlsx. This is a Nominal MSA example with 30 samples, 3
appraisers and 2 trials. The response is text “Type_1”, “Type_2” and “Type_3.” The Expert
Reference column is the reference standard from an expert appraisal. Note that the worksheet
data must be in stacked column format and the reference values must be consistent for each
sample.
2. Click SigmaXL > Measurement Systems Analysis > Attribute MSA (Nominal). Ensure that the
entire data table is selected. Click Next.
3. Select Sample No., Appraiser, Assessed Result and Expert Reference as shown. Check Report
Information and enter Attribute MSA – Nominal for Product/Unit Name. Select Percent
Confidence Interval Type – Wilson Score:
Tip: While this report is quite extensive, a quick assessment of the attribute measurement system
can be made by viewing the Kappa color highlights: Green - very good agreement (Kappa >= 0.9);
Yellow - marginally acceptable, improvement should be considered (Kappa 0.7 to < 0.9); Red -
unacceptable (Kappa < 0.7). See Attribute MSA (Binary) for a detailed discussion of the report
graphs and tables. Here we will just look at Kappa.
155
SigmaXL: Measure Phase Tools
Fleiss’ Kappa statistic is a measure of agreement that is analogous to a “correlation coefficient” for
discrete data. Kappa ranges from -1 to +1: A Kappa value of +1 indicates perfect agreement. If
Kappa = 0, then agreement is the same as would be expected by chance. If Kappa = -1, then there
is perfect disagreement. “Rule-of-thumb” interpretation guidelines: >= 0.9 very good agreement
(green); 0.7 to < 0.9 marginally acceptable, improvement should be considered (yellow); < 0.7
unacceptable (red). See Appendix Kappa for further details on the Kappa calculations and “rule-of-
thumb” interpretation guidelines.
Fleiss’ Kappa P-Value: H0: Kappa = 0. If P-Value < alpha (.05 for specified 95% confidence level),
reject H0 and conclude that agreement is not the same as would be expected by chance. Significant
P-Values are highlighted in red.
Fleiss' Kappa LC (Lower Confidence) and Fleiss' Kappa UC (Upper Confidence) limits use a kappa
normal approximation. Interpretation Guidelines: Kappa lower confidence limit >= 0.9: very good
agreement. Kappa upper confidence limit < 0.7: the attribute agreement is unacceptable. Wide
confidence intervals indicate that the sample size is inadequate.
Fleiss’ Kappa Overall is an overall Kappa for all of the response levels.
Fleiss’ Kappa Individual gives Kappa for each response level (Type_1, Type_2 and Type_3). This is
useful to identify if an appraiser has difficulty assessing a particular defect type.
This is a very good attribute measurement system, but Appraiser C is marginal so a refresher would
be helpful.
156
SigmaXL: Measure Phase Tools
These templates are also located at SigmaXL > Process Capability > Basic Process Capability
Templates.
See Measure Phase Part B – Templates and Calculators for Process Capability Template examples:
157
SigmaXL: Measure Phase Tools
2. Let’s begin with a view of the data using Histograms and Descriptive Statistics. Click SigmaXL >
Graphical Tools > Histograms & Descriptive Statistics.
3. Ensure that entire data table is selected. If not, check Use Entire Data Table. Click Next.
4. Select Delivery Time Deviation, click Numeric Data Variable (Y) >>. Click OK.
Note that the data has an average of 6, so on average the breakfast delivery is 6 minutes late.
Also note the wide variation in delivery times (StDev = 7.2 minutes). With the AD P-Value > .05,
this data can be assumed to have a normal distribution.
158
SigmaXL: Measure Phase Tools
5. Click Sheet 1 Tab. Click SigmaXL > Process Capability > Histograms & Process Capability (or
SigmaXL > Graphical Tools > Histograms & Process Capability).
6. Select Delivery Time Deviation, click Numeric Data Variable (Y) >>. Enter USL = 10, Target = 0,
LSL = -10, check Normal Curve as shown below:
7. Click OK. Resulting Histogram & Process Capability report is shown below:
With the Process Performance indices Pp and Ppk < 1, this process is clearly in need of
improvement. Note that the difference between Pp and Ppk is due to the off-center process
mean. (Cp and Cpk indices are provided in the Capability Combination Report or optionally
when creating Control Charts).
159
SigmaXL: Measure Phase Tools
3. Select Delivery Time Deviation, click Numeric Data Variable (Y) >>. Enter USL = 10, Target = 0,
LSL = -10, as shown below:
4. Click OK. The resulting Process Capability Combination report is shown below:
160
SigmaXL: Measure Phase Tools
5. To print this report, click SigmaXL Chart Tools > Print Charts. Do not use Excel’s print function.
6. Cp is the “best case” potential capability index. Ppk is the “worst case” actual performance. If
Cp is excellent (say Cp =2) and Ppk is poor (say Ppk < 1), this indicates that the process can be
dramatically improved through centering and stabilizing. If Cp is poor, you will have to
implement fundamental improvements using DMAIC to address common cause variation. If
Ppk = Cp, this indicates that the process is centered and stable, and the short-term (within
subgroup) variation is the same as the long-term (overall) variation.
7. Cpm is a similar statistic but it incorporates a penalty for deviation from target.
8. On a technical note regarding compatibility with other software tools, please be advised that
SigmaXL does not use unbiasing constants when calculating Overall StDev, Pp, Ppk and Cpm.
This is done in order to ensure that the calculated Overall StDev matches the results given in
Descriptive Statistics. Note that Within StDev, Cp and Cpk are based on control chart
techniques (MR-bar/d2 for Individuals Chart, R-bar/d2 for X-Bar & R, S-bar/c4 for X-Bar & S).
161
SigmaXL: Measure Phase Tools
2. Select B2:F22; here, we will only use the first 20 subgroups to determine the process capability
(subgroups 21 to 25 are studied later in Control Phase Tools: Statistical Process Control (SPC)
Charts: Part B - X-Bar & Range Charts).
3. Select SigmaXL > Process Capability > Capability Combination Report (Subgroups).
5. Click Next. Select Subgroups across Rows, select Shot 1, Shot 2, Shot 3, click Numeric Data
Variables (Y) >>. Select Operator, click Optional X-Axis Labels >>. Select X-Bar & R Charts.
Check Tests for Special Causes. Enter USL = 108, Target = 100, LSL = 92, as shown:
162
SigmaXL: Measure Phase Tools
6. Click OK. The resulting Process Capability Combination report is shown below:
163
SigmaXL: Measure Phase Tools
Box-Cox Transformation (includes an automatic threshold option so that data with negative values
can be transformed)
Johnson Transformation
Distributions supported:
Half-Normal
Logistic
For technical details, see Appendix: Statistical Details for Nonnormal Distributions and
Transformations. See also Andrew Sleeper, Six Sigma Distribution Modeling, for further
information on these methods.
Note that these transformations and distributions are particularly effective for inherently skewed
data but should not be used with bimodal data or where the nonnormality is due to outliers
(typically identified with a Normal Probability Plot). In these cases, you should identify the reason
for the bimodality or outliers and take corrective action. Another common reason for nonnormal
data is poor measurement discrimination leading to “chunky” data. In this case, attempts should
be made to improve the measurement system.
164
SigmaXL: Measure Phase Tools
Box-Cox Transformation
SigmaXL’s default setting is to use the Box-Cox transformation which is the most common approach
to dealing with nonnormal data. Box-Cox is used to convert nonnormal data to normal by applying
a power transformation, Y^lambda, where lambda varies from -5 to +5. You may select rounded or
optimal lambda. Rounded is typically preferred since it will result in a more “intuitive”
transformation such as Ln(Y) (lambda=0) or SQRT(Y) (lambda=0.5). If the data includes zero or
negative values, select Lambda & Threshold. SigmaXL will solve for an optimal threshold which is a
shift factor on the data so that all of the values are positive.
1. Open the file Nonnormal Cycle Time2.xlsx. This contains continuous data of process cycle
times. The Critical Customer Requirement is: USL = 1000 minutes.
2. Let’s begin with a view of the data using Histograms and Descriptive Statistics. Click SigmaXL >
Graphical Tools > Histograms & Descriptive Statistics.
3. Ensure that entire data table is selected. If not, check Use Entire Data Table. Click Next.
4. Select Cycle Time (Minutes), click Numeric Data Variable (Y) >>. Click OK.
Clearly this is a process in need of improvement. To start, we would like to get a baseline
process capability. The problem with using regular Capability analysis is that the results will be
incorrect due to the nonnormality in the data. The Histogram and AD P-Value < .05 clearly
show that this data is not normal.
5. We will confirm the nonnormality by using a Normal Probability Plot. Click Sheet 1 Tab (or F4).
Click SigmaXL > Graphical Tools > Normal Probability Plots.
6. Ensure that the entire data table is selected. If not, check Use Entire Data Table. Click Next.
165
SigmaXL: Measure Phase Tools
7. Select Cycle Time (Minutes), click Numeric Data Variable (Y) >>. Click OK. A Normal Probability
Plot of Cycle Time data is produced:
2.33
1.83
1.33
0.83
NSCORE
0.33
-0.17
-0.67
-1.17
-1.67
-2.17
-4000
-2000
2000
4000
6000
8000
Cycle Time (Minutes)
8. The curvature in this normal probability plot confirms that this data is not normal.
9. For now, let us ignore the nonnormal issue and perform a Process Capability study assuming a
normal distribution. Click Sheet 1 Tab. Click SigmaXL > Process Capability > Capability
Combination Report (Individuals).
10. Select Cycle Time (Minutes), click Numeric Data Variable (Y) >>. Enter USL = 1000; delete
previous Target and LSL settings.
166
SigmaXL: Measure Phase Tools
11. Click OK. The resulting Process Capability Report is shown below:
167
SigmaXL: Measure Phase Tools
Notice the discrepancy between the Expected Overall (Theoretical) Performance and Actual
(Empirical) Performance. This is largely due to the nonnormality in the data, since the expected
performance assumes that the data is normal. So why not just use the actual performance and
disregard the expected? This would not be reliable because the sample size, n = 30, is too small to
estimate performance using pass/fail (discrete) criteria.
Also note that the process appears to be out-of-control on both the individuals and moving range
charts.
12. We will now perform a process capability analysis using the Capability Combination Report for
Nonnormal Individuals. Click Sheet 1 Tab (or F4). Click SigmaXL > Process Capability >
Nonnormal > Capability Combination Report (Individuals Nonnormal). Ensure that the entire
data table is selected. If not, check Use Entire Data Table. Click Next.
13. Select Cycle Time (Minutes), click Numeric Data Variable (Y) >>. Enter USL = 1000. We will use
the default selection for Transformation/Distribution Options: Box-Cox Transformation with
Rounded Lambda. Check Tests for Special Causes as shown:
168
SigmaXL: Measure Phase Tools
14. Click OK. The resulting Process Capability Combination report is shown below:
169
SigmaXL: Measure Phase Tools
The AD Normality P-Value Transformed Data value of 0.404 confirms that the Box-Cox
transformation to normality was successful. The process capability indices and expected
performance can now be used to establish a baseline performance. Note that there are no out-of-
control signals on the control charts, so the signals observed earlier when normality was assumed
were false alarms.
The Individuals – Original Data chart displays the untransformed data with control limits calculated
as:
CL = 50th percentile
The benefit of displaying this chart is that one can observe the original untransformed data. Since
the control limits are based on percentiles, this represents the overall, long term variation rather
than the typical short term variation. The limits will likely be nonsymmetrical.
The Individuals/Moving Range – Normalized Data chart displays the transformed z-values with
control limits calculated using the standard Shewhart formulas for Individuals and Moving Range
charts. The benefit of using this chart is that tests for special causes can be applied and the control
limits are based on short term variation. The disadvantage is that one is observing transformed
data on the chart rather than the original data.
Now we will redo the capability analysis using the Automatic Best Fit option.
15. Click Recall SigmaXL Dialog menu or press F3 to recall last dialog. Select Automatic Best Fit as
shown:
170
SigmaXL: Measure Phase Tools
16. Click OK. The resulting Process Capability Combination report is shown below. Please note that
due to the extensive computations required, this could take up to 1 minute (or longer for large
datasets):
171
SigmaXL: Measure Phase Tools
The 2 Parameter Loglogistic distribution was selected as the best fit distribution. For details on
how this selection was made, see Appendix: Statistical Details for Nonnormal Distributions
and Transformations.
The Anderson Darling statistic for the Loglogistic distribution is 0.245 which is less than the 0.37
value for the AD Normality test of the Box-Cox transformation indicating a better fit. (Note that
published AD P-Values for this distribution are limited to a maximum value of 0.25. The best fit
selection uses a P-Value estimate that is obtained by transforming the data to normality and
then using a modified Anderson Darling Normality test on the transformed data).
Distribution Fitting
Another helpful tool to evaluate transformations and distributions is Distribution Fitting.
1. Click Sheet 1 Tab (or F4). Click SigmaXL > Process Capability > Nonnormal > Distribution
Fitting. Ensure that the entire data table is selected. If not, check Use Entire Data Table. Click
Next.
2. Select Cycle Time (Minutes), click Numeric Data Variable (Y) >>. We will use the default
selection for Transformation/Distribution Options: All Transformations & Distributions as
shown:
172
SigmaXL: Measure Phase Tools
3. Click OK. The resulting Distribution Fitting report is shown below. Please note that due to the
extensive computations required, this could take up to 1 minute (or longer for large datasets):
173
SigmaXL: Measure Phase Tools
The distributions and transformations are sorted in descending order using the AD Normality P-
Value on the transformed z-score values. Note that the first distribution shown may not be the
selected “best fit”, because the best fit procedure also looks for models that are close but with
fewer parameters.
The reported AD P-Values are those derived from the particular distribution. The AD P-Value is not
available for distributions with a threshold (except Weibull), so the AD Normality P-Value on the
transformed z-score values is used (labeled as Z-Score Est.).
Since the sort order is based on the AD P-Values from Z-Score estimates, it is possible that the
reported distribution based AD P-Values may not be in perfect descending order. However any
discrepancies based on sort order will likely not be statistically or practically significant.
Some data will have distributions and transformations where the parameters cannot be solved
(e.g., 2-parameter Weibull with negative values). These are excluded from the Distribution Fitting
report.
The parameter estimates and percentile report includes a confidence interval as specified in the
Distribution Fitting dialog, with 95% being the default. Note that the wide intervals here are due
to the small sample size, n = 30.
The control limits for the percentile based Individuals chart will be the 0.135% (lower control limit),
50% (center line, median) and 99.865% (upper control limit). Additional percentiles may be entered
in the Distribution Fitting dialog.
After reviewing this report, if you wish to perform a process capability analysis with a particular
transformation or distribution, simply select Specify Distribution from the
Transformation/Distribution Options in the Capability Combination Report (Individuals -
Nonnormal) dialog as shown below (using 2 Parameter Loglogistic):
174
SigmaXL: Measure Phase Tools
175
SigmaXL: Measure Phase Tools
Box-Cox Transformation
This is a standalone tool that allows you to visually see how the Box-Cox transformation selects a
rounded or optimal lambda value.
2. Click SigmaXL > Process Capability > Nonnormal > Box-Cox Transformation (or SigmaXL > Data
Manipulation > Box-Cox Transformation or SigmaXL > Control Charts > Nonnormal > Box-Cox
Transformation). Ensure that the entire data table is selected. If not, check Use Entire Data
Table. Click Next.
3. Select Cycle Time (Minutes), click Numeric Data Variable (Y) >>. The selected variable must
contain all positive values.
Tip: Note that while this tool is often successful to transform the data to normality, there may
not be a suitable transformation to make the data normal. The output report indicates the
Anderson-Darling P-Value for the transformed data. You may wish to check Do not store if
transformed data is not normal. Another option is Do not store transformed data if Lambda =
1 falls within 95%CI. This latter option prevents you from using transformations that do not
result in a statistically significant improvement towards normality.
176
SigmaXL: Measure Phase Tools
5. The fact that Lambda=1 falls outside of the 95% confidence interval tells us that the
transformation is statistically significant. The Anderson-Darling P-Value of 0.4041 indicates that
we cannot reject the null hypothesis that the transformed data is normal, so the Ln
transformation has successfully converted the data to normality.
177
SigmaXL: Measure Phase Tools
Reliability/Weibull Analysis
The Weibull distribution is a continuous distribution that was publicized by Waloddi Weibull in
1951. It has become widely used, especially in the reliability field. The Weibull distribution's
popularity resulted from its ability to be used with small sample sizes and its flexibility. See
Dodson, B. “The Weibull Analysis Handbook” Second Edition, ASQ, for further information. For
SigmaXL technical details, see Appendix: Statistical Details for Nonnormal Distributions and
Transformations.
The Weibull distribution can also be used as an alternative to the Box-Cox Transformation to
determine Process Capability Indices or Control Limits for nonnormal data.
Weibull is particularly effective when the data are censored, with exact failure times unknown or
suspension of test. Suspension of test is also referred to as right censored. SigmaXL can analyze
complete or right censored data.
1. Open the file Weibull Dodson.xlsx. This contains right censored time-to-failure data from
Dodson, page 28.
2. Click SigmaXL > Reliability/Weibull Analysis. Ensure that the entire data table is selected. Click
Next.
3. Select Time-to-Fail, click Numeric Response (Y) >>. Note that the selected variable must
contain all positive values. Select Censor Code, click Right Censor Column >>. The default
censor value of 1 will be used. We will use the default method of estimation, Maximum
Likelihood (some reliability practitioners prefer Least Squares Regression (X on Y) for small
sample sizes). Other defaults include Threshold = 0, and Confidence Level = 90%. Enter time
values to estimate survival probabilities as shown:
178
SigmaXL: Measure Phase Tools
179
SigmaXL: Measure Phase Tools
0.500
0.000
-0.500
WEIBULL SCORE
-1.000
-1.500
-2.000
-2.500
-3.000
-3.500
3.70
3.90
4.10
4.30
4.50
4.70
4.90
5.10
5.30
5.50
Ln (Time-to-Fail)
180
SigmaXL: Measure Phase Tools
5. The Hollander-Proschan test is used when there are right censored observations (see Dodson).
If the P-Value is less than 0.05, this indicates significant lack of fit, so Weibull would not be an
appropriate distribution. Given the P-Value of 0.9771, there is no lack-of-fit with this time-to-
fail data. This is confirmed in the Weibull probability plot with the data roughly following a
straight line.
7. The shape parameter, also referred to as Beta, is the slope of the line on the Weibull probability
plot. The Weibull distribution can be used in a wide variety of situations and dependent on the
value of Beta, is equal to or can approximate several other distributions. For example, if Beta =
1, the Weibull distribution is identical to the exponential distribution; if Beta = 2.5, the Weibull
distribution approximates the lognormal distribution; and if Beta = 3.6, the Weibull distribution
approximates the normal distribution. Because of this flexibility, there are few observed failure
rates that cannot be accurately modeled by the Weibull distribution. Furthermore if Beta is < 1,
the system is suffering from infant mortality failures; if Beta is = 1, failures are random – they
are occurring at a constant failure rate and are independent of time; if Beta is > 1, components
are wearing out with age.
8. The scale parameter is the characteristic life. This is the point at which we could expect 63.2%
of the population under study to have failed. The survival probability report shows that the
survival probability at time 200 = 0.386. The cumulative failure probability at 200 is 1 – 0.386 =
0.614.
9. The percentile report tells us that we can expect a 50% failure rate (one-half of the population
will fail) at time = 179.73, i.e. the Median = 179.73.
10. The percentile report can also be used to determine control limits for nonnormal data. The
lower control limit is the 0.135 % value = 22.052; the center line is the 50% value = 179.73; the
upper control limit is the 99.865 % value = 383.55.
181
SigmaXL: Analyze Phase Tools
SigmaXL’s Pareto tool allows you to create Basic (Single) or Advanced (Multiple) Pareto Charts.
Advanced Pareto charts are particularly useful in the Analyze Phase because of the ease with which
you can slice and dice (or stratify) your data. Of course, Pareto charts are not limited to the
Analyze Phase – they can also be used to aid project selection and to prioritize in the Measure
Phase.
Consider the following guidelines to help ensure that your Pareto analysis is successful:
• Your Pareto analysis will only be as good as the quality of the data collected. Ensure that
you have the right data and that the data is correct. Use other graphs such as run charts to
apply a sanity check to your data.
• Check process stability using appropriate control charts. If the process is not in control,
your prioritization of defects and root causes could be invalid.
• Avoid collecting data over too short a time period. Your data may not be representative of
the process as a whole. Also keep in mind that since the data is discrete, a minimum sample
size of 500 is recommended with 1000 preferred.
• Conversely, data gathered over too long a time period may include process changes that
could lead to incorrect conclusions. SigmaXL provides a date subsetting feature that allows
you to easily explore different time periods.
• If your initial Pareto analysis does not yield useful results, explore other categories that may
be important. SigmaXL’s Advanced Charts makes it easy for you to 'slice and dice' your data
with different X categories.
• Consider Pareto charting measures such as cost and severity, in addition to defect counts.
SigmaXL enables you to chart multiple Y responses.
185
SigmaXL: Analyze Phase Tools
2. Ensure that entire data table is selected. If not, check Use Entire Data Table. Click Next.
Tip: SigmaXL will automatically count the number of unique items in the Pareto Category. If we
had a separate column with a count (or cost), this count column would be selected as the
Optional Numeric Count (Y).
186
SigmaXL: Analyze Phase Tools
Tip: After you have saved your defaults, you can bypass the above options, by clicking Finish
instead of Next at the original Basic Pareto Chart dialog box. The saved defaults will
automatically be applied.
Note: For SigmaXL Mac version, click SigmaXL > Graphical Tools > Basic Pareto Chart Options
to set options. The Basic Pareto Chart is displayed after you click Finish.
187
SigmaXL: Analyze Phase Tools
3. Set Order of Bars to Same Order on the “Other Bar”/Cum Sum options tab.
This is typically used for comparative purposes. The Descending Order option makes each
Chart a true Pareto Chart, but is less useful for comparison.
4. Click Chart Options tab. Set according to choice – in this case we have selected Data Labels for
the Bars but not for the Cum Sum line.
5. Ensure that Save Defaults is checked. Note that these options will be saved and applied to all
Advanced Pareto Charts. Click Finish
6. SigmaXL automatically takes you to the next step of Chart Generation (This is equivalent to
clicking SigmaXL > Graphical Tools > Advanced Pareto Charts). If necessary, check Use Entire
Data Table.
7. Click Next.
188
SigmaXL: Analyze Phase Tools
8. Select Major Complaint, click Pareto Category (X1) >>; select Customer Type, click Group
Category (X2) >>.
9. Click OK. A Pareto Chart of Major Customer Complaints is produced for each Customer Type.
10. Click Sheet 1 Tab, Click SigmaXL > Graphical Tools > Advanced Pareto Charts.
11. Ensure that entire data table is selected. If not, check Use Entire Data Table. Click Next. (Steps
10 and 11 can be bypassed with the Recall SigmaXL Dialog menu or by pressing F3 to recall last
dialog).
189
SigmaXL: Analyze Phase Tools
12. Select Major Complaint, click Pareto Category (X1) >>; select Size of Customer, click Group
Category (X2) >>; select Product Type, click Group Category (X3) >>.
If a Numeric Count (Y) variable is not specified, SigmaXL automatically determines the counts
from the Pareto Category (X1).
Normally we would use a text column of discrete Xs, but be aware that numeric columns are
also allowed. Be careful here – this could easily generate a very large number of charts.
The total number of charts generated = (# of levels in X2) * (# of levels in X3) * (# of levels in X4)
* (# of Y variables). Please note, Group Category (X3) and Group Category (X4)
190
SigmaXL: Analyze Phase Tools
Part B - EZ-Pivot
One of the most powerful features in Excel is the Pivot table. SigmaXL’s EZ-Pivot tool simplifies the
creation of Pivot tables using the familiar X and Y dialog box found in the previous Pareto tools.
2. Ensure that entire data table is selected. If not, check Use Entire Data Table. Click Next.
3. Select Major Complaint, click Count Category (X1) >>. Note that if Y is not specified, the Pivot
Table Data is based on a count of X1, hence the name Count Category.
4. Select Customer Type, click Group Category (X2) >>; select Size of Customer, click Group
Category (X3) >> as shown.
5. Click OK. Resulting Pivot Table of Major Complaint by Customer Type is shown:
191
SigmaXL: Analyze Phase Tools
6. This Pivot table shows the counts for each Major Complaint (X1), broken out by Customer Type
(X2), for all Sizes of Customers (X3). (Grand Totals can be added to the Pivot Table by using
Pivot Table Toolbar > Table Options. Check Grand Totals for Columns, Grand Totals for Rows).
7. To display counts for a specific Customer Size, click the arrow adjacent to Size of Customer
(All). Select Large.
Note that the Major Complaint “Not-Available” is not shown. Pivot tables only show rows
where there is at least a count of one.
9. The Pivot Chart can be seen by clicking the EZ Pivot Chart (1) tab; reset Size of Customer to All
as shown below:
192
SigmaXL: Analyze Phase Tools
Count of Major-Complaint
70
60
50
40 Customer Type
3
2
30 1
20
10
0
Difficult-to-order Not-available Order-takes-too-long Return-calls Wrong-color
Major-Complaint
10. Drag Size of Customer from Report Filter to Axis Fields (shown under PivotChart Fields to the
right of the chart) and Excel will automatically split the Pivot Chart showing both Large and
Small Customers.
Count of Major-Complaint
40
35
30
25
Customer Type
3
20
2
1
15
10
0
Difficult-to- Order-takes- Return-calls Wrong-color Difficult-to- Not-available Order-takes- Return-calls Wrong-color
order too-long order too-long
Large Small
193
SigmaXL: Analyze Phase Tools
2. Select Customer Type, click Count Category (X1) >>; select Size of Customer, click Group
Category (X2) >>; select Product Type, click Group Category (X3); select Overall Satisfaction,
click Numeric Responses (Y) >>. Note that the Label for X1 changed from Count Category to
Group Category. The Pivot Table data will now be based on Y data.
3. The Response default uses a Sum of Y. This however can be changed to Average or Standard
Deviation. Select Average. Uncheck Create Pivot Charts (Since we are looking at averages, the
stacked bar Pivot Charts would not be very useful, unless they are changed to clustered column
format using Chart > Chart Type).
5. Note that the table now contains Averages of the Customer Satisfaction scores (Y). Again
Product Type (X3) can be varied to show Consumer, Manufacturer, or All. Double clicking on
Average of Overall Satisfaction allows you to switch to Standard Deviation (StDev).
194
SigmaXL: Analyze Phase Tools
2. Select Customer Type, click Group Category (X1) >>; select Size of Customer, click Group
Category (X2) >>; select Product Type, click Group Category (X3) >>. Select Avg Days Order to
Delivery, Loyalty – Likely to Respond, Overall Satisfaction, click Numeric Responses (Y) >>.
Select Average and One Pivot Table (default is separate Pivot Tables for each Y). Uncheck
Create Pivot Charts.
195
SigmaXL: Analyze Phase Tools
These confidence interval templates are also located at SigmaXL > Statistical Tools > Basic
Statistical Templates.
See Measure Phase Part B – Templates and Calculators for Confidence Interval template
examples.
4. Select Overall Satisfaction, click Numeric Data Variables (Y) >>. Select Customer Type, click
Group Category (X1) >>; Confidence Level default is 95%:
196
SigmaXL: Analyze Phase Tools
5. Click OK. Descriptive Statistics are given for Customer Satisfaction grouped by Customer Type:
We are given the 95% confidence interval for each sample Mean (95% CI Mean) as well as the
95% confidence interval for the Standard Deviation (95% CI Sigma – do not confuse this with
Process Sigma Quality Level).
These confidence intervals are very important in understanding our data and making decisions
from the data. How often are we driven by sample estimates only and fail to consider the
confidence interval or margin of error? For example, newspapers will often fail to take into
account the confidence interval when reporting opinion poll results. (To calculate confidence
intervals for discrete proportion data, use SigmaXL > Templates and Calculators > Basic
Statistical Templates > 1 Proportion Test and Confidence Interval).
Note that a confidence interval of 95% implies that, on average, the true population parameter
(Mean, Median, Standard Deviation, or Proportion) will lie within the interval 19 times out of
20.
A confidence interval or margin of error does not take into account measurement error or
survey bias, so the actual uncertainty may be greater than stated. This should be addressed
with good data collection, reliable measurement systems, and good survey design.
Confidence Intervals for the Mean can be obtained in several ways with SigmaXL: Descriptive
Statistics, Histograms & Descriptive Statistics, 1-Sample t-test & Confidence Intervals, One-Way
ANOVA, and Multi-Vari Charts.
To illustrate confidence intervals for the mean of Overall Satisfaction graphically, we have
generated a Multi-Vari Chart (with 95% CI Mean Options) using the Customer Data.xlsx data.
This chart type will be covered later (Part Q).
197
SigmaXL: Analyze Phase Tools
4.5
4
Overall Satisfaction (Mean Options)
3.5
2.5
1.5
1
#1 #2 #3
Customer Type
The dots correspond to individual data points. The tick marks show the 95% upper confidence
limit, mean, and 95% lower confidence limit. Clearly we can see that Customer Type 2 has a
significantly higher level of mean satisfaction; the lower limit does not overlap with the upper
limit for Types 1 and 3. On the other hand, we see overlap of the CI’s when comparing types 1
and 3. Hypothesis testing will now be used to compare the mean satisfaction scores more
precisely and determine statistical significance for the results.
198
SigmaXL: Analyze Phase Tools
These templates are also located at SigmaXL > Statistical Tools > Basic Statistical Templates.
See Measure Phase Part B – Templates and Calculators for One Sample Z, t and Equivalence
Test template examples:
Basic Statistical Templates – 1 Sample Z-Test and Confidence Interval for Mean
Basic Statistical Templates – 1 Sample t-Test and Confidence Interval for Mean
199
SigmaXL: Analyze Phase Tools
2. Ensure that Stacked Column Format is selected. Select Overall Satisfaction, click Numeric Data
Variable (Y) >>, select Customer Type, click Optional Group Category (X) >>.
3. Historically, our average customer satisfaction score has been 3.5. We would like to see if this
has changed, with the results grouped by customer type. Null Hypothesis H0: μ=3.5;
Alternative Hypothesis Ha: μ≠3.5
4. Enter 3.5 for the Null Hypothesis H0: Mean value. Set Ha as Not Equal To, Confidence Level =
95.0%, and check Display Test Assumptions Report.
200
SigmaXL: Analyze Phase Tools
6. Note the P-Values. Customer Type 2 shows a significant change (increase) in Satisfaction Mean
(P-Value < .05), whereas Customer Types 1 and 3 show no change (P-Value ≥ .05). Also note the
confidence intervals around each mean match the results from Descriptive Statistics.
7. In the Measure Phase we determined that Overall Satisfaction for Customer Type 2 has
nonnormal data but this does not imply that the P-Value for the 1 Sample t-test is wrong. The
Central Limit Theorem applies here: the distribution of averages tends to be normal, even if the
individual observations are not-normal. With a sample size of 42, the t-test is fairly robust
against skewed data.
This is a text report with color highlight: Green (OK), Yellow (Warning) and Red (Serious
Violation).
Each sample is tested for Normality using the Anderson Darling test. If not normal, the
minimum sample size for robustness of the 1 sample t-Test is determined utilizing Monte Carlo
201
SigmaXL: Analyze Phase Tools
regression equations (see Basic Statistical Templates – Minimum Sample Size for Robust t-
Tests and ANOVA). If the sample size is inadequate, a warning is given and a suitable
Nonparametric Test is recommended (Wilcoxon if symmetric, Sign Test if not symmetric).
Each sample is tested for Outliers defined as: Potential: Tukey's Boxplot (> Q3 + 1.5*IQR or < Q1
– 1.5*IQR); Likely: Tukey's Boxplot 2.2*IQR; Extreme: Tukey's Boxplot 3*IQR. If outliers are
present, a warning is given and recommendation to review the data with a Boxplot and Normal
Probability Plot. Here we have a potential outlier for Customer Type 2.
Tip: If the removal of outlier(s) result in an Anderson Darling P-Value that is > 0.1, a notice is
given that excluding the outlier(s), the sample data are inherently normal.
Each sample is also tested for Randomness using the Exact Nonparametric Runs Test. If the
sample data is not random, a warning is given and recommendation to review the data with a
Run Chart.
202
SigmaXL: Analyze Phase Tools
Tip: Typical sample size rules of thumb address confidence interval size and robustness to
normality (e.g. n=30). Computing power is more difficult because it involves the magnitude of
change in mean to be detected, so one needs to use the power and sample size calculator.
Please use the following guidelines when using the power and sample size calculator:
Power >= .99 (Beta Risk is <= .01) is considered Very High Power
Power >= .95 and < .99 (Beta Risk is <= .05) is High Power
Power >= .8 and < .95 (Beta Risk is <= .2) is Medium Power. Typically a power value of .9 to detect
a difference of 1 standard deviation is considered adequate for most applications. If the data
collection is difficult and/or expensive, then .8 might be used.
Power >= .5 and < .8 (Beta Risk is <= .5) is Low Power – not recommended.
Power < .5 (Beta Risk is > .5) is Very Low Power – do not use!
1. Click SigmaXL > Statistical Tools > Power and Sample Size Calculators > 1 Sample t-Test
Calculator. We will only consider the statistics from Customer Type 3 here. We will treat the
problem as a two sided test with Ha: Not Equal To to be consistent with the original test.
2. Enter 27 in Sample Size (N). The difference to be detected in this case would be the difference
between the sample mean and the hypothesized value i.e. 3.6411 – 3.5 = 0.1411. Enter 0.1411
in Difference. Leave Power value blank, with Solve For Power selected (default). Given any
two values of Power, Sample size, and Difference, SigmaXL will solve for the remaining selected
third value. Enter the sample standard deviation value of 0.6705 in Standard Deviation. Keep
Alpha and Ha at the default values as shown:
203
SigmaXL: Analyze Phase Tools
4. A power value of 0.1836 is very poor. It is the probability of detecting the specified difference.
Alternatively, the associated Beta risk is 1-0.1836 = 0.8164 which is the probability of failure to
detect such a difference. Typically, we would like to see Power > 0.9 or Beta < 0.1. In order to
detect a difference this small we would need to increase the sample size. We could also set the
difference to be detected as a larger value.
5. First we will determine what sample size would be required in order to obtain a Power value of
0.9. Click Recall SigmaXL Dialog menu or press F3 to recall last dialog. Select the Solve For
Sample Size button as shown. It is not necessary to delete the entered sample size of 27 – it
will be ignored. Enter a Power Value of .9:
204
SigmaXL: Analyze Phase Tools
7. A sample size of 240 would be required to obtain a power value of 0.9. The actual power is
rarely the same as the desired power due to the restriction that the sample size must be an
integer. The actual power will always be greater than or equal to the desired power.
8. Now we will determine what the difference would have to be to obtain a Power value of 0.9,
given the original sample size of 27. Click Recall SigmaXL Dialog menu or press F3 to recall last
dialog. Select the Solve For Difference button as shown:
205
SigmaXL: Analyze Phase Tools
10. A difference of 0.435 would be required to obtain a Power value of 0.9, given the sample size of
27.
206
SigmaXL: Analyze Phase Tools
1. Open the file Sample Size and Difference Worksheet.xlsx, select the Sample Size & Diff sheet
tab. Click SigmaXL > Statistical Tools > Power & Sample Size with Worksheets > 1 Sample t-
Test. If necessary, check Use Entire Data Table. Click Next.
2. Ensure that Solve For Power (1 – Beta) is selected. Select Sample Size (N) and Difference
columns as shown. Enter the Standard Deviation value of 1. Enter .05 as the Significance Level
value:
207
SigmaXL: Analyze Phase Tools
4. To create a graph showing the relationship between Power, Sample Size and Difference, click
SigmaXL > Statistical Tools > Power & Sample Size Chart. Check Use Entire Data Table. Click
Next.
5. Select Power (1 – Beta), click Y Axis (Y); select Sample Size (N), click X Axis (X1); select
Difference, click Group Category (X2). Click Add Title. Enter Power & Sample Size: 1 Sample t-
Test:
6. Click OK. Click OK. The resulting Power & Sample Size Chart is displayed:
208
SigmaXL: Analyze Phase Tools
Nonparametric tests use Medians and Ranks, thus they are robust to outliers in the data. If,
however, the data are normal and free of outliers, nonparametric tests are less powerful than
normal based tests to detect a real difference when one exists.
Nonparametric tests should be used when the data are nonnormal, the data cannot be readily
transformed to normality, and the sample size is too small for robustness due to the Central Limit
Theorem.
It is important to note that while exact P-Values are “correct,” they do not increase (or decrease)
the power of a small sample test, so they are not a solution to the problem of failure to detect a
change due to inadequate sample size.
For data that require more computation time than specified, Monte Carlo P-Values provide an
approximate (but unbiased) P-Value that typically matches exact to two decimal places using
10,000 replications. One million replications give a P-Value that is typically accurate to three
decimal places. A confidence interval (99% default) is given for the Monte Carlo P-Values. Note
that the Monte Carlo confidence interval for P-Value is not the same as a confidence interval on
the test statistic due to data sampling error. The confidence level for the hypothesis test statistic is
still 95%, so all reported P-Values less than .05 will be highlighted in red to indicate significance.
The 99% Monte Carlo P-Value confidence interval is due to the uncertainty in Monte Carlo
sampling, and it becomes smaller as the number of replications increases (irrespective of the data
209
SigmaXL: Analyze Phase Tools
sample size). The Exact P-Value will lie within the stated Monte Carlo confidence interval 99% of
the time.
These are sample size guidelines for when exact nonparametric tests should be used rather than
“large sample” asymptotic based on the Normal or Chi-Square approximation. It is always
acceptable to use an exact test, but computation time can become an issue especially for tests with
two or more samples. In those cases, one can always use a Monte Carlo P-Value with 99%
confidence interval.
SigmaXL Exact P-Values are validated by comparison to textbook examples, published paper
examples and other exact software such as StatXact, SPSS Exact, SAS, and Matlab.
Monte Carlo P-Values are validated using 1e6 replications and compared against exact. Repeated
simulations are used to validate the confidence intervals.
For further details and references refer to the Appendix Exact and Monte Carlo P-Values for
Nonparametric and Contingency Tests.
210
SigmaXL: Analyze Phase Tools
Historically, our Median customer satisfaction score has been 3.5. We would like to see if this has
changed, with the results grouped by customer type (H0: Median=3.5, Ha: Median≠3.5, α = 0.05).
1. Open Customer Data.xlsx, select Sheet 1 tab. Click SigmaXL > Statistical Tools >
Nonparametric Tests>1 Sample Sign. If necessary, check Use Entire Data Table, click Next.
2. Ensure that Stacked Column Format is selected. Select Overall Satisfaction, click Numeric Data
Variable (Y) >>; select Customer Type, click Optional Group Category (X) >>.
3. Enter 3.5 for the Null Hypothesis H0: Median value. Set Ha as Not Equal To.
211
SigmaXL: Analyze Phase Tools
Note the P-Values. Customer Type 2 shows a significant change (increase) in Satisfaction
Median (P-Value < .05), whereas Customer Types 1 and 3 show no change (P-Value ≥ .05).
While the P-Values are not the same as those given by the 1 sample t-Test, the conclusions do
match.
If Count (N) is less than or equal to 50, the Sign Test computes an exact P-Value using the
binomial distribution. For N > 50, the P-Value is estimated using a normal approximation. Since
this is always done automatically and is very fast, the Sign Test is not included in the separate
Nonparametric Exact menu.
The One Sample Wilcoxon Test is a nonparametric equivalent to the parametric One Sample t-Test.
Historically, our Median customer satisfaction score has been 3.5. We would like to see if this has
changed, with the results grouped by customer type (H0: Median=3.5, Ha: Median≠3.5, α = 0.05).
1. Open Customer Data.xlsx, select Sheet 1 tab (or press F4 to activate last worksheet). Click
SigmaXL > Statistical Tools > Nonparametric Tests > 1 Sample Wilcoxon. If necessary, check
Use Entire Data Table, click Next.
2. Ensure that Stacked Column Format is selected. Select Overall Satisfaction, click Numeric Data
Variable (Y) >>; select Customer Type, click Optional Group Category (X) >>.
3. Enter 3.5 for the Null Hypothesis H0: Median value. Keep Ha as Not Equal To.
212
SigmaXL: Analyze Phase Tools
Note the P-Values. Customer Type 2 shows a significant change (increase) in Satisfaction
Median (P-Value < .05), whereas Customer Types 1 and 3 show no change (P-Value ≥ .05).
Although the P-Values are not identical to the sign test and t-Test, the conclusions match.
(Note, in the case of Customer Type 2, the Sign Test is preferred since the data is not
symmetrical but skewed).
1. Open Customer Data.xlsx, select Sheet 1 tab (or press F4 to activate last worksheet). Click
SigmaXL > Statistical Tools > Nonparametric Tests – Exact > 1 Sample Wilcoxon - Exact. If
necessary, check Use Entire Data Table, click Next.
213
SigmaXL: Analyze Phase Tools
2. Ensure that Stacked Column Format is selected. Select Overall Satisfaction, click Numeric Data
Variable (Y) >>; select Customer Type, click Optional Group Category (X) >>.
3. Enter 3.5 for the Null Hypothesis H0: Median value. Keep Ha as Not Equal To.
The Wilcoxon Statistics are identical to the above “large sample” or “asymptotic” results. The
Exact P-Values are close but slightly different. This was expected because the sample sizes are
reasonable (N > 15), so the “large sample” P-Values are valid using a normal approximation for
the Wilcoxon Statistic.
Note, if Count (N) is greater than 1000, the Exact P-Value is estimated using a continuity -
corrected normal approximation. Since the Wilcoxon Exact P-Value is computed very quickly
for sample sizes as large as 1000, Monte Carlo P-Values are not required.
5. Now we will consider a small sample problem. Open Nonnormal Task Time Difference – Small
Sample.xlsx. A study was performed to determine the effectiveness of training to reduce the
time required to complete a short but repetitive process task. Fifteen operators were randomly
selected and the difference in task time was recorded in seconds (after training – before
training). A negative value denotes that the operator completed the task in less time after
training than before.
214
SigmaXL: Analyze Phase Tools
6. Click SigmaXL > Graphical Tools > Histograms & Descriptive Statistics. If necessary, check Use
Entire Data Table. Click Next. Select Difference (Seconds), click Numeric Data Variables (Y) >>.
Click OK.
4 Difference (Seconds)
Count = 15
3 Mean = -7.067
Stdev = 12.842
Range = 35.00
2 Minimum = -25
25th Percentile (Q1) = -20
50th Percentile (Median) = -2
1 75th Percentile (Q3) = 6
Maximum = 10
10.0
-25.0
-20.0
-15.0
-10.0
0.0
5.0
-5.0
This small sample data fails the Anderson Darling Normality Test (P-Value = .023). Note that
this is due to the data being uniform or possibly bimodal, not due to a skewed distribution.
Now we will perform a 1 Sample t-Test and review the assumptions.
7. Select Task Time Difference tab (or press F4 to activate last worksheet). Click SigmaXL >
Statistical Tools > 1 Sample t-Test & Confidence Intervals. If necessary, check Use Entire Data
Table, click Next.
8. Ensure that Stacked Column Format is selected. Select Difference (Seconds), click Numeric
Data Variable (Y) >>.
9. This is a one sided test because we have no reason to expect that the task time will increase, so
the Null Hypothesis H0: μ = 0; and the Alternative Hypothesis Ha: μ < 0. Enter 0 for the H0:
Mean value. Select Ha as Less Than, Confidence Level = 95.0%, and check Display Test
Assumptions Report.
215
SigmaXL: Analyze Phase Tools
The 1 Sample t-Test Assumptions Report highlights that the data are not normal, but note that
Kurtosis equal to -1.84 is the issue here, not Skewness. This was observed in the Histogram
above with the data being uniform or possibly bimodal.
The sample size is too small for a robust t-Test, so the Exact One Sample Wilcoxon Test is
recommended. The Wilcoxon Test is recommended over the Sign Test because it is a more
powerful test and meets the requirement that the data be symmetrical. The Exact test is
recommended because the sample size is very small (N <= 15).
216
SigmaXL: Analyze Phase Tools
11. Select Task Time Difference tab (or press F4 to activate last worksheet). Click SigmaXL >
Statistical Tools > Nonparametric Tests – Exact > 1 Sample Wilcoxon - Exact. If necessary,
check Use Entire Data Table, click Next.
12. Ensure that Stacked Column Format is selected. Select Difference (Seconds), click Numeric
Data Variable (Y) >>. Enter 0 for the H0: Median value. Select Ha as Less Than.
With the P-Value = .0497 we reject H0 and conclude that the Median Task Time Difference is
significantly less than 0, so the training is effective.
By way of comparison we will now rerun the analysis using the large sample (asymptotic)
Wilcoxon test.
14. Select Task Time Difference tab (or press F4 to activate last worksheet). Click SigmaXL >
Statistical Tools > Nonparametric Tests > 1 Sample Wilcoxon. If necessary, check Use Entire
Data Table, click Next.
15. Ensure that Stacked Column Format is selected. Select Difference (Seconds), click Numeric
Data Variable (Y) >>. Enter 0 for the H0: Median value. Select Ha as Less Than.
217
SigmaXL: Analyze Phase Tools
The difference between exact and large sample P-Value is small but it was enough to lead us to
falsely conclude that the training is ineffective.
In conclusion, whenever you have a small sample size and are performing a nonparametric test,
always use the Exact option.
218
SigmaXL: Analyze Phase Tools
These templates are also located at SigmaXL > Statistical Tools > Basic Statistical Templates.
See Measure Phase Part B – Templates and Calculators for Two Sample t-Test template
examples:
Basic Statistical Templates – 2 Sample t-Test and Confidence Interval (Compare 2 Means)
219
SigmaXL: Analyze Phase Tools
2. Click SigmaXL > Statistical Tools > 2 Sample t-test. If necessary, check Use Entire Data Table,
click Next.
3. With stacked column format checked, select Overall Satisfaction, click Numeric Data Variable Y
>>; select Customer Type, click Group Category X >>; H0: Mean Diff = 0; Ha: Not Equal To;
Confidence Level: 95%; ensure that Assume Equal Variances and Display Test Assumption
Report are checked:
220
SigmaXL: Analyze Phase Tools
Given the P-Value of .0007 we reject H0 and conclude that Mean Customer Satisfaction is
significantly different between Customer type 2 and 3. This confirms previous findings.
221
SigmaXL: Analyze Phase Tools
This is a text report with color highlight: Green (OK), Yellow (Warning) and Red (Serious
Violation).
Each sample is tested for Normality using the Anderson-Darling test. If not normal, the
minimum sample size for robustness of the 2 sample t-Test is determined utilizing Monte Carlo
regression equations (see Basic Statistical Templates – Minimum Sample Size for Robust t-
Tests and ANOVA). If the sample size is inadequate, a warning is given and the Nonparametric
Mann-Whitney Test is recommended.
Each sample is tested for Outliers defined as: Potential: Tukey's Boxplot (> Q3 + 1.5*IQR or < Q1
– 1.5*IQR); Likely: Tukey's Boxplot 2.2*IQR; Extreme: Tukey's Boxplot 3*IQR. If outliers are
present, a warning is given and recommendation to review the data with a Boxplot and Normal
Probability Plot. Here we have a potential outlier for Customer Type 2.
Tip: If the removal of outlier(s) result in an Anderson Darling P-Value that is > 0.1, a notice is
given that excluding the outlier(s), the sample data are inherently normal.
Each sample is tested for Randomness using the Exact Nonparametric Runs Test. If the sample
data is not random, a warning is given and recommendation to review the data with a Run
Chart.
222
SigmaXL: Analyze Phase Tools
A test for Equal Variances is also applied. If all sample data are normal, the F-Test is utilized,
otherwise Levene’s Test is used. If the variances are unequal and the test being used is the
equal variance option, then a warning is given and Welch’s test is recommended.
223
SigmaXL: Analyze Phase Tools
Paired t-Test
1. Open the file Dietcola.xlsx. These are the results of a Before and After taste test on sweetness
for diet cola. Ten tasters were used and one month elapsed with the cola in warm storage
between the before and after results. Do a one sample t-test on the column of differences.
2. Click SigmaXL > Statistical Tools > 1-Sample t-test & Confidence Intervals. Ensure that entire
data table is selected. If not, check Use Entire Data Table. Click Next. Select Difference, click
Numeric Data Variable (Y) >>, set H0: Mean = 0, Ha: Less Than (this is a one-sided or one-tail
test – sweetness cannot increase). Uncheck Display Test Assumptions Report for now:
Given the P-Value of .012, we reject H0 and conclude that the sweetness has in fact decreased.
4. Now redo the analysis using the paired t-test: Click Sheet 1 Tab; Click SigmaXL > Statistical
Tools > Paired t-Test; Click Next; Select After, click Numeric Data Variable 1 >>, select Before as
224
SigmaXL: Analyze Phase Tools
Numeric Data Variable 2 >>, H0: Mean Diff = 0, Ha: Less Than. Check Display Test
Assumptions Report.
5. Click OK. Results are identical to the One sample t-test analysis of difference column, with the
added assumptions report showing that all assumptions are met:
225
SigmaXL: Analyze Phase Tools
2. Click SigmaXL > Statistical Tools > 2 Sample t-Test. Ensure that entire data table is selected. If
not, check Use Entire Data Table. Click Next.
3. Check Unstacked Column Format. Select After, Before and click Numeric Data Variables (Y) >>.
H0: Mean Diff = 0, Ha: Less Than; check Assume Equal Variances. Uncheck Display Test
Assumptions Report:
Now the P-Value is .07, indicating a fail to reject H0. What changed? Hint: Compare the SE
Mean of the Paired t-test to the Std Error Difference of the unpaired two-sample t-test. Where
226
SigmaXL: Analyze Phase Tools
does the additional variability come from in the two-sample t-test? The paired t-test is the
appropriate test to use here.
1. Click SigmaXL > Statistical Tools > Power & Sample Size Calculators > 2 Sample t-Test
Calculator.
2. Select Solve For Power (1 – Beta). Enter Sample Size and Difference as shown:
Note that we are calculating the power or likelihood of detection given that Mean1 – Mean2 =
1, with sample size for each group = 30, standard deviation = 1, significance level = .05, and Ha:
Not Equal To (two-sided test).
A power value of 0.97 is good, hence we have the basis for the “minimum sample size n=30”
rule of thumb used for continuous data.
227
SigmaXL: Analyze Phase Tools
4. To determine Power & Sample Size using a Worksheet, click SigmaXL > Statistical Tools >
Power & Sample Size with Worksheet > 2 Sample t-Test.
5. A graph showing the relationship between Power, Sample Size and Difference can then be
created using SigmaXL > Statistical Tools > Power & Sample Size Chart. See Part E for an
example using the 1 Sample t-Test.
228
SigmaXL: Analyze Phase Tools
229
SigmaXL: Analyze Phase Tools
2. Click SigmaXL > Statistical Tools > 2 Sample Comparison Tests. Click Next. Check Stacked
Column Format. Select Overall Satisfaction, click Numeric Data Variable (Y) >>, select
Customer Type, click Group Category (X) >>.
230
SigmaXL: Analyze Phase Tools
4. Click OK.
Customer Type 2 has nonnormal data. This makes Levene’s test the appropriate test for
unequal variance. Levene’s test indicates that Customer type 2 has a significantly lower
variance, or standard deviation. The lower standard deviation translates to a consistent level of
satisfaction.
Since Levene’s test indicates unequal variance, the appropriate t-test assumes unequal
variance. The t-test indicates that Customer Type 2 has a significantly higher mean satisfaction.
Clearly the next step would be to determine a root cause or best practices to reduce the
variability in overall satisfaction and increase the mean for all customer types.
231
SigmaXL: Analyze Phase Tools
1. Open Customer Data.xlsx, click Sheet 1 tab (or press F4 to activate last worksheet).
2. Click SigmaXL > Statistical Tools > Nonparametric Tests > 2 Sample Mann-Whitney. If
necessary, check Use Entire Data Table, click Next.
3. With Stacked Column Format checked, select Overall Satisfaction, click Numeric Data Variable
(Y) >>; select Customer Type, click Group Category (X) >>; and Ha: Not Equal To:
232
SigmaXL: Analyze Phase Tools
Given the P-Value of .0008 we reject H0 and conclude that Median Customer Satisfaction is
significantly different between Customer types 2 and 3. This confirms previous findings and
matches the results of the 2 Sample t-Test.
1. Open Customer Data.xlsx, click Sheet 1 tab (or press F4 to activate last worksheet).
2. Click SigmaXL > Statistical Tools > Nonparametric Tests – Exact > 2 Sample Mann-Whitney –
Exact. If necessary, check Use Entire Data Table, click Next.
3. With Stacked Column Format checked, select Overall Satisfaction, click Numeric Data Variable
(Y) >>; select Customer Type, click Group Category (X) >> and Ha: Not Equal To. Select Monte
Carlo Exact with the default Number of Replications = 10000 and Confidence Level for P-Value
= 99%.
233
SigmaXL: Analyze Phase Tools
Tip 1: If Exact is selected and the computation time limit is exceeded, a dialog will prompt you
to use Monte Carlo or to increase the computation time.
Tip 2: 10,000 replications will result in a Monte Carlo P-Value that is typically correct to two
decimal places. One million (1e6) replications will result in three decimal places of accuracy
and typically require less than 60 seconds to solve for any data set.
Tip 3: The Monte Carlo 99% confidence interval for P-Value is not the same as a confidence
interval on the test statistic due to data sampling error. The confidence level for the hypothesis
test statistic is still 95%, so all reported P-Values less than .05 will be highlighted in red to
indicate significance. The 99% Monte Carlo P-Value confidence interval is due to the
uncertainty in Monte Carlo sampling, and it becomes smaller as the number of replications
increases (irrespective of the data sample size). The Exact P-Value will lie within the stated
Monte Carlo confidence interval 99% of the time.
234
SigmaXL: Analyze Phase Tools
Given the Monte Carlo P-Value of .0004 we reject H0 and conclude that Median Customer
Satisfaction is significantly different between Customer types 2 and 3. The Mann Whitney
Statistic is very close to the above “large sample” or “asymptotic” result. This was expected
because the sample size is reasonable (each sample N > 10), so the “large sample” P-Values are
valid using a normal approximation for the Mann-Whitney Statistic.
The Monte Carlo P-Value 99% confidence interval is 0.0000 to 0.0009. Note that the Monte
Carlo P-Value will be slightly different every time it is run (the Monte Carlo seed value is derived
from the system clock). This was demonstrated using 10,000 replications, but with a P-Value
this low, it is recommended that the number of replications be increased to 1e5 or 1e6 to get a
better estimate.
5. Now we will consider a small sample problem. Open Stimulant Test.xlsx. This data is from:
Narayanan, A. and Watts, D. “Exact Methods in the NPAR1WAY Procedure,” SAS Institute Inc.,
Cary, NC. http://support.sas.com/rnd/app/stat/papers/exact.pdf
6. Select Reaction Time tab. Click SigmaXL > Statistical Tools > Nonparametric Tests – Exact > 2
Sample Mann-Whitney - Exact. If necessary, check Use Entire Data Table, click Next.
7. With Stacked Column Format checked, select Reaction Time, click Numeric Data Variable (Y)
>>; select Stimulant, click Group Category (X) >>; and Ha: Less Than. Select Exact with the
default Time Limit for Exact Computation = 60 seconds.
235
SigmaXL: Analyze Phase Tools
This sets the order for the one sided test, so the alternative hypothesis Ha is Median 1 <
Median 2.
With the P-Value = .0527 we fail to reject H0, so cannot conclude that there is a difference in
median reaction times. This exact P-Value matches that given in the reference paper.
236
SigmaXL: Analyze Phase Tools
By way of comparison we will now rerun the analysis using the “large sample” or “asymptotic”
Mann-Whitney test.
10. Select Reaction Time tab (or press F4 to activate last worksheet). Click SigmaXL > Statistical
Tools > Nonparametric Tests > 2 Sample Mann-Whitney. If necessary, check Use Entire Data
Table, click Next.
11. With Stacked Column Format checked, select Reaction Time, click Numeric Data Variable (Y)
>>; select Stimulant, click Group Category (X) >>; and Ha: Less Than.
The difference between exact and large sample P-Value is small but it was enough to lead us to
falsely conclude that stimulant 1 resulted in a reduced median reaction time.
In conclusion, whenever you have a small sample size and are performing a nonparametric test,
always use the Exact option.
237
SigmaXL: Analyze Phase Tools
2. Open Customer Data.xlsx, click on Sheet 1 tab (or press F4 to activate last worksheet).
3. Click SigmaXL > Statistical Tools > One-Way ANOVA & Means Matrix. Ensure that the entire
data table is selected. If not, check Use Entire Data Table.
4. Click Next. Ensure that Stacked Column Format is checked. Select Overall Satisfaction, click
Numeric Data Variable (Y) >>; select Customer Type, click Group Category (X) >>. Uncheck
Display ANOVA Table Details. Check Display Test Assumptions Report.
238
SigmaXL: Analyze Phase Tools
6. The ANOVA P-Value of 0.0000 tells us that at least one pairwise set of means are not equal.
From the Pairwise Mean Difference (Means Matrix), we conclude that Mean Overall
Satisfaction is significantly different between Customer Type 2 and 3, as well as 1 and 2. Note
that the default probabilities are Fisher Pairwise. The P-Values will be slightly different than the
previous 2 Sample t-Test results because the variances from all 3 customer types are “pooled”
here. This also results in slightly different confidence intervals. See below for more details on
the multiple comparison options.
7. A graphical view of the Overall Satisfaction Mean and 95% Confidence Intervals are given to
complement the Means Matrix. The fact that the CI’s for Customer Type 2 do not overlap those
of Type 1 or 3, clearly shows that Customer Type 2 has a significantly higher mean satisfaction
score. The overlap of CI’s for Type 1 and 3 shows that the mean satisfaction scores for 1 and 3
are not significantly different.
8. Note that the Confidence Level shown in the graph can be modified by clicking the Options
button and setting Confidence Level. This also changes the alpha value used to highlight the P-
Values in red:
239
SigmaXL: Analyze Phase Tools
9. The R-Square (R-Sq) value of 20.95% indicates that Customer Type “explains” approximately
21% of the variation in Overall Satisfaction. We need to “drill down” to understand the root
causes and best practices associated with Customer Type 2.
This is a text report with color highlight: Green (OK), Yellow (Warning) and Red (Serious
Violation).
Each sample is tested for Normality using the Anderson Darling test. If not normal, the
minimum sample size for robustness of the ANOVA Test is determined utilizing Monte Carlo
regression equations (see Basic Statistical Templates – Minimum Sample Size for Robust t-
Tests and ANOVA). If the sample size is inadequate, a warning is given and the appropriate
Nonparametric test is recommended (Kruskal-Wallis if there are no extreme outliers, Mood’s
Median if there are extreme outliers).
Each sample is tested for Outliers defined as: Potential: Tukey's Boxplot (> Q3 + 1.5*IQR or < Q1
– 1.5*IQR); Likely: Tukey's Boxplot 2.2*IQR; Extreme: Tukey's Boxplot 3*IQR. If outliers are
present, a warning is given and recommendation to review the data with a Boxplot and Normal
Probability Plot. Here we have a potential outlier for Customer Type 2.
Tip: If the removal of outlier(s) result in an Anderson Darling P-Value that is > 0.1, a notice is
given that excluding the outlier(s), the sample data are inherently normal.
Each sample is tested for Randomness using the Exact Nonparametric Runs Test. If the sample
data is not random, a warning is given and recommendation to review the data with a Run
Chart.
240
SigmaXL: Analyze Phase Tools
A test for Equal Variances is also applied. If all sample data are normal, Bartlett’s Test is
utilized, otherwise Levene’s Test is used. If the variances are unequal, then a warning is given
and Welch’s ANOVA is recommended.
• Fisher:
o Does not correct for family-wise error rate, so should only be used for k = 3 means
and in the restricted case where the ANOVA P-Value is < alpha (this is also known as
Protected Fisher LSD). For k = 3 means, Protected Fisher LSD is more powerful than
Tukey.
• Tukey:
o Similar to LSD, uses pairwise tests with pooled standard deviation, but is a
studentized range statistic that corrects for family-wise error rate. Recommended
for k > 3.
o If one of the groups are a control reference group, Dunnett with Control is more
powerful than Tukey because it is doing fewer pairwise comparisons (only considers
those pairwise against the control group).
o Uses pooled standard deviation and a multivariate t distribution that corrects for
family-wise error rate.
o This chart is also available at SigmaXL > Graphical Tools > Analysis of Means
(ANOM). See Part P – Analysis of Means (ANOM) Charts.
241
SigmaXL: Analyze Phase Tools
1. Press F3 or click Recall SigmaXL Dialog to recall last dialog. Uncheck Display Test Assumptions
Report. Click the Options button. Select Tukey. Check Display Residual Charts and Display
ANOM Normal One-Way Chart; Confidence Level = 95.0 as shown:
Note: The Confidence Level is used to set the level in the Mean/Confidence Interval Plot, the
alpha level (alpha = (100 – CI)/100) used to highlight the P-Values and the alpha level for the
ANOM chart. However, the confidence level used in the Residuals Normal Probability Plot is
always 95%.
2. Click OK. The Pairwise Means Difference (Means Matrix) and Tukey Probability results are:
242
SigmaXL: Analyze Phase Tools
3. The significant Tukey Probability values above have not changed as compared to Fisher, but
note that they are larger than Fisher to compensate for the family-wise error rate:
4. Press F3 or click Recall SigmaXL Dialog to recall last dialog. Click Options. Uncheck all Display
options and select Dunnett with Control Level = 2 (i.e. we are treating Customer Type 2 as the
Control Group):
243
SigmaXL: Analyze Phase Tools
5. Click OK. The Pairwise Means Difference (Means Matrix) and Dunnett Probability results are:
All three mean differences are shown but only two probability values are calculated: Customer
Type 1 – Control Type 2 and Control Type 2 – Type 3. (Typically one would show 1 – 2 and 3 – 2
for Dunnett, but we are displaying the upper triangle for consistency with the other options).
Note that the Dunnett 2 – 3 P-Value is .003 which is smaller than the Tukey .004 but larger than
the Fisher .002. Dunnett is more powerful than Tukey (due to having fewer tests), but less
powerful than Fisher. It does, however, have an advantage over Fisher because it protects the
family-wise error rate for the comparisons being made.
244
SigmaXL: Analyze Phase Tools
6. Click the ANOVA Residuals sheet tab to display the Residual Plots (created at steps 1, 2):
Residuals are the unexplained variation from the ANOVA model (Actual – Predicted/Fitted
values). We expect to see the residuals approximately normally distributed with no obvious
patterns in the above graphs, which is the case here.
245
SigmaXL: Analyze Phase Tools
7. Click ANOM_Normal_One_Way sheet tab to display the ANOM Chart (created at steps 1, 2):
Here we see that Customer Type 1 mean satisfaction score is significantly below the overall
mean and Customer Type 2 is significantly higher. This is consistent with the results that are
observed in the the Pairwise Means Difference (Means Matrix), but is easier to interpret.
The varying decision limits are due to the varying sample sizes for each Customer Type, with
smaller sample size giving wider limits in a manner similar to a control chart. If the data are
balanced the decision limit lines will be constant.
246
SigmaXL: Analyze Phase Tools
1. Click SigmaXL > Statistical Tools > Power & Sample Size Calculators > One-Way ANOVA
Calculator.
2. Select Solve For Power (1 – Beta). Enter Sample Size and Maximum Difference as shown:
Note that we are calculating the power or likelihood of detection given that the maximum
difference between group means = 1, with sample size for each group = 30, 3 groups, standard
deviation = 1, significance level = .05, and Ha: Not Equal To (two sided test).
A power value of 0.94 is acceptable. Note that this value is less than the power value of 0.97
obtained with the two-sample t-Test.
247
SigmaXL: Analyze Phase Tools
4. Press F3 or click Recall SigmaXL Dialog to recall last dialog. Change the number of groups to 4.
Click OK. Note that the power value is 0.907. If the number of groups (levels) increases, you
will have to increase the sample size in order to maintain statistical power.
5. To determine Power & Sample Size using a Worksheet, click SigmaXL > Statistical Tools >
Power & Sample Size with Worksheet > One-Way ANOVA.
6. A graph showing the relationship between Power, Sample Size and Maximum Difference can
then be created using SigmaXL > Statistical Tools > Power & Sample Size Chart. See Part E for
an example using the 1 Sample t-Test.
248
SigmaXL: Analyze Phase Tools
Two-Way ANOVA
Two-Way ANOVA tests the following:
H0 (Factor X1): μ1 = μ2 = …= μk
Ha (Factor X1): at least one pairwise set of means are not equal (μi ≠ μj);
H0 (Factor X2): μ1 = μ2 = …= μk
Ha (Factor X2): at least one pairwise set of means are not equal (μi ≠ μj);
A Two-Way ANOVA analysis will typically have balanced data from a designed experiment, with an
equal number of observations for each combination level of X1 and X2. SigmaXL will also
accommodate unbalanced data. The minimum requirement is one observation per combination
level of X1 and X2. An error message will be produced if this minimum requirement is not met.
1. Open Customer Data.xlsx, click on Sheet 1 tab (or press F4 to activate last worksheet).
2. Click SigmaXL > Statistical Tools > Two-Way ANOVA. Ensure that the entire data table is
selected. If not, check Use Entire Data Table.
3. Click Next. Select Avg No of Orders per Mo, click Numeric Data Variable (Y) >>; select Customer
Type, click Group Category Factor (X1) >>; select Size of Customer, click Group Category Factor
(X2) >>. Remove Interaction (Fit Additive Model) should remain unchecked.
Tip: If the Two-Way ANOVA report gives a P-Value for the interaction that is high (> 0.1), you
should then press F3 or click Recall SigmaXL Dialog to recall this dialog, check Remove
Interaction (Fit Additive Model) and rerun the analysis.
Display ANOM Normal Two-Way Chart should be unchecked. We will discuss this tool later in
Part P – Analysis of Means (ANOM) Charts. Note that this option is only available when
Remove Interaction (Fit Additive Model) is unchecked because ANOM Normal Two-Way
always includes the interaction term in the model. The ANOM alpha is (100 – Confidence
Level)/100).
249
SigmaXL: Analyze Phase Tools
250
SigmaXL: Analyze Phase Tools
5. Scroll down to view the Main Effects (with confidence intervals) and Interaction Plots:
Note that the mean values shown are fitted (predicted least squares) means not data means.
This compensates for unbalanced data.
6. From the ANOVA table, we can see that the Size of Customer term is significant with a P-Value
less than .05. Customer Type is not significant by itself, but the Interaction term is significant.
This indicates that the effect of Size of Customer on Average Number of Orders per Month
depends on Customer Type (we could also equivalently say that the effect of Customer Type
depends on Size of Customer).
7. This is also confirmed looking at the Main Effects and Interaction plots. Customer Type by itself
is not significant. Size of Customer is obviously significant. Looking at the Interaction plot, the
different slopes illustrate that the change in Average Number of Orders per Month across
Customer Types depends on Customer Size, albeit this is a relatively small effect.
251
SigmaXL: Analyze Phase Tools
25 3.00
2.00
20
1.00
15
NSCORE
Frequency
0.00
10
-1.00
5 -2.00
-3.00
0
00
0
00
00
00
00
0
0
00
00
00
00
00
00
2
.1
.1
.0
.6
.7
.7
.8
.0
5
.0
.0
.0
.0
8.
5.
2.
5.
0.
5.
0.
2.
5.
8.
0.
5.
11
14
17
-9
-6
-3
-0
-5
10
15
20
25
-1
-1
-1
-2
-2
-1
-1
Residuals Residuals
Residuals vs Data Order for: Avg No. of orders per mo Residuals vs Predicted Values for: Avg No. of orders per mo
20.00 20.00
15.00 15.00
10.00 10.00
5.00 5.00
Residuals
Residuals
0.00 0.00
-5.00 -5.00
-10.00 -10.00
-15.00 -15.00
-20.00
-20.00
-25.00
-25.00
0
.0
.0
.0
.0
.0
1
1
11
21
31
41
51
61
71
81
91
10
24
29
34
39
44
Observation Order Fitted Values
16.47
11.47
6.47
Residuals
1.47
-3.54
-8.54
-13.54
-18.54
1
3
2
8
1.
1.
1.
1.
2.
2.
2.
2.
Customer Type
Residuals are the unexplained variation from the ANOVA model (Actual – Predicted/Fitted
values). We expect to see the residuals approximately normally distributed with no obvious
patterns in the above graphs, which is the case here.
The Residuals versus Size of Customer graph is not shown because Size of Customer is text. In
order to display this plot, Size could be coded numerically with 1 = Small and 2 = Large. Simply
create a new column called Size-Coded and use the following Excel formula to create the coded
values for the first record:
=IF(K2="Small",1,IF(K2="Large",2))
Copy and Paste this formula to obtain coded values for all 100 records. Rerun the Two-Way
ANOVA analysis to create the residual graphs
252
SigmaXL: Analyze Phase Tools
Bartlett’s Test
Bartlett’s Test is similar to the 2 sample F-Test (SigmaXL > Statistical Tools > 2 Sample Comparison
Test) but allows for multiple group comparison of variances (or standard deviations). Like the F-
Test, Bartlett’s requires that the data from each group be normally distributed but is more
powerful than Levene’s Test.
• F-Test Pairwise
o Does not correct for family-wise error rate, so should only be used for k = 3 groups
and in the restricted case where the Bartlett P-Value is < alpha.
o This chart is also available at SigmaXL > Graphical Tools > Analysis of Means
(ANOM). See Part P – Analysis of Means (ANOM) Charts.
253
SigmaXL: Analyze Phase Tools
2. Click SigmaXL > Statistical Tools > Equal Variance Tests > Bartlett. Ensure that the entire data
table is selected. If not, check Use Entire Data Table.
3. Click Next. Ensure that Stacked Column Format is checked. Select Delivery Time Deviation,
click Numeric Data Variable (Y) >>; select Floor, click Group Category (X) >>.
5. All 10 Anderson-Darling Test P-Values are > .05 indicating that all group data are normal. Since
the assumption of normality is met, Bartlett’s is the appropriate test to use. If any one of the
groups have a low P-Value for the Normality test, then Levene’s test should be used.
6. With the P-Value = 0.63 we fail to reject H0; we do not have evidence to show that the group
variances are unequal (practically speaking we will assume that the variances are equal).
7. If the Equal Variance test is only being used to test the assumption for use in ANOVA then it is
not necessary to examine the Multiple Comparison of Variances. However, in the context of a
process improvement project, we often do want to know which groups are significantly
different. This can give us important clues to identify opportunities for variance reduction.
254
SigmaXL: Analyze Phase Tools
10. Press F3 or click Recall SigmaXL Dialog to recall last dialog. Click Options. Check Display ANOM
Variances Chart.
Note: The Confidence Level determines the alpha level (alpha = (100 – CI)/100) used to
highlight the P-Values and the alpha level for the ANOM chart. However, the alpha level used
to highlight P-Values in the Anderson-Darling Normality Test is always 0.05.
11. We will not run F-Test with Bonferroni Correction in this example, but typically that would be
used when there are more than 3 groups.
255
SigmaXL: Analyze Phase Tools
12. Click OK. Click ANOM_Variances sheet tab to display the ANOM Chart:
The ANOM Variances chart visually shows that none of the group standard deviations are
significantly different from the grand mean of all the standard deviations. It is called an ANOM
Variances Chart but displays Standard Deviations for ease of interpretation (similar to a
Standard Deviation S Control Chart). This does, however, result in non-symmetrical decision
limits. The ANOM Variances chart in SigmaXL > Graphical Tools > Analysis of Means (ANOM) >
ANOM Variances has an option to display Variances.
256
SigmaXL: Analyze Phase Tools
Levene’s Test
Levene’s Test for multiple group comparison of variances is less powerful that Bartlett’s Test, but is
robust to the assumption of normality. (This is a modification of the original Levene’s Test,
sometimes referred to as the Browne-Forsythe Test that uses Absolute Deviations from the Median
or ADM).
• Levene Pairwise
o Does not correct for family-wise error rate, so should only be used for k = 3 groups
and in the restricted case where the Levene P-Value is < alpha.
o This post-hoc test is unique to SigmaXL, inspired by the method used in ANOM
Levene Variances.
o This chart is also available at SigmaXL > Graphical Tools > Analysis of Means
(ANOM). See Part P – Analysis of Means (ANOM) Charts.
257
SigmaXL: Analyze Phase Tools
2. Click SigmaXL > Statistical Tools > Equal Variance Tests > Levene. Ensure that the entire data
table is selected. If not, check Use Entire Data Table.
3. Click Next. Ensure that Stacked Column Format is checked. Select Responsive to Calls, click
Numeric Data Variable (Y) >>; select Customer Type, click Group Category (X) >>.
5. The Levene’s Test P-Value of 0.0144 tells us that we reject H0. At least one pairwise set of
variances are not equal. The normality test P-Values indicate that all 3 groups have nonnormal
data (P-Values < .05). Since Levene’s Test is robust to the assumption of normality, it is the
correct test for equal variances (rather than Bartlett’s Test).
6. If the Equal Variances test is only being used to test the assumption for use in ANOVA then it is
not necessary to examine the Multiple Comparison of Variances. However, in the context of a
258
SigmaXL: Analyze Phase Tools
process improvement project, we often do want to know which groups are significantly
different. This can give us important clues to identify opportunities for variance reduction.
Press F3 or click Recall SigmaXL Dialog to recall last dialog. Click Options. Select Tukey ADM
(Absolute Deviations from Median). Check Display ANOM Levene Robust Variances Chart.
Note: The Confidence Level determines the alpha level (alpha = (100 – CI)/100) used to
highlight the P-Values and the alpha level for the ANOM chart. However, the alpha level used
to highlight P-Values in the Anderson-Darling Normality Test is always 0.05.
8. Click OK. The Multiple Comparison of Variances is a matrix of Tukey ADM Probabilities:
259
SigmaXL: Analyze Phase Tools
The 1 – 2 P-Value is significant but larger than the Levene Pairwise because it adjusts for the
family-wise error rate. Note that it is smaller than the Bonferroni corrected value = .0036 * 3 =
.011, so more powerful than Bonferroni. The difference in power between Tukey and
Bonferroni becomes more prominent with a larger number of groups, so Bonferroni is not
included as an option.
10. The ANOM chart clearly shows Customer Type 1 has significantly higher variance (ADM) than
overall and Customer Type 2 has significantly lower variance.
11. The varying decision limits are due to the varying sample sizes for each Customer Type, with
smaller sample size giving wider limits in a manner similar to a control chart. If the data are
balanced, the decision limit lines will be constant.
12. Now that we have determined that the variances are not equal, we are presented with a
problem if we want to test for equal group means. Classical ANOVA assumes that the group
variances are equal, so should not be used. A modified ANOVA called Welch’s ANOVA is robust
to the assumption of equal variances and will be demonstrated next.
260
SigmaXL: Analyze Phase Tools
• Welch Pairwise:
o Pairwise 2 sample t-tests with unpooled standard deviation and weighted degrees of
freedom (2 sample t-test for unequal variance)
o Does not correct for family-wise error rate, so should only be used for k = 3 means
and in the restricted case where the Welch ANOVA P-Value is < alpha.
• Games-Howell:
o Similar to Welch Pairwise, uses unpooled standard deviation and weighted degrees
of freedom, but is a studentized range statistic that corrects for family-wise error
rate. Recommended for k > 3.
o It is an extension of the Tukey test, but does not assume equal variance.
• ANOM Chart for Welch ANOVA is not included in SigmaXL since it requires two-stage
sampling (see Nelson, Wludyka, Copeland, 2005, Chapter 8, Heteroscedastic Data)
References for Analysis of Means (ANOM) Charts.
261
SigmaXL: Analyze Phase Tools
1. Open Customer Data.xlsx, click on Sheet 1 tab (or press F4 to activate last worksheet).
2. Click SigmaXL > Statistical Tools > Equal Variance Tests > Welch’s ANOVA (Assume Unequal
Variance) or SigmaXL > Statistical Tools > Welch’s ANOVA (Assume Unequal Variance). Ensure
that the entire data table is selected. If not, check Use Entire Data Table.
3. Click Next. Ensure that Stacked Column Format is checked. Select Responsive to Calls, click
Numeric Data Variable (Y) >>; select Customer Type, click Group Category (X) >>. Check
Display Test Assumptions Report.
262
SigmaXL: Analyze Phase Tools
5. The P-Value for Welch’s ANOVA is 0.0135, therefore we reject H0 and conclude that the group
means for Responsive to Calls are not equal.
6. From the Pairwise Mean Difference (Means Matrix), we conclude that Mean Responsive to Calls
is significantly different between Customer Type 1 and 2. Note that the default probabilities
are Welch Pairwise. See below for more details on the multiple comparison options.
8. A graphical view of the Responsive to Calls Mean and 95% Confidence Intervals are given to
complement the Means Matrix. Note that the standard deviations are unpooled, resulting in
different CI widths for each group. The fact that the CI’s for Customer Type 1 do not overlap
those of Type 2, visually shows that there is a significant difference in mean Responsive to Calls.
The overlap of CI’s for Type 2 and 3 shows that the mean scores for 2 and 3 are not significantly
different.
9. Later, we will explore the relationship between Responsive to Calls and Overall Satisfaction.
263
SigmaXL: Analyze Phase Tools
This is a text report with color highlight: Green (OK), Yellow (Warning) and Red (Serious
Violation).
Each sample is tested for Normality using the Anderson-Darling test. If not normal, the
minimum sample size for robustness of the ANOVA Test is determined utilizing Monte Carlo
regression equations (see Basic Statistical Templates – Minimum Sample Size for Robust t-
Tests and ANOVA). If the sample size is inadequate, a warning is given and the appropriate
Nonparametric test is recommended (Kruskal-Wallis if there are no extreme outliers, Mood’s
Median if there are extreme outliers).
Each sample is tested for Outliers defined as: Potential: Tukey's Boxplot (> Q3 + 1.5*IQR or < Q1
– 1.5*IQR); Likely: Tukey's Boxplot 2.2*IQR; Extreme: Tukey's Boxplot 3*IQR. If outliers are
present, a warning is given and recommendation to review the data with a Boxplot and Normal
Probability Plot
Tip: If the removal of outlier(s) result in an Anderson-Darling P-Value that is > 0.1, a notice is
given that excluding the outlier(s), the sample data are inherently normal.
Each sample is tested for Randomness using the Exact Nonparametric Runs Test. If the sample
data is not random, a warning is given and recommendation to review the data with a Run
Chart.
A test for Equal Variances is also applied. If all sample data are normal, Bartlett’s Test is
utilized, otherwise Levene’s Test is used. Since we are using Welch’s ANOVA, it is confirmed as
appropriate.
264
SigmaXL: Analyze Phase Tools
11. Press F3 or click Recall SigmaXL Dialog to recall last dialog. Uncheck Display Test Assumptions
Report. Click the Options button. Select Games-Howell. Check Display Residual Charts as
shown:
Note: The Confidence Level is used to set the level in the Mean/Confidence Interval Plot and
the alpha level (alpha = (100 – CI)/100) used to highlight the P-Values. However, the confidence
level used in the Residuals Normal Probability Plot is always 95%.
12. Click OK. The Pairwise Means Difference (Means Matrix) and Games-Howell Probability results
are:
265
SigmaXL: Analyze Phase Tools
13. The significant Games-Howell Probability values above have not changed as compared to Welch
Pairwise, but note that they are larger to compensate for the family-wise error rate:
Note also that the 1 – 2 Games-Howell probability is smaller than the Bonferroni corrected
value = .0045 * 3 = .0135, so more powerful than Bonferroni. The difference in power between
Games-Howell and Bonferroni becomes more prominent with a larger number of groups, so
Bonferroni is not included as an option.
266
SigmaXL: Analyze Phase Tools
14. Click the Welch Residuals sheet tab to display the Residual Plots:
Residuals are the unexplained variation from the ANOVA model (Actual – Predicted or Fitted
values). Note that the residuals are not normally distributed - as expected from the
assumptions report - but like the regular ANOVA, Welch’s ANOVA is quite robust to the
assumption of normality. Also, as expected from Levene’s test for equal variances, the
variability for Type 2 is less than Type 1, but Welch’s ANOVA is robust to the assumption of
equal variances, so we can trust that the P-Values are valid.
267
SigmaXL: Analyze Phase Tools
Kruskal-Wallis
The Kruskal-Wallis test is an extension of the Mann-Whitney Rank test, allowing for more than
2 samples. It is a nonparametric equivalent to the parametric One-Way ANOVA. The Null
Hypothesis is: H0: Median1 = Median2 = … = MedianK. Ha: At least two Medians are different.
The test does not assume sample normality but does assume that the samples have equal
shapes. If the shapes are different, the null hypothesis is that the distributions are the same.
1. Open Customer Data.xlsx, click on Sheet 1 tab (or press F4 to activate last worksheet).
2. Click SigmaXL > Statistical Tools > Nonparametric Tests > Kruskal-Wallis. Ensure that the
entire data table is selected. If not, check Use Entire Data Table.
3. Click Next. Ensure that Stacked Column Format is checked. Select Overall Satisfaction, click
Numeric Data Variable (Y) >>; select Customer Type, click Group Category (X) >>.
268
SigmaXL: Analyze Phase Tools
Click on cell B16 to view the P-Value with more decimal place precision (or change the cell format
to scientific notation). The P-Value of 0.000023 (2.3 e-5) tells us that we reject H0. At least one
pairwise set of medians are not equal.
5. The Kruskal-Wallis Statistic is based on comparing mean ranks for each group versus the mean
rank for all observations. The Z value for Customer Type 3 is −1.56, the smallest absolute Z-
value. This size indicates that the mean rank for Type 3 differed least from the mean rank for
all observations. The Z value for Customer Type 2 is 4.53, the largest absolute Z-value. This size
indicates that the mean rank differed most from the mean rank for all observations.
269
SigmaXL: Analyze Phase Tools
6. A graphical view of the Overall Satisfaction Median and 95% Confidence Intervals are given to
complement the Z scores. The fact that the CI’s for Customer Type 2 do not overlap those of
Type 1 or 3, clearly shows that Customer Type 2 has a significantly higher median satisfaction
score. The overlap of CI’s for Type 1 and 3 shows that the median satisfaction scores for 1 and
3 are not significantly different.
270
SigmaXL: Analyze Phase Tools
Kruskal-Wallis – Exact
We will now redo the above example to estimate the exact P-Value using Monte Carlo. Typically
this would not be necessary unless the sample sizes were smaller (each sample N <= 5 for Kruskal-
Wallis), but this gives us continuity on the example. We will consider a small sample problem later.
Computing an exact P-Value for Kruskal-Wallis is very computationally intensive. The Network
Model by Mehta and Patel cannot be used for this test (see Appendix Exact and Monte Carlo P-
Values for Nonparametric and Contingency Tests). In this example, the total number of
permutations are:
(i.e., more than the number of stars in the observable universe). So, we will not attempt to
compute the exact, but rather use Monte Carlo!
1. Open Customer Data.xlsx, click Sheet 1 tab (or press F4 to activate last worksheet).
2. Click SigmaXL > Statistical Tools > Nonparametric Tests – Exact > Kuskal-Wallis – Exact. If
necessary, check Use Entire Data Table, click Next.
3. Ensure that Stacked Column Format is checked. Select Overall Satisfaction, click Numeric Data
Variable (Y) >>; select Customer Type, click Group Category (X) >>. Select Monte Carlo Exact
with the Number of Replications = 1e6 and Confidence Level for P-Value = 99%. One million
replications are used because the expected P-Value is very small as estimated from the “large
sample” Kruskal-Wallis above. This will take up to a minute to run, so if you have a slow
computer, use 1e5 replications instead of 1e6.
271
SigmaXL: Analyze Phase Tools
Tip: The Monte Carlo 99% confidence interval for P-Value is not the same as a confidence
interval on the test statistic due to data sampling error. The confidence level for the hypothesis
test statistic is still 95%, so all reported P-Values less than .05 will be highlighted in red to
indicate significance. The 99% Monte Carlo P-Value confidence interval is due to the
uncertainty in Monte Carlo sampling, and it becomes smaller as the number of replications
increases (irrespective of the data sample size). The Exact P-Value will lie within the stated
Monte Carlo confidence interval 99% of the time.
4. Click OK.
Click on cell B16 to view the P-Value with more decimal place precision (or change the cell
format to scientific notation). The Monte Carlo P-Value here is 0.000009 (9 e-6) with a 99%
confidence interval of .000002 (2 e-6) to 0.000016 (1.6 e-5). This will be slightly different every
time it is run (the Monte Carlo seed value is derived from the system clock). So we reject H0: at
least one pairwise set of medians are not equal.
Note that the large sample (asymptotic) P-Value of 2.3 e-5 lies outside of the Monte Carlo exact
confidence interval.
5. Now we will consider a small sample problem. Open Snore Study.xlsx. This data is from:
Gibbons, J.D. and Chakraborti, S. (2010). Nonparametric Statistical Inference (5th Edition). New
York: Chapman & Hall, (Example 10.2.1 data, page 347; Example 10.4.2 analysis, pp. 360 – 362).
An experiment was conducted to determine which device is the most effective in stopping
snoring or at least in reducing it. Fifteen men who are habitual snorers were divided randomly
into three groups to test the devices. Each man’s sleep was monitored for one night by a
machine that measures amount of snoring on a 100-point scale while using a device.
272
SigmaXL: Analyze Phase Tools
6. Select Snore Study Data tab. Click SigmaXL > Statistical Tools > Nonparametric Tests – Exact >
Kruskal-Wallis - Exact. If necessary, check Use Entire Data Table, click Next.
7. With Unstacked Column Format checked, select Device A, Device B and Device C, click Numeric
Data Variables (Y) >>. Select Exact with the default Time Limit for Exact Computation = 60
seconds.
With the Exact P-Value = 0.0042 we reject H0, and conclude that there is a significant difference
in median snore study scores. This exact P-Value matches that given in the reference textbook
using SAS and StatXact.
273
SigmaXL: Analyze Phase Tools
By way of comparison we will now rerun the analysis using the “large sample” or “asymptotic”
Kruskal-Wallis test.
9. Select Snore Study Data tab (or press F4 to activate last worksheet). Click SigmaXL > Statistical
Tools > Nonparametric Tests > Kruskal-Wallis. If necessary, check Use Entire Data Table, click
Next.
10. With Unstacked Column Format checked, select Device A, Device B and Device C, click Numeric
Data Variables (Y) >>.
With the P-Value = .0118 we reject H0 (using alpha = .05), but note that if we were using alpha
= 0.01, we would have incorrectly failed to reject the null hypothesis. This “large sample” P-
Value matches that given in the reference textbook using Minitab.
In conclusion, whenever you have a small sample size and are performing a nonparametric test,
always use the Exact option.
274
SigmaXL: Analyze Phase Tools
Mood's median test is more robust to outliers than the Kruskal-Wallis test, but is less powerful in
the absence of outliers. You should first look at your data with Boxplots. If there are extreme
outliers, then Mood’s Median should be used rather than Kruskal-Wallis.
1. Open Customer Data.xlsx, click on Sheet 1 tab (or press F4 to activate last worksheet).
2. Click SigmaXL > Statistical Tools > Nonparametric Tests > Mood’s Median Test. Ensure that
the entire data table is selected. If not, check Use Entire Data Table.
3. Click Next. Ensure that Stacked Column Format is checked. Select Overall Satisfaction, click
Numeric Data Variable (Y) >>; select Customer Type, click Group Category (X) >>.
275
SigmaXL: Analyze Phase Tools
The P-Value of 0.0012 tells us that we reject H0. At least one pairwise set of medians are not
equal.
5. A graphical view of the Overall Satisfaction Median and 95% Confidence Intervals are given.
This is the same graph provided in the Kruskal-Wallis test report.
276
SigmaXL: Analyze Phase Tools
1. Open Customer Data.xlsx, click Sheet 1 tab (or press F4 to activate last worksheet).
2. Click SigmaXL > Statistical Tools > Nonparametric Tests – Exact > Mood’s Median Test – Exact.
If necessary, check Use Entire Data Table, click Next.
3. With Stacked Column Format checked, select Overall Satisfaction, click Numeric Data Variable
(Y) >>; select Customer Type, click Group Category (X) >>. Select Exact with the default Time
Limit for Exact Computation = 60 seconds. Set Values Equal to Overall Median: to Counted as
Below.
Tip: If none of the observations are equal to the overall median, the options for “Values Equal
to Overall Median:” do not affect the exact P-Value. That is the case with this Customer Data
example. In cases where there are observations equal to the overall median, then the exact P-
Value will be different for each option selected. For regular Mood’s Median, SigmaXL uses
Counted as Below (in agreement with Minitab) but will use Counted as Above for small sample
cases if the average cell expected value is higher. StatXact uses Counted as Below for Mood’s
Median. Matlab uses Not Counted. Users should try Counted as Below and Counted as Above
to ensure that the P-Values agree on reject or fail to reject H0. Use of this option will be
demonstrated in the small sample example.
277
SigmaXL: Analyze Phase Tools
Tip: If the exact computation time limit is exceeded a dialog will prompt you to use Monte Carlo
or to increase the computation time. When this occurs, Monte Carlo is recommended.
Click on cell B17 to view the Exact P-Value with more decimal place precision Given the P-Value
of .00124 we reject H0 and conclude that at least one pairwise set of medians are not equal.
The Exact P-Value is very close to the above “large sample” or “asymptotic” result (.00121). This
was expected because the sample size is reasonable (N > 10), so the “large sample” Mood’s
Median P-Values are valid using a chi-square approximation. Note also that the counts in B8:D9
form a two way contingency table and the expected counts are all greater than 5.
The Exact P-Value was computed in seconds, but if the data set was larger, the required
computation time could become excessive, and Monte Carlo would be required.
5. Press F3 or click Recall SigmaXL Dialog to recall last dialog. Select Monte Carlo Exact with the
Number of Replications = 1e6 and Confidence Level for P-Value = 99%. One million
replications are used because the expected P-Value is small as estimated from the “large
sample” Mood’s Median above. This will take up to a minute to run, so if you have a slow
computer, use 1e5 replications instead of 1e6.
278
SigmaXL: Analyze Phase Tools
Tip: The Monte Carlo 99% confidence interval for P-Value is not the same as a confidence
interval on the test statistic due to data sampling error. The confidence level for the hypothesis
test statistic is still 95%, so all reported P-Values less than .05 will be highlighted in red to
indicate significance. The 99% Monte Carlo P-Value confidence interval is due to the
uncertainty in Monte Carlo sampling, and it becomes smaller as the number of replications
increases (irrespective of the data sample size). The Exact P-Value will lie within the stated
Monte Carlo confidence interval 99% of the time.
6. Click OK.
279
SigmaXL: Analyze Phase Tools
Click on cell B17 to view the Monte Carlo P-Value with more decimal place precision. The
Monte Carlo P-Value here is .001249 with a 99% confidence interval of .00117 (B19) to .00133
(B18). This will be slightly different every time it is run (the Monte Carlo seed value is derived
from the system clock). The true Exact P-Value = .00124 lies within this confidence interval.
So we reject H0: at least one pairwise set of medians are not equal.
7. Now we will consider the small sample problem used in Kruskal-Wallis. Open Snore Study.xlsx.
This data is from:
Gibbons, J.D. and Chakraborti, S. (2010). Nonparametric Statistical Inference (5th Edition). New
York: Chapman & Hall, (Example 10.2.1, pp. 347-348).
An experiment was conducted to determine which device is the most effective in stopping
snoring or at least in reducing it. Fifteen men who are habitual snorers were divided randomly
into three groups to test the devices. Each man’s sleep was monitored for one night by a
machine that measures amount of snoring on a 100-point scale while using a device.
8. Select Snore Study Data tab. Click SigmaXL > Statistical Tools > Nonparametric Tests – Exact >
Mood’s Median Test - Exact. If necessary, check Use Entire Data Table, click Next.
9. With Unstacked Column Format checked, select Device A, Device B and Device C, click Numeric
Data Variables (Y) >>. Select Exact with the default Time Limit for Exact Computation = 60
seconds.
280
SigmaXL: Analyze Phase Tools
With the Exact P-Value = .0676 we fail to reject H0, and cannot conclude that there is a
significant difference in median snore study scores. This exact P-Value matches that given in
the reference textbook using StatXact. Comparing Mood’s Median to Kruskal-Wallis we see
that the Exact P-Value is higher because Mood’s Median is not as powerful as Kruskal-Wallis.
We will now rerun the analysis using the Counted as Above option.
11. Press F3 or click Recall SigmaXL Dialog to recall last dialog. Set Values Equal to Overall
Median: to Counted as Above.
281
SigmaXL: Analyze Phase Tools
With the Exact P-Value = 0.3007 we fail to reject H0, and cannot conclude that there is a
significant difference in median snore study scores. Note that setting Values Equal to Overall
Median to Counted as Above resulted in a different count for Device C, which results in the
dramatic difference in Exact P-Value.
In conclusion, when using Mood’s Median Exact, always try Counted as Below and Counted as
Above to ensure that the P-Values agree with each other.
282
SigmaXL: Analyze Phase Tools
1. Open Customer Data.xlsx, click on Sheet 1 tab (or press F4 to activate last worksheet).
2. Click SigmaXL > Statistical Tools > Nonparametric Tests > Runs Test. Ensure that the entire
data table is selected. If not, check Use Entire Data Table.
3. Click Next. Select Overall Satisfaction, click Numeric Data Variable (Y) >>:
283
SigmaXL: Analyze Phase Tools
5. With all of the P-Values being greater than 0.01, we fail to reject H0, and conclude that the data
is random (or statistically independent). Recall from the run chart of this data that there were
no obvious trends or patterns.
a. Clustering appears as a group of points in one area of the chart. It may indicate
special cause variation such as sampling or measurement problems.
284
SigmaXL: Analyze Phase Tools
b. Mixtures appear as an absence of data points near the center line. A mixture may
indicate a bimodal distribution due to a regular change of shift, machinery, or raw
materials.
Note that the P-Value for Mixtures = 1 – P-Value for Clustering. They are mutually
exclusive. The P-Value for Lack of Randomness = 2 * minimum of (P-Value
Clustering, P-Value Mixtures).
c. Trends appear as an upward or downward drift in the data and may be due to
special causes such as tool wear.
285
SigmaXL: Analyze Phase Tools
Note that the P-Value for Trends = 1 – P-Value for Oscillation. They are mutually
exclusive.
286
SigmaXL: Analyze Phase Tools
It is important to note that while exact P-Values are “correct,” they do not increase (or decrease)
the power of a small sample test, so they are not a solution to the problem of failure to detect a
trend due to inadequate sample size.
If Count (N) is greater than 1000, the exact P-Value is estimated using a continuity-corrected
normal approximation. Since the Runs Test Exact P-Value is computed very quickly for sample sizes
as large as 1000, Monte Carlo P-Values are not required.
Exact P-Values are derived from published tables, given the sample size and the number of
up/down runs. The exact tables apply to N <= 50. If N > 50, a continuity-corrected normal
approximation is used.
For further details and references refer to the Appendix Exact and Monte Carlo P-Values for
Nonparametric and Contingency Tests.
We will now redo the Customer Data Overall Satisfaction example to compute exact P-Values.
Typically this would not be necessary unless the sample sizes were smaller (N <= 50), but this gives
us continuity on the example. We will also consider the above small sample examples later.
2. Click SigmaXL > Statistical Tools > Nonparametric Tests – Exact > Runs Test – Exact. If
necessary, check Use Entire Data Table, click Next.
3. Select Overall Satisfaction, click Numeric Data Variables (Y) >>. Set Values Equal to Median: to
Counted as Below.
287
SigmaXL: Analyze Phase Tools
Note on Options for Clustering, Mixtures and Lack of Randomness (Runs Above/Below)
If none of the observations are equal to the sample median, the options for “Values Equal to
Median:” do not affect the exact P-Value. That is the case with this Customer Data example. In
cases where there are observations equal to the sample median, then the exact P-Value will be
different for each option selected. For regular Runs Test, SigmaXL uses Counted as Below (in
agreement with Minitab). StatXact uses Counted as Above for the Runs Test. Matlab uses Not
Counted. Users should try Counted as Below and Counted as Above to ensure that the P-
Values agree on reject or fail to reject H0.
The Values Equal to Median options also apply to the Runs Up/Down Test: zeros of the first
order differences (i.e., two consecutive values that are the same). Counted as Below denotes
“counted as negative.” Counted as Above denotes “counted as positive.” Not Counted
denotes that zero first order differences are deleted. This consolidation of options was done to
keep the dialog simple.
288
SigmaXL: Analyze Phase Tools
With all of the P-Values being greater than 0.01 (alpha = .01 is preferred for the Runs Test to
minimize false alarms), we fail to reject H0, and conclude that the data is random (or
statistically independent).
The Above/Below and Up/Down Runs counts are identical to the above “large sample” or
“asymptotic” results. The Clustering/Mixtures/Randomness (Above/Below) Exact P-Values are
close but slightly different. This was expected because the sample size is reasonable (N > 50), so
the “large sample” P-Values are valid using a normal approximation.
The P-Values for Trends and Oscillation (Up/Down) use a normal approximation because the
sample size is greater than 50. They are, however, slightly different than the “large sample”
above because a continuity correction is now applied to the normal approximation.
5. Now we will consider the small sample examples. Open Runs Test Example Data.xlsx, click
Runs Test Example Data tab.
6. Click SigmaXL > Statistical Tools > Nonparametric Tests – Exact > Runs Test – Exact. If
necessary, check Use Entire Data Table, click Next.
7. Select Clustering and shift click to Oscillation, click Numeric Data Variables (Y) >>. Set Values
Equal to Median: to Counted as Below.
289
SigmaXL: Analyze Phase Tools
The Exact P-Values are close to the above “large sample” or “asymptotic” results, but note that
some of the values are now greater than .01 so they would fail-to reject H0. Note that this does
not imply that the large sample runs test is more powerful, but rather we cannot conclude
some of the previously identified patterns at the 99% confidence level.
We will now rerun the analysis using the Counted as Above option.
9. Press F3 or click Recall SigmaXL Dialog to recall last dialog. Set Values Equal to Median: to
Counted as Above.
290
SigmaXL: Analyze Phase Tools
Note the change in Trends and Oscillation Number of Runs and resulting dramatic change to
the Exact P-Values.
In conclusion, when using Runs Test Exact, always try Counted As Below and Counted as Above
to ensure that the P-Values agree with each other. Also, whenever you have a small sample size
and are performing a nonparametric test, always use Exact.
291
SigmaXL: Analyze Phase Tools
See Basic Statistical Templates – 1 Proportion Test and Confidence Interval for an example of the
1 Proportion Test and Confidence Interval template.
See below for an example of the 2 Proportions Test and Confidence Interval template.
See Basic Statistical Templates – 2 Proportions Equivalence Test for an example of the 2
Proportions Equivalence Test template.
See Basic Statistical Templates – One-Way Chi-Square Goodness-of-Fit Test for an example of the
One-Way Chi-Square Goodness-of-Fit Test template.
292
SigmaXL: Analyze Phase Tools
See Basic Statistical Templates – One-Way Chi-Square Goodness-of-Fit Test - Exact for an example
of the One-Way Chi-Square Goodness-of-Fit Test - Exact template.
See Basic Statistical Templates – 1 Poisson Rate Test and Confidence Interval for an example of
the 1 Poisson Rate Test and Confidence Interval template.
See Basic Statistical Templates – 2 Poisson Rates Test and Confidence Interval for an example of
the 2 Poisson Rates Test and Confidence Interval template.
See Basic Statistical Templates – 2 Poisson Rates Equivalence Test for an example of the 2 Poisson
Equivalence Test template.
293
SigmaXL: Analyze Phase Tools
2. Click SigmaXL > Templates & Calculators > Basic Statistical Templates > 2 Proportions Test and
Confidence Interval (or SigmaXL > Statistical Tools > Basic Statistical Templates > 2
Proportions Test and Confidence Interval).
Since the Fisher’s exact P-Value of 0.141 is greater than .05, we fail to reject H0. We do not
have enough evidence to show that there is a significant difference between Day Shift and
Night shift. This does not mean that we have proven that they are the same. In practice
however, we either assume that they are the same or we collect more data.
Note: Fisher’s exact P-Values should be used for any real world problem. The approximate P-
Values based on the normal distribution are provided for instructional purposes, e.g.,
comparing to hand calculations.
4. Now enter x1 = 160, n1 = 200, x2 = 140, n2 = 200. Note that the Fisher’s exact P-Value is now
.028, so we reject H0.
294
SigmaXL: Analyze Phase Tools
2. Click SigmaXL > Statistical Tools > Chi-Square Tests > Chi-Square Test & Association – Two-
Way Table Data. Note the selection of data includes the Row and Column labels (if we had Row
and Column Totals these would NOT be selected). Do not check Advanced Tests and Measures
of Association.
See Appendix Chi-Square Tests and (Contingency) Table Associations for the Chi-Square Table
Statistics formula details.
295
SigmaXL: Analyze Phase Tools
The P-Value matches that of the 2 Proportions test. Since the P-Value of 0.1 is greater than .05,
we fail to reject H0.
4. Now click Example 2 Sheet tab. The Yields have not changed but we have doubled the sample
size. Repeat the above analysis. The resulting output is:
Since the P-Value is < .05, we now reject the Null Hypothesis, and conclude that Day Shift and
Night Shift are significantly different. The Std. (Standardized) Residuals tell us that Day Shift
failures are less than expected (assuming equal proportions), and Night Shift failures are more
than expected.
Note, by doubling the sample size, we improved the power or sensitivity of the test.
5. Click the Example 3 Sheet tab. In this scenario we have 3 suppliers and an additional marginal
level. A random sample of 100 units per supplier is tested. The null hypothesis here is: No
relationship between Suppliers and Pass/Fail/Marginal rates, but in this case, we can state it as
No difference across suppliers. Redoing the above analysis (for selection B2:E5) yields the
following:
296
SigmaXL: Analyze Phase Tools
The P-Value tells us that we do not have enough evidence to show that there is a difference
across the 3 suppliers.
6. Click the Example 4 Sheet tab. Here we have doubled the sample size to 200 per supplier. Note
that the percentages are identical to example 3. Redoing the above analysis yields the
following:
With the P-Value < .05 we now conclude that there is a significant difference across suppliers.
Examining the Std. (Standardized) Residuals tells us that Supplier A has fewer failures than
297
SigmaXL: Analyze Phase Tools
expected (if there was no difference across suppliers), Supplier B has more marginal parts than
expected and Supplier C has fewer marginal parts than expected.
7. The table row and column cell percentages can be visualized using Excel’s 100% Stacked
Column Chart. Select cells A3:D6 of the Chi-Square sheet. Click Excel’s Insert > Insert Column
or Bar Chart and select 100% Stacked Column as shown.
8. Click to create the 100% stacked column chart. Uncheck the Chart Title as shown.
9. The rows and columns can easily be switched by clicking Design > Switch Row/Column
298
SigmaXL: Analyze Phase Tools
10. These charts make it easy to visualize the cell row and column percentages.
299
SigmaXL: Analyze Phase Tools
300
SigmaXL: Analyze Phase Tools
o Cramer’s V
Cramer's V is an extension of Phi for larger tables. It is the most popular
measure of association for tables of any size.
It varies from 0 to 1, with 0 = no association and 1 = perfect association.
Use Cohen’s rules-of-thumb given above for Phi.
o Contingency Coefficient
An alternative to Phi, varies from 0 to < 1.
Use Cohen’s rules-of-thumb given above for Phi.
o Cohen's Kappa (Agreement - Square Table)
Kappa is used to measure agreement between two assessors evaluating the
same parts or items.
For an extended Attribute Measurement Systems Analysis use SigmaXL >
Measurement Systems Analysis > Attribute MSA.
• http://course1.winona.edu/bdeppa/STAT%20701%20Online/stat_701%20home.htm
Click 4a) Measures of Association. This is a narrated presentation, but a regular PowerPoint
301
SigmaXL: Analyze Phase Tools
is also available on the web site. It is part of a Biostatistics course by Dr. Brant Deppa,
Department of Mathematics and Statistics, Winona State University.
1. Press F3 or click Recall SigmaXL Dialog to recall last dialog. Check Nominal Categories as
shown:
302
SigmaXL: Analyze Phase Tools
The adjusted residuals are equivalent to normal z values, so for a specified 95% confidence
level, any value greater than 1.96 (or less than -1.96) is highlighted in red. This results in a slight
difference in interpretation from that of the standardized residuals, but the 3 largest magnitude
residuals are consistent.
As noted above, the Chi-Square P-Value tells us that there is a significant difference across
suppliers, in other words, there is association between Supplier and Pass/Fail/Marginal, but it
does not tell us the degree or strength of that association. Cramer’s V is used for tables larger
than 2x2 and from the rules-of-thumb, the 0.1 value is considered weak (or small effect).
303
SigmaXL: Analyze Phase Tools
Note, coding Shift as continuous numeric is easier to interpret and is valid because it has a range of
1.
Analyze using SigmaXL > Statistical Tools > Regression > Binary Logistic Regression. Select
Response Count (Y) / Sample Size. Pass(Y) is selected as Numeric Response Count (Y), Trials (N) is
Numeric Sample Size (Trials) and Shift (X) is selected as Continuous Predictors (X):
Click OK. The Odds Ratio and Confidence Limits are given as:
The shift change 1 to 2 is 0.583 times as likely to produce passed (good) product with a 95%
confidence interval of 0.368 to 0.924.
304
SigmaXL: Analyze Phase Tools
The P-Value for this hypothesis test is from Kendall’s Tau-B, but is the same
for all of the Concordant – Discordant ordinal measures: Tau-C, Gamma and
Somers’ D. See Agresti (2010). This may differ from other software using an
approximation formula.
o Spearman Rank Correlation
• Measures of Association for Ordinal Categories with Confidence Intervals
o Spearman Rank Correlation
Equivalent to Pearson’s correlation on ranks
o Kendall's Tau-B (Square Table), Kendall-Stuart Tau-C (Rectangular Table), Goodman-
Kruskal Gamma
Use Tau-B for square tables (no. rows = no. columns) and Tau-C for
rectangular tables (no. rows <> no. columns).
Tau-B, Tau-C and Gamma are asymmetric measures, so will give the same
result regardless of variable assignment to Rows and Columns.
305
SigmaXL: Analyze Phase Tools
306
SigmaXL: Analyze Phase Tools
1. Open the file Attribute Data.xlsx, click Example 5 – Salary Sat Sheet tab. This data is in two-
way table format and has ordinal categories: Salary in the Rows and Satisfaction Level in the
Columns. Note that cells A1:E5 have been pre-selected.
2. Click SigmaXL > Statistical Tools > Chi-Square Tests > Chi-Square Test & Association – Two-
Way Table Data. Note the selection of data includes the Row and Column labels (if we had Row
and Column Totals these would NOT be selected). Check Nominal Categories and Ordinal
Categories as shown:
Tip: Even if the categories are ordinal, it is sometimes useful to select nominal categories as
well for comparison purposes.
307
SigmaXL: Analyze Phase Tools
Note that the Chi-Square P-Value is 0.1, indicating that there is no significant association
between Salary and Satisfaction when they are treated as nominal categories (although the
significant result for McNemar-Bowker does show that there is lack of symmetry in the off
diagonals).
Since the Chi-Square P-Value is not significant, the Adjusted Residuals are not highlighted, even
though some values are greater than 1.96 (and less than -1.96). This follows the concept used
in ANOVA called “Fisher Protected” where one considers the significance of post-hoc tests only
when the overall test is significant.
Note: 4 out of 16 cells have expected counts less than 5. If more than 20% of the cells have
expected counts less than 5 (or if any of the cells have an expected count less than 1), the Chi-
Square approximation may be invalid, and Fisher’s Exact should be used. This will be discussed
308
SigmaXL: Analyze Phase Tools
later, but for this example the Fisher’s Monte-Carlo Exact P-Value = 0.095 so does not change
the interpretation of the results for the above Chi-Square analysis.
When Salary and Satisfaction are treated as ordinal categories, the more powerful Concordant
– Discordant and Spearman Rank Correlation P-Values clearly show that there is a significant
association. The Measures of Association for Ordinal Categories show that this is positive, i.e.,
an increase in Salary is associated with an increase in Satisfaction. However, using the rules-of-
thumb given above, we see that the association is weak.
4. The table row and column cell percentages can be visualized using Excel’s 100% Stacked
Column Chart. Select cells A3:E7 of the Chi-Square sheet. Click Excel’s Insert > Insert Column
or Bar Chart and select 100% Stacked Column as shown.
5. Click to create the 100% stacked column chart (uncheck the Chart Title):
6. The rows and columns can easily be switched by clicking Design > Switch Row/Column
309
SigmaXL: Analyze Phase Tools
2. Click SigmaXL > Statistical Tools > Chi-Square Tests > Chi-Square Test & Association. Ensure
that the entire data table is selected. If not, check Use Entire Data Table. Click Next.
3. Select Major-Complaint, click Rows Category >>; select Customer Type, click Columns Category
>>. We will not be using Options in this example.
310
SigmaXL: Analyze Phase Tools
With the P-Value = 0.142 we fail to reject H0, so we do not have enough evidence to show a
difference in customer complaints across customer types.
Note: 9 out of 15 cells have expected counts less than 5. If more than 20% of the cells have
expected counts less than 5 (or if any of the cells have an expected count less than 1), the Chi-
Square approximation may be invalid. Use Chi-Square Test – Fisher’s Exact).
Tip: Use Advanced Pareto Analysis and Excel’s 100% Stacked Column Chart to complement Chi-
Square Analysis.
Caution: When using stacked column format data with Ordinal Category variables that are text,
SigmaXL will sort alphanumerically which may not result in the correct ascending order for
analysis. We recommend coding text Ordinal variables as numeric (e.g., 1,2,3) or modified text
(e.g., Sat_0, Sat_1).
311
SigmaXL: Analyze Phase Tools
5. Press F3 or click Recall SigmaXL Dialog to recall last dialog. Select Loyalty – Likely to
Recommend, click Rows Category >>; select Sat-Discrete, click Columns Category >>. Click
Options, check Ordinal Categories.
Sat-Discrete is derived from Overall Satisfaction where a score >= 3.5 is considered a 1, and
scores < 3.5 are considered a 0.
312
SigmaXL: Analyze Phase Tools
As expected there is a strong positive association between Loyalty and Discrete Satisfaction.
Note: Kendall-Stuart Tau-C should be used here rather than Tau-B because it is a rectangular
table. Since Satisfaction leads to Loyalty, Loyalty is the Dependent variable, so Rows
Dependent Somers’ D should be used rather than Cols Dependent or Symmetric.
7. In order to visualize the row and column percentages with Excel’s 100% Stacked Column Chart,
we will need to modify the numeric row and column labels as shown, converting to text as
shown:
8. Select cells A3:C8 of the Chi-Square sheet. Click Excel’s Insert > Insert Column or Bar Chart and
select 100% Stacked Column as shown.
313
SigmaXL: Analyze Phase Tools
9. Click to create the 100% stacked column chart (uncheck the Chart Title):
This clearly shows the strong relationship between the Loyalty scores and Discrete Satisfaction.
314
SigmaXL: Analyze Phase Tools
It is important to note that while exact P-Values are “correct,” they do not increase (or
decrease) the power of a small sample test, so they are not a solution to the problem of failure
to detect a change due to inadequate sample size!
1. We will now re-analyze the above Major Complaint by Customer Type data. Open Customer
Data.xlsx. Click Sheet 1 tab. The discrete data of interest is Complaints and Customer Type,
i.e., does the type of complaint differ across customer type? Formally the Null Hypothesis is
that there is no relationship (or independence) between Customer Type and Complaints.
2. Click SigmaXL > Statistical Tools > Chi-Square Tests – Exact > Chi-Square Test – Fisher’s Exact.
Ensure that the entire data table is selected. If not, check Use Entire Data Table. Click Next.
3. Select Major-Complaint, click Rows Category >>; select Customer Type, click Column Category
>>. Select Exact with the default Time Limit for Exact Computation = 60 seconds.
Tip: If the exact computation time limit is exceeded a dialog will prompt you to use Monte Carlo
or to increase the computation time. When this occurs, Monte Carlo is recommended.
315
SigmaXL: Analyze Phase Tools
With Fisher’s Exact P-Value = 0.1469 we fail to reject H0, so we cannot conclude that there is a
difference in customer complaints across customer types. This is close to the approximate Chi-
Square P-Value of 0.142, so either P-Value results in the same conclusion.
However, there are cases where the “large sample” Chi-Square P-Value leads to one conclusion
but Fisher’s Exact P-Value gives another. This will be demonstrated later using Two Way Table
Data – Fisher’s Exact.
The Exact P-Value was computed very quickly, but if the data set was larger, the required
computation time could become excessive, and Monte Carlo would be required. We will rerun
this analysis with Monte Carlo and discuss the output report.
5. Press F3 or click Recall SigmaXL Dialog to recall last dialog. Select Monte Carlo Exact with the
default Number of Replications = 10000 and Confidence Level for P-Value = 99%.
316
SigmaXL: Analyze Phase Tools
Tip: As discussed above 10,000 replications will result in a Monte Carlo P-Value that is correct
to two decimal places. One million (1e6) replications will result in three decimal places of
accuracy and typically require less than 60 seconds to solve for any data set.
Tip: The Monte Carlo 99% confidence interval for P-Value is not the same as a confidence
interval on the test statistic due to data sampling error. The confidence level for the hypothesis
test statistic is still 95%, so all reported P-Values less than .05 will be highlighted in red to
indicate significance. The 99% Monte Carlo P-Value confidence interval is due to the
uncertainty in Monte Carlo sampling, and it becomes smaller as the number of replications
increases (irrespective of the data sample size). The Exact P-Value will lie within the stated
Monte Carlo confidence interval 99% of the time.
317
SigmaXL: Analyze Phase Tools
Fisher’s Monte Carlo P-Value here is 0.1486 with a 99% confidence interval of 0.1403 to 0.1569.
This will be slightly different every time it is run (the Monte Carlo seed value is derived from the
system clock). The true Exact P-Value = 0.1469 lies within this confidence interval. If the
confidence interval is too wide (e.g., upper “fail-to-reject” H0 and lower “reject” H0), simply
rerun the Monte Carlo option using a larger number of replications (use 1e5 or 1e6).
318
SigmaXL: Analyze Phase Tools
1. Open the file Oral_Lesions.xlsx. This data is in Two –Way Contingency Table (or pivot table)
format. Note that cells A1:D10 have been pre-selected.
2. Click SigmaXL > Statistical Tools > Chi-Square Tests – Exact > Chi-Square Test – Two-Way Table
Data – Fisher’s Exact. Note the selection of data includes the Row and Column labels.
319
SigmaXL: Analyze Phase Tools
Note: 25 out of 27 cells have expected counts less than 5. Since more than 20% of the cells
have expected counts less than 5 and several cells have an expected count less than 1, the Chi-
Square P-Value = 0.14 is invalid and leads to an incorrect conclusion: Fail to Reject H0, Site of
Oral Lesion and Geographic Region are independent. However with Fisher’s Exact P-Value =
320
SigmaXL: Analyze Phase Tools
0.0101, we have strong evidence to reject H0, and conclude that there is indeed a relationship
between Site of Oral Lesion and Region.
While this example is originally from a dental health study, it could have been Defect Type
versus Supplier and highlights that the use of Fisher’s Exact for contingency tables with sparse
data can make the difference between a good business decision and a bad business decision!
The Exact P-Value was computed very quickly, but if the data set was larger, the required
computation time could become excessive, and Monte Carlo would be required. We will rerun
this analysis with Monte Carlo and discuss the output report.
4. Press F3 or click Recall SigmaXL Dialog to recall last dialog. Select Monte Carlo Exact with the
default Number of Replications = 10000 and Confidence Level for P-Value = 99%.
Tip: As discussed above 10,000 replications will result in a Monte Carlo P-Value that is correct
to two decimal places. One million (1e6) replications will result in three decimal places of
accuracy and typically require less than 60 seconds to solve for any data set.
Tip: The Monte Carlo 99% confidence interval for P-Value is not the same as a confidence
interval on the test statistic due to data sampling error. The confidence level for the hypothesis
test statistic is still 95%, so all reported P-Values less than .05 will be highlighted in red to
indicate significance. The 99% Monte Carlo P-Value confidence interval is due to the
uncertainty in Monte Carlo sampling, and it becomes smaller as the number of replications
increases (irrespective of the data sample size). The Exact P-Value will lie within the stated
Monte Carlo confidence interval 99% of the time.
321
SigmaXL: Analyze Phase Tools
Fisher’s Monte Carlo P-Value here is 0.0094 with a 99% confidence interval of 0.0072 to 0.0116.
This will be slightly different every time it is run (the Monte Carlo seed value is derived from the
322
SigmaXL: Analyze Phase Tools
system clock). The true Exact P-Value = 0.0101 lies within this confidence interval. If the
confidence interval is too wide (e.g., upper is a “fail-to-reject” H0 and lower is a “reject” H0),
simply rerun the Monte Carlo option using a larger number of replications (use 1e5 or 1e6).
[2] Mehta, C.R. ; Patel, N.R. (1998). "Exact Inference for Categorical Data." In P. Armitage and T.
Colton, eds., Encyclopedia of Biostatistics, Chichester: John Wiley, pp. 1411–1422.
323
SigmaXL: Analyze Phase Tools
1. Click SigmaXL > Statistical Tools > Power & Sample Size Calculators > 1 Proportion Test
Calculator.
2. Select Solve For Power (1 – Beta). Enter Sample Size and Alternative Proportion as shown:
Note that we are calculating the power or likelihood of detection given that the hypothesized
proportion is 0.5, but the alternative proportion is 0.6, sample size = 300, significance level =
.05, and Ha: Not Equal To (two sided test).
A power value of 0.94 is acceptable, but note that the sample size n = 300, and the difference in
proportion value is 0.1 or 10%! The sample size requirements for discrete data are much higher
than those for continuous data.
324
SigmaXL: Analyze Phase Tools
4. To determine Power & Sample Size using a Worksheet, click SigmaXL > Statistical Tools >
Power & Sample Size with Worksheets > 1 Proportion Test.
5. A graph showing the relationship between Power, Sample Size and Proportion Value can then
be created using SigmaXL > Statistical Tools > Power & Sample Size Chart. See Part E for an
example using the 1 Sample t-Test.
325
SigmaXL: Analyze Phase Tools
1. Click SigmaXL > Statistical Tools > Power & Sample Size Calculators > 2 Proportions Test
Calculator.
2. Select Solve For Power (1 – Beta). Enter Sample Size and Proportion 1 values as shown:
Note that we are calculating the power or likelihood of detection given that P1 = 0.5 and P2 =
0.6, sample size for each group = 300, significance level = .05, and Ha: Not Equal To (two sided
test).
A power value of 0.69 is unacceptable. Note that this value is much less than the power for the
one proportion test (0.94).
326
SigmaXL: Analyze Phase Tools
4. To compensate, we will double the sample size per group. Press F3 or click Recall SigmaXL
Dialog to recall last dialog. Change the sample size per group from 300 to 600. Note that the
power value is now 0.94.
5. To determine Power & Sample Size using a Worksheet, click SigmaXL > Statistical Tools >
Power & Sample Size with Worksheets > 2 Proportions Test.
6. A graph showing the relationship between Power, Sample Size and Proportion Values can then
be created using SigmaXL > Statistical Tools > Power & Sample Size Chart. See Part E for an
example using the 1 Sample t-Test.
327
SigmaXL: Analyze Phase Tools
A statistical procedure for troubleshooting industrial processes and analyzing the results of
experimental designs with factors at fixed levels. It provides a graphical display of data. Ellis
R. Ott developed the procedure in 1967 because he observed that nonstatisticians had
difficulty understanding analysis of variance. Analysis of means is easier for quality
practitioners to use because it is (like) an extension of the control chart.
From the preface of the book, “The Analysis of Means: A Graphical Method for Comparing Means,
Rates, and Proportions” by Peter R. Nelson, Peter S. Wludyka, and Karen A. F. Copeland:
The goal of statistical data analysis is to use data to gain and communicate knowledge
about processes and phenomena. Comparing means is often part of an analysis, for data
arising in both experimental and observational studies.
The analysis of means (ANOM) is an alternative procedure (to ANOVA) for comparing
means.
ANOM has the advantages of being much more intuitive and providing an easily understood
graphical result, which clearly indicates any means that are different (from the overall
mean) and allows for easy assessment of practical as well as statistical significance.
328
SigmaXL: Analyze Phase Tools
where:
h = critical value from multivariate t distribution – SigmaXL uses table exact critical values
(Table B.1)
N = sample size
I = number of levels
Formula for One-Way Balanced Binomial Proportions (from ANOM book 2.6)
This is based on the normal approximation to the Binomial, so requires sample sizes large enough
that np and n(p-1) > 5. SigmaXL automatically checks this and gives a warning message if the
condition is not met.
329
SigmaXL: Analyze Phase Tools
Formula for One-Way Balanced Poisson Rates (from ANOM book 2.9)
This is based on the normal approximation to the Poisson, so requires sample sizes large enough
that nu > 5. SigmaXL automatically checks this and gives a warning message if the condition is not
met.
and the above formula for balanced (or unbalanced) Normal is applied to compute decision limits.
where U, L are critical values from Table B.4 (n <= 35). For unbalanced and large sample size see
Appendix Analysis of Means (ANOM) Charts.
and the above formula for balanced (or unbalanced) Normal is applied to compute decision limits.
330
SigmaXL: Analyze Phase Tools
Two-Way Normal
Main Effects for Two-Way ANOM are similar to One-Way but the mean square error (MSE) is
derived from the ANOVA.
Slice Charts are a modified ANOM chart developed by Dr. Peter Wludyka that enables one to easily
interpret the effects in the presence of an interaction (Wludyka 2013, 2015). The basic idea is to
compare the levels of one factor for each level of the other factor with MSE derived from the Two-
Way ANOVA. A yellow highlight automatically recommends Main Effects (if interaction is not
significant, i.e., P-Value >= 0.1) or Slice Chart (if interaction is significant, P-Value < 0.1). The
interaction P-Value is determined from ANOVA
There is an option to specify correction to alpha for multiple chart family-wise error rate:
Note that this is strictly to adjust for multiple charts. As seen in the formulas above, the individual
charts also have correction built-in for the number of levels.
A Two-Way ANOM analysis will typically have balanced data from a designed experiment, with an
equal number of observations for each combination level of X1 and X2. SigmaXL will allow
unbalanced data, but we recommend that you use balanced data or slightly unbalanced data.
ANOM charts use actual data means, not fitted (predicted least squares) means as used in Two-
Way ANOVA Main Effects and Interaction Plots.
As with Normal, the basic idea is to compare the levels of one factor for each level of the other
factor. MSE is derived from the whole model
A yellow highlight automatically recommends Main Effects (if interaction is not significant) or Slice
Chart (if interaction is significant).
The interaction P-Value is automatically determined from Logistic regression for Binomial
Proportions and Poisson regression for Poisson Rates.
See Appendix Analysis of Means (ANOM) Charts for further formula details and references.
331
SigmaXL: Analyze Phase Tools
2. Click SigmaXL > Graphical Tools > Analysis of Means (ANOM) > ANOM Normal One-Way.
Ensure that the entire data table is selected. If not, check Use Entire Data Table.
3. Click Next. Ensure that Stacked Column Format is checked. Select Overall Satisfaction, click
Numeric Data Variable (Y) >>; select Customer Type, click Group Category (X) >>. Set Alpha
Level = 0.05.
332
SigmaXL: Analyze Phase Tools
5. Here we see that Customer Type 1 mean satisfaction score is significantly below the overall
mean and Customer Type 2 is significantly higher. This is consistent with the results that were
observed in the ANOVA analysis, but is easier to interpret.
6. The varying decision limits are due to the varying sample sizes for each Customer Type, with
smaller sample size giving wider limits in a manner similar to a control chart. If the data are
balanced the decision limit lines will be constant.
333
SigmaXL: Analyze Phase Tools
2. Open ANOM Examples.xlsx, click on Hemoglobin – Normal tab. This data is from a factorial
design to study hemoglobin levels in males (Example 5.1). The factors are Therapy (Chemo or
Radiation) and Drug (Type 1, 2, 3).
3. Click SigmaXL > Graphical Tools > Analysis of Means (ANOM) > ANOM Normal Two-Way.
Ensure that the entire data table is selected. If not, check Use Entire Data Table.
4. Click Next. Select Hemoglobin Level, click Numeric Data Variable (Y) >>; select Therapy, click
Group Category Factor (X1) >>; select Drug, click Group Category Factor (X2) >>. Alpha Level =
0.05, Adjust chart alpha for family-wise error rate is unchecked:
5. Click OK. The ANOM Normal Two-Way Main Effects, Slice charts and Interaction plots are
shown below:
334
SigmaXL: Analyze Phase Tools
6. Since the Interaction is not significant (P-Value is >= 0.1), the Main Effects charts are
highlighted. The decision limits match the manual calculations given in the ANOM book (pages
87, 88). Since the average hemoglobin level for radiation falls below the lower decision limit,
and the average for chemotherapy falls above the upper decision limit, therapy has an (α =
0.05) effect. There is also an effect on hemoglobin level due to the drug since the average
hemoglobin level for drug 3 is above the upper decision line.
Tip: Two-Way ANOM Normal is complimentary to Two-Way ANOVA. Analyzing the above data
using Two-Way ANOVA shows that Therapy is significant with a P-Value = 0.0001, Drug is also
significant with a P-Value = 0.0105 and the Interaction is not significant, with a P-Value =
0.5942.
335
SigmaXL: Analyze Phase Tools
7. Click Recall SigmaXL Dialog menu or press F3 to recall last dialog. Check Adjust chart alpha for
family-wise error rate.
This will apply a Bonferroni correction to the specified alpha values to control the overall
family-wise error rate for the charts. So, the Main Effects alpha will be = 0.05/2 = 0.025 and the
Slice Charts alpha will be = 0.05/2 = 0.025 as well.
336
SigmaXL: Analyze Phase Tools
8. Click OK. The ANOM Normal Two-Way Main Effects and Slice charts are shown below:
9. With the Bonferroni corrected alpha of 0.025, the significant results do not change, but they are
closer to the decision lines. The trade-off for this chart family-wise adjustment is loss of power.
337
SigmaXL: Analyze Phase Tools
10. In ANOM Examples.xlsx, click on Process Yield – Normal tab. This data is from a factorial
design to study the effect of three monomers and two levels of UV exposure (1 = UV; 2 = No
UV) on the percent yield for contact lens production (Example 5.3).
11. Click SigmaXL > Graphical Tools > Analysis of Means (ANOM) > ANOM Normal Two-Way.
Ensure that the entire data table is selected. If not, check Use Entire Data Table.
12. Click Next. Select Process Yield, click Numeric Data Variable (Y) >>; select UV, click Group
Category Factor (X1) >>; select Monomer, click Group Category Factor (X2) >>. Alpha Level =
0.05, Adjust chart alpha for family-wise error rate is unchecked:
13. Click OK. The ANOM Normal Two-Way Main Effects, Slice charts and Interaction plots are
shown below:
338
SigmaXL: Analyze Phase Tools
14. Since the Interaction P-Value is < 0.1, the Slice charts are highlighted. The decision limits for
the Slice charts match the calculations given in Wludyka (2013). See Appendix References for
Analysis of Means (ANOM) Charts. The Slice plot clearly shows the effect of the strong
interaction. When UV = 1 (UV Exposure), the three monomers differ. In particular, the average
for monomer 1 is below the overall average and the average for monomer 3 is above the
overall average. However, for UV = 2 (no UV), the monomers also differ but in an entirely
different manner than in the UV = 1 case, revealing the nature of the interaction.
339
SigmaXL: Analyze Phase Tools
15. In ANOM Examples.xlsx, click on Process Yield – Normal Unbal tab. This data is from a
factorial design to study the effect of three monomers and two levels of UV exposure (1 = UV; 2
= No UV) on the percent yield for contact lens production, but two of the observations are
missing for A = 2 (no UV) and B = 3 (monomer 3).
16. Click SigmaXL > Graphical Tools > Analysis of Means (ANOM) > ANOM Normal Two-Way.
Ensure that the entire data table is selected. If not, check Use Entire Data Table.
17. Click Next. Select Process Yield, click Numeric Data Variable (Y) >>; select UV, click Group
Category Factor (X1) >>; select Monomer, click Group Category Factor (X2) >>. Alpha Level =
0.05, Adjust chart alpha for family-wise error rate is unchecked:
18. Click OK. The ANOM Normal Two-Way Main Effects, Slice charts and Interaction plots are
shown below:
340
SigmaXL: Analyze Phase Tools
19. The significant results of the experiment have not changed, but the decision limits are now
varying due to the unbalanced data and use critical values for unbalanced data. The decision
limits for the Slice charts match the calculations given in Wludyka (2013).
341
SigmaXL: Analyze Phase Tools
2. Click SigmaXL > Graphical Tools > Analysis of Means (ANOM) > ANOM Binomial Proportions
One-Way. Ensure that the entire data table is selected. If not, check Use Entire Data Table.
3. Click Next. Select Proficient, click Numeric Data Variable (Y) >>; select Enrollment, click
Subgroup Column or Size >>; select School, click Optional Group Category (X) >>. Set Alpha
Level = 0.01:
4. Click OK. The ANOM Binomial Proportions One-Way chart is shown below:
342
SigmaXL: Analyze Phase Tools
5. The resulting ANOM decision chart shows that three schools are performing at significantly low
levels and two schools are performing at significantly high levels.
6. SigmaXL automatically checks to see if the sample sizes are large enough for the normal
approximation to the Binomial to be valid, i.e., np and n(p-1) are > 5. Here we see: Warning:
Sample sizes are too small to use the normal approximation to the binomial distribution.
Sample #1 has minimum n(1-p) = 2.0. Note that the warning does not show all occurrences,
just the sample(s) with smallest failed np or n(1-p).
7. This does not mean that the chart results are invalid, an obvious “out” or “in” will not likely be
affected, but the results should be used with caution, and if possible, more data collected.
343
SigmaXL: Analyze Phase Tools
2. Click SigmaXL > Graphical Tools > Analysis of Means (ANOM) > ANOM Binomial Proportions
Two-Way. Ensure that the entire data table is selected. If not, check Use Entire Data Table.
3. Click Next. Select LOS > 6 days, click Numeric Data Variable (Y) >>; select Sample Size, click
Subgroup Column or Size >>; select Drug, click Group Category Factor (X1) >>; select
Treatment, click Group Category Factor (X2) >>. Alpha Level = 0.05, Adjust chart alpha for
family-wise error rate is unchecked:
344
SigmaXL: Analyze Phase Tools
4. Click OK. The ANOM Binomial Proportions Two-Way Main Effects, Slice charts and Interaction
plots are shown below:
5. Since the Interaction P-Value is < 0.1 (automatically determined from Logistic regression), the
Slice charts are highlighted. The Slice plot clearly shows the effect of the strong interaction.
When Drug = A, Treatment 2 results in a significantly lower mean proportion LOS and
Treatment 3 results in a significantly higher LOS. When Drug = B, the Treatment does not have
a significant effect on LOS.
345
SigmaXL: Analyze Phase Tools
2. Click SigmaXL > Graphical Tools > Analysis of Means (ANOM) > ANOM Poisson Rates One-
Way. Ensure that the entire data table is selected. If not, check Use Entire Data Table.
3. Click Next. Select Nonconformities, click Numeric Data Variable (Y) >>; enter 100 (or select
Sample Size) for Subgroup Column or Size >>; select Belt Speed, click Optional Group Category
(X) >>. Set Alpha Level = 0.05:
346
SigmaXL: Analyze Phase Tools
4. Click OK. The ANOM Poisson Rates One-Way chart is shown below:
5. From the ANOM chart, one sees that a belt speed of 15 results in a significantly low number of
nonconformities and a belt speed of 30 results in a significantly high number of
nonconformities.
347
SigmaXL: Analyze Phase Tools
2. Click SigmaXL > Graphical Tools > Analysis of Means (ANOM) > ANOM Poisson Rates Two-
Way. Ensure that the entire data table is selected. If not, check Use Entire Data Table.
3. Click Next. Select ER Visits, click Numeric Data Variable (Y) >>; select Sample Size, click
Subgroup Column or Size >>; select Intervention, click Group Category Factor (X1) >>; select
Payor, click Group Category Factor (X2) >>. Alpha Level = 0.01, Adjust chart alpha for family-
wise error rate is unchecked:
348
SigmaXL: Analyze Phase Tools
4. Click OK. The ANOM Poisson Rates Two-Way Main Effects, Slice charts and Interaction plots are
shown below:
5. Since the Interaction is not significant (P-Value >= 0.1, automatically determined from Poisson
regression), the Main Effects charts are highlighted. Both factors Intervention and Payor are
significant at alpha = 0.01. “Home nurse assigned” results in a significantly lower rate of ER
visits versus “no home nurse assigned.” Payor Type 1 has a significantly lower rate of ER visits
and Type 2 has a significantly higher rate.
349
SigmaXL: Analyze Phase Tools
2. Click SigmaXL > Graphical Tools > Analysis of Means (ANOM) > ANOM Nonparametric
Transformed Ranks. Ensure that the entire data table is selected. If not, check Use Entire Data
Table.
3. Click Next. Ensure that Stacked Column Format is checked. Select Overall Satisfaction, click
Numeric Data Variable (Y) >>; select Customer Type, click Group Category (X) >>. Set Alpha
Level = 0.05.
4. Click OK. The ANOM Nonparametric Transformed Ranks chart is shown below:
350
SigmaXL: Analyze Phase Tools
5. Here we see that Customer Type 1 mean transformed rank satisfaction score is significantly
below the grand mean and Customer Type 2 is significantly higher. This is consistent with the
results that were observed in the Kruskal-Wallis analysis (see Kruskal-Wallis).
6. The varying decision limits are due to the varying sample sizes for each Customer Type, with
smaller sample size giving wider limits in a manner similar to a control chart. If the data are
balanced, the decision limit lines will be constant.
7. Open ANOM Examples.xlsx, click on the Exponential Data – TR tab. This is exponential data,
Example 9.16 from the ANOM book with alpha level = 0.01 (used with author permission).
8. Click SigmaXL > Graphical Tools > Analysis of Means (ANOM) > ANOM Nonparametric
Transformed Ranks. Ensure that the entire data table is selected. If not, check Use Entire Data
Table.
9. Click Next. Ensure that Stacked Column Format is checked. Select Y, click Numeric Data
Variable (Y) >>; select X, click Group Category (X) >>. Set Alpha Level = 0.01.
351
SigmaXL: Analyze Phase Tools
10. Click OK. The ANOM Nonparametric Transformed Ranks chart is shown below:
11. X = 1 shows as significantly higher than the grand mean transformed rank and X = 3 is
significantly lower (at alpha = 0.01). These results match those given in the ANOM book.
352
SigmaXL: Analyze Phase Tools
12. By way of comparison, rerun the analysis using ANOM One-Way Normal (with alpha = 0.01).
The results are:
13. ANOM Normal fails to detect the significant differences in mean noted above.
353
SigmaXL: Analyze Phase Tools
ANOM Variances
The ANOM Variances chart is complementary to Bartlett’s Test for Equal Variance and is also
available as an option in the Bartlett’s Test dialog. It assumes that the data are normally
distributed.
2. Click SigmaXL > Graphical Tools > Analysis of Means (ANOM) > ANOM Variances. Ensure that
the entire data table is selected. If not, check Use Entire Data Table.
3. Click Next. Ensure that Stacked Column Format is checked. Select Delivery Time Deviation,
click Numeric Data Variable (Y) >>; select Floor, click Group Category (X) >>; Select Standard
Deviations; Set Alpha Level = 0.05.
We are analyzing the same normal data used in Bartlett’s Test for Equal Variances. See
Bartlett’s Test.
354
SigmaXL: Analyze Phase Tools
5. The ANOM Variances chart visually shows that none of the group standard deviations are
significantly different from the grand mean of all the standard deviations. It is called an ANOM
Variances Chart but displays Standard Deviations for ease of interpretation (similar to a
Standard Deviation S Control Chart). This does however result in non-symmetrical decision
limits. To display Variances rerun the above analysis with option Variances selected.
355
SigmaXL: Analyze Phase Tools
2. Click SigmaXL > Graphical Tools > Analysis of Means (ANOM) > ANOM Levene Robust
Variances. Ensure that the entire data table is selected. If not, check Use Entire Data Table.
3. Click Next. Ensure that Stacked Column Format is checked. Select Responsive to Calls, click
Numeric Data Variable (Y) >>; select Customer Type, click Group Category (X) >>; Set Alpha
Level = 0.05.
We are analyzing the same nonnormal data used in Levene’s Test for Equal Variances. See
Levene’s Test.
356
SigmaXL: Analyze Phase Tools
5. The ANOM chart clearly shows Customer Type 1 has significantly higher variance (ADM) than
overall and Customer Type 2 has significantly lower variance.
6. The varying decision limits are due to the varying sample sizes for each Customer Type, with
smaller sample size giving wider limits in a manner similar to a control chart. If the data are
balanced the decision limit lines will be constant.
357
SigmaXL: Analyze Phase Tools
Multi-Vari Charts
The Multi-Vari chart is a powerful tool to identify dominant Sources of Variation (SOV). The three
major “families” of variation are: Within Unit, Between Unit, and Temporal (Over Time). We will
look at examples of each type of SOV and then use the Multi-Vari Chart to study Overall
Satisfaction in the Customer Data.xlsx file.
1. Open Multi-Vari Data.xlsx, click Sheet Within. Select SigmaXL > Graphical Tools > Multi-Vari
Options.
2. The charts shown will be updated as options are selected. Note that they are for
demonstration purposes and are not based on the Multi-Vari Data.xlsx data (sample charts are
not displayed in Excel for Mac). Ensure that all general Options are selected (Range Line,
Individual Data Points, Min and Max, Standard Deviation Chart). Select Mean Options, ensure
that Show Means, Connect Means, and Group Means are checked. Ensure that Save Defaults
is checked. These settings would be typical for a Multi-Vari chart. (The Median options provide
the ability to display percentiles as an alternative to the Means).
Tip: Multi-Vari Charts can be used to display Confidence Intervals (as we did earlier in Part C).
To do this, check the 95% Confidence Interval.
3. Click Finish. SigmaXL automatically starts the Multi-Vari Chart procedure (this is equivalent to
clicking SigmaXL > Graphical Tools > Multi-Vari Charts).
5. Note that the input X’s can be text or numeric but should be discrete. Y’s must be numeric -
typically continuous, but can also be count or proportion data.
358
SigmaXL: Analyze Phase Tools
6. Select Measurement, click Numeric Response (Y) >>; select unit, click Group Category (X1) >>;
select time, click Group Category (X2) >>.
7. Click OK. Resulting Multi-Vari Chart illustrating dominant “Within Unit” Source of Variation:
8. Using Multi-Vari Data.xlsx, click Sheet Between. Select SigmaXL > Graphical Tools > Multi-Vari
Charts. Check Use Entire Data Table. Click Next.
9. Select Measurement, click Numeric Response (Y) >>; select unit, click Group Category (X1) >>;
select time, click Group Category (X2) >>.
10. Click OK. Resulting Multi-Vari Chart illustrating dominant “Between Unit” Source of Variation:
359
SigmaXL: Analyze Phase Tools
11. Using Multi-Vari Data.xlsx, click Sheet OverTime. Select SigmaXL > Graphical Tools > Multi-
Vari Charts. Check Use Entire Data Table. Click Next.
12. Select Measurement, click Numeric Response (Y) >>; select unit, click Group Category (X1) >>;
select time, click Group Category (X2) >>.
13. Click OK. Resulting Multi-Vari Chart illustrating dominant “Over Time” Source of Variation:
14. Open Customer Data.xlsx, click Sheet 1. Select SigmaXL > Graphical Tools > Multi-Vari Charts.
360
SigmaXL: Analyze Phase Tools
15. Select Overall Satisfaction, click Numeric Response (Y) >>; select Customer Type, click Group
Category (X1) >>; select Size of Customer, click Group Category (X2) >>; select Product Type,
click Group Category (X3) >>.
Examining this Multi-Vari chart reveals that the dominant Source of Variation is “within”
Customer Type, followed by “between” Customer Type. Furthermore, it would be worthwhile
to examine the combination of Customer Type 2, Customer Size Large, and Product Type
Consumer.
Other tools that can help us identify potential X factors that may explain some of the large
“Within” variability are the Scatter Plot, Scatter Plot Matrix and Correlation Matrix.
361
SigmaXL: Analyze Phase Tools
Scatter Plots
1. Open Customer Data.xlsx. Click Sheet 1 Tab. Click SigmaXL > Graphical Tools > Scatter Plots; if
necessary, click Use Entire Data Table, click Next.
2. Select Overall Satisfaction, click Numeric Response (Y) >>; select Avg Days Order Time to
Delivery, click Numeric Predictor (X1) >>. Check Trendline, 95% Confidence Interval and 95%
Prediction Interval as shown:
3. Click OK. The resulting Scatter Plot is shown with equation, trendline, 95% confidence interval
(blue lines – for a given X value this is the 95% confidence interval for predicted mean Overall
Satisfaction) and 95% prediction interval (red lines – for a given X value this is the 95%
confidence interval for predicted individual values of Overall Satisfaction).
The equation is based on linear regression, using the method of least squares. R-squared * 100
is the percent variation of Y explained by X (here 10.3%).
362
SigmaXL: Analyze Phase Tools
4. Now we want to redo the Scatter Plot and stratify by Customer Type. Press F3 or click Recall
SigmaXL Dialog to recall last dialog. (Or, Click Sheet 1 Tab; Click SigmaXL > Graphical Tools >
Scatter Plots; click Next.)
5. Select Overall Satisfaction, click Numeric Response (Y) >>; select Average Days Order Time to
Delivery, click Numeric Predictor (X1) >>; select Customer Type, click Group Category (X2) >>,
Uncheck 95% Confidence Interval and 95% Prediction Interval. Click OK:
Clearly, according to the analysis, Customer Type 3 is happier if orders take longer! But, does
this make sense? Of course not! Customer Sat scores should not increase with Order to
Delivery time. What is happening here? This is a coincidental situation. Something else is
driving customer satisfaction. We will now look at the Scatter Plot Matrix to help us investigate
other factors influencing Customer Satisfaction.
Tip: Be careful when interpreting scatter plots (and other statistical tools): Y versus X
correlation or statistical significance does not always mean that we have a causal relationship.
Umbrella sales are highly correlated to traffic accidents, but we cannot reduce the rate of traffic
accidents by purchasing fewer umbrellas! The best way to validate a relationship is to perform
a Design of Experiments (see Improve Phase).
363
SigmaXL: Analyze Phase Tools
2. Ensure that entire data table is selected. If not, check Use Entire Data Table. Click Next.
3. Select the variable Avg No. of orders per month; shift-click on Staff Knowledge and click
Numeric Data Variables (Y) >> as shown:
364
SigmaXL: Analyze Phase Tools
The correlation matrix complements the scatterplot matrix by quantifying the degree of
association. The following table shows the approximate relationship between r, R-squared, and
degree of association:
Correlation Matrix
1. Open Customer Data.xlsx. Click Sheet 1 tab (or press F4 to activate last worksheet). Click
SigmaXL > Statistical Tools > Correlation Matrix. Ensure that entire data table is selected. If
not, check Use Entire Data Table. Click Next.
2. Select the variable Avg No. of orders per month; shift-click on Staff Knowledge and click
Numeric Data Variables (Y) >> as shown:
365
SigmaXL: Analyze Phase Tools
Correlations highlighted in red are considered significant (P-Values < .05). The corresponding
correlation coefficients above the P-Values are also highlighted in red. (Compare these to the
Scatterplot Matrix.)
Note that Spearman’s Rank Correlation complements Pearson’s Correlation, in that it provides
a robust measure of association. Spearman’s rank is based on correlated ranks, which are not
sensitive to outliers or departures from normality.
Reference
Doornik, J.A. and Hansen, H. “An Omnibus Test for Univariate and Multivariate Normality,” Oxford
Bulletin Of Economics And Statistics, 70, Supplement (2008).
366
SigmaXL: Analyze Phase Tools
Multiple Regression
Multiple Regression analyzes the relationship between one dependent variable (Y) and multiple
independent variables (X's). It is used to discover the relationship between the variables and
create an empirical equation of the form:
This equation can be used to predict a Y value for a given set of input X values. SigmaXL uses the
method of least squares to solve for the model coefficients and constant term. Statistical tests of
hypothesis are provided for the model coefficients.
1. Open Customer Data.xlsx. Click Sheet 1 Tab (or press F4 to activate last worksheet). Click
SigmaXL > Statistical Tools > Regression > Multiple Regression. If necessary, click Use Entire
Data Table, click Next.
2. Select Overall Satisfaction, click Numeric Response (Y) >>, select Responsive to Calls and Ease of
Communications, click Continuous Predictors (X) >>. Fit Intercept and Display Residual Plots
are checked by default.
Note: Fit Intercept is optional. If unchecked the model will not fit an intercept (constant) for
the model. This is useful when you have strong a priori reasons to believe that Y = 0 when the X
or X’s are equal to 0 and the relationship is linear. An example would be Y=automobile fuel
consumption, X=automobile weight.
367
SigmaXL: Analyze Phase Tools
4. The variance inflation factor (VIF) and Tolerance scores are used to measure multicollinearity or
correlation among predictors. VIF = 1 indicates no relation among predictors (which is highly
desirable – you will see examples of this in Design of Experiments); VIF > 1 indicates that the
predictors are correlated; VIF > 5 indicates that the regression coefficients are strongly
correlated and this will lead to poor estimates of the coefficients. Tolerance = 1/VIF.
5. The Durbin-Watson Test is used to determine if the residuals are autocorrelated. If either P-
Value is < .05, then there is significant autocorrelation. This is likely due to an external bias
factor affecting your process over time (e.g. “warm-up” effect, seasonality). This will also be
evident in the plot of residuals versus observation order. Autocorrelation in the residuals is not
a problem for this model.
6. If the Fit Intercept option is unchecked, the following diagnostics are not reported due to
statistical issues associated with the removal of the constant. These are: R-Square, R-Square
Adjusted, VIF (Variance Inflation Factor) and Tolerance collinearity diagnostics, DW (Durbin-
Watson) autocorrelation report, Anova report for Discrete factors (discrete factors will not be
permitted when "Fit Intercept" is unchecked). All other tests including residual diagnostics are
reported. Users can run the model with the "Fit Intercept" on to get the above statistics and
then run with "Fit Intercept" off.
368
SigmaXL: Analyze Phase Tools
7. A predicted response calculator allows you to enter X values and obtain the predicted response
value, including the 95% confidence interval for the long term mean and 95% prediction
interval for individual values:
8. Click the Mult Reg Residuals Sheet and scroll right to view the Residual Plots shown below:
Histogram of Regular Residuals for: Overall Satisfaction Normal Probability Plot of Regular Residualsfor: Overall Satisfaction
60 3.00
50 2.00
40 1.00
Frequency
NSCORE
30 0.00
20 -1.00
10 -2.00
0 -3.00
14
32
50
68
86
04
22
00
50
00
4
0
.9
.7
.5
.4
.2
.0
.0
.5
0.
0.
0.
0.
0.
1.
1.
0.
0.
1.
-0
-0
-0
-0
-0
-0
-1
-0
Regular Residuals Residuals
Regular Residuals vs Data Order for: Overall Satisfaction Regular Residuals vs Predicted Values for: Overall Satisfaction
1.50 1.500000
1.00 1.000000
0.50 0.500000
Regular Residuals
Regular Residuals
0.00 0.000000
-0.50 -0.500000
-1.00 -1.000000
-1.50 -1.500000
50
00
50
00
50
00
50
00
1
1
21
41
61
81
10
1.
2.
2.
3.
3.
4.
4.
5.
Observation Order Fitted Values
1.058332953 1.058332953
0.558332953 0.558332953
Regular Residuals
Regular Residuals
0.058332953 0.058332953
-0.441667047 -0.441667047
-0.941667047 -0.941667047
00
50
00
50
00
50
00
50
00
40
90
40
90
40
90
40
90
1.
1.
2.
2.
3.
3.
4.
4.
5.
1.
1.
2.
2.
3.
3.
4.
4.
Responsive to Calls Ease of Communications
Residuals are the unexplained variation from the regression model (Y - Ŷ). We expect to see the
residuals normally distributed with no obvious patterns in the above graphs. Clearly this is not
the case here, with the Residuals versus Predicted Values indicating there is likely some other X
factor influencing the Overall Satisfaction. It would be appropriate to consider other factors in
the model.
9. SigmaXL also provides Standardized Residuals and Studentized (Deleted t) Residuals. The
standardized residual is the residual, divided by an estimate of its standard deviation. This
makes it easier to detect outliers. Standardized residuals greater than 3 and less than -3 are
considered large (these outliers are highlighted in red). Studentized deleted residuals are
369
SigmaXL: Analyze Phase Tools
computed in the same way that standardized residuals are computed, except that the ith
observation is removed before performing the regression fit. This prevents the ith observation
from influencing the regression model, resulting in unusual observations being more likely to
stand out.
10. Click Recall Last Dialog (or press F3). Change the Residual type to Standardized. Click OK. The
resulting Standardized Residual plots are displayed:
Histogram of Standardized Residuals for: Overall Satisfaction Normal Probability Plot of Standardized Residuals for: Overall Satisfaction
70 3.00
60
2.00
50
1.00
Frequency
40
NSCORE
0.00
30
-1.00
20
-2.00
10
0 -3.00
00
00
00
00
0
65
40
14
89
63
38
12
2
00
00
00
00
00
00
00
.8
.0
.3
.5
.8
.0
.0
.0
.0
.0
0.
1.
2.
2.
3.
4.
5.
0.
1.
2.
3.
4.
5.
6.
-3
-3
-2
-1
-0
-0
-4
-3
-2
-1
Standardized Residuals Standardized Residuals
Standardized Residuals vs Data Order for: Overall Satisfaction Standardized Residuals vs Predicted Values for: Overall Satisfaction
6.00 6.000
5.00 5.000
4.00 4.000
3.00 3.000
Standardized Residuals
Standardized Residuals
2.00 2.000
1.00 1.000
0.00 0.000
-1.00 -1.000
-2.00 -2.000
-3.00 -3.000
-4.00 -4.000
-5.00 -5.000
50
00
50
00
50
00
50
00
1
1
21
41
61
81
10
1.
2.
2.
3.
3.
4.
4.
5.
Observation Order Fitted Values
4.179854 4.179854
3.179854 3.179854
Standardized Residuals
Standardized Residuals
2.179854 2.179854
1.179854 1.179854
0.179854 0.179854
-0.820146 -0.820146
-1.820146 -1.820146
-2.820146 -2.820146
-3.820146 -3.820146
00
50
00
50
00
50
00
50
00
40
90
40
90
40
90
40
90
1.
1.
2.
2.
3.
3.
4.
4.
5.
1.
1.
2.
2.
3.
3.
4.
4.
Responsive to Calls Ease of Communications
11. Other diagnostic measures reported but not plotted include Leverage, Cook's Distance
(Influence), and DFITS. Leverage is a measure of how far an individual X predictor value
deviates from its mean. High leverage points can potentially have a strong effect on the
estimate of regression coefficients. Leverage values fall between 0 and 1. Cook's distance and
DFITS are overall measures of influence. An observation is said to be influential if removing the
observation substantially changes the estimate of coefficients. Cook’s distance can be thought
of as the product of leverage and the standardized residual; DFITS as the product of leverage
and the studentized residual.
12. The easiest way to identify observations with high leverage and/or influence is to plot the
measures on a run chart.
370
SigmaXL: Analyze Phase Tools
371
SigmaXL: Analyze Phase Tools
15. The parameter estimates table now includes Customer Type 2 and Customer Type 3, but where
is Customer Type 1? Since Customer Type is a discrete predictor, SigmaXL applies “dummy
coding” (i.e. Customer Type 1 corresponds to Coded Variable 1 = 0 and Coded Variable 2 = 0;
Customer Type 2 has Coded Variable 1 = 1 and Coded Variable 2 = 0; and Customer Type 3 has
Coded Variable 1 = 0 and Coded Variable 2=1). Hence Customer Type 1 becomes the “hidden”
or reference value. Other statistical software packages may use a different coding scheme
based on -1,0,+1 instead of 0,1. The advantage of a 0,1 coding scheme is the relative ease of
interpretation when making predictions with the model.
Note that for Customer Type 1, you would enter Customer Type 2 = 0 and Customer Type 3 = 0;
for Customer Type 2, as shown, you entered Customer Type 2 = 1 and Customer Type 3 = 0; for
Customer Type 3 you would enter Customer Type 2 = 0 and Customer Type 3 = 1.
17. Note the addition of the Analysis of Variance for Categorical (Discrete) Predictors. The
Customer Type P-Value is .08, so we do not have strong evidence to keep this term in the
model. However, many practitioners will use an alpha value of 0.1 as a criteria for removal.
You are probably wondering why the P-Value for Customer Type is not lower given the results
we saw earlier using ANOVA. The change in P-Value is due to the inclusion of Responsive to
Calls and Ease of Communications in the model. We have higher scores for Ease of
Communications and Responsive to Calls with Customer Type 2. Statistically, Customer Type is
somewhat correlated to Responsive to Calls and Ease of Communications (VIF for Customer
Type = 1.66 and 1.39).
372
SigmaXL: Analyze Phase Tools
This equation can be used to predict an event probability Y value for a given set of input X values.
SigmaXL uses the method of maximum likelihood to solve for the model coefficients and constant
term. Statistical tests of hypothesis and odds-ratios are provided for the model coefficients. The
odds-ratios identify change in likelihood of the event for one unit change in X.
An example application from medical research would be Y=Disease (Yes/No) and X’s = Age, Smoker
(Yes/No), Number Years of Smoking and Weight. The model coefficient P-Values would indicate
which X’s are significant and the odds-ratios would provide the relative change in risk for each unit
change in X.
We will analyze the familiar Customer Satisfaction data using Y=Discrete Satisfaction where the
values have been coded such that an Overall Satisfaction score >= 3.5 is considered a 1, and scores
< 3.5 are considered a 0. Please note, we are not advising that continuous data be converted to
discrete data in actual practice, but simply using the Discrete Satisfaction score for continuity with
the previous analysis.
1. Open Customer Data.xlsx. Click Sheet 1 Tab (or press F4 to activate last worksheet). Click
SigmaXL > Statistical Tools > Regression > Binary Logistic Regression. If necessary, click Use
Entire Data Table, click Next.
2. Select Sat-Discrete, click Binary Response (Y) >>; select Responsive to Calls and Ease of
Communications, click Continuous Predictors (X) >>; select Customer Type, click Categorical
Predictors (X) >>.
373
SigmaXL: Analyze Phase Tools
Note that Response Count (Y)/Sample Size (Trials) should be used when each record contains
both the number of occurrences along with associated sample size. This is common when
tracking daily quality data or performing design of experiments where each run contains a
response of the number of defects and sample size.
374
SigmaXL: Analyze Phase Tools
4. The Likelihood Ratio P-Value < .05 tells us that the model is significant. The P-Values for the
coefficients in the Parameter Estimates table confirm that Responsive to Calls and Ease of
Communications are significant.
5. The P-Value in Wald Estimates for Categorical (Discrete) Predictors table tells us that Customer
Type is not significant here.
Tip: Significance for categorical predictors should be based on the Wald Estimates not the P-
Values given in the Parameter Estimates table.
6. Note that Customer Type 1 is not displayed in the Parameter Estimates table. This is the
“hidden” reference value for Customer Type. Categorical predictors must have one level
selected as a reference value. SigmaXL sorts the levels alphanumerically and selects the first
level as the reference value.
7. Now we will rerun the binary logistic regression but remove Customer Type as a predictor.
Press F3 or click Recall SigmaXL Dialog to recall last dialog. Remove Customer Type by
highlighting Customer Type and double-clicking (or press the Remove button).
375
SigmaXL: Analyze Phase Tools
9. The Odds Ratios in the Parameter Estimates table tell us that for every unit increase in
Responsive to Calls we are 11.7 times more likely to obtain a satisfied customer. For every unit
increase in Ease of Communications we are 13.3 times more likely to obtain a satisfied
customer.
10. McFadden’s Pseudo R-Square mimics the R-square found in linear regression. This value varies
between 0 and 1 but is typically much lower than the traditional R-squared value. A value less
than 0.2 indicates a weak relationship; 0.2 to 0.4 indicates a moderate relationship; greater
than 0.4 indicates a strong relationship. Here we have an R-square value of 0.66 indicating a
strong relationship. This is also confirmed with the Percent Correctly Predicted value of 96%.
376
SigmaXL: Analyze Phase Tools
11. The Pearson, Deviance and Hosmer-Lemeshow Goodness of Fit tests are used to confirm if the
binary logistic model fits the data well. P-Values < .05 for any of these tests indicate a
significant lack of fit. Here the Hosmer-Lemeshow test is indicating lack of fit. Residuals
analysis will help us to see where the model does not fit the data well.
12. The measures of association are used to indicate the relationship between the observed
responses and the predicted probabilities. Larger values for Goodman-Kruskal Gamma, Somers’
D and Kendall’s Tau-a indicate that the model has better predictive ability.
13. The residuals report is given on the Sheet Binary Logistic Residuals. Three types of residuals
are provided: Pearson, Standardized Pearson and Deviance. The Standardized Pearson
Residual is most commonly used and is shown here plotted on a Run Chart. To create, click
SigmaXL > Graphical Tools > Run Chart, check Use Entire Data Table to select the Residuals
data, click Next, select Std. Pearson Residual as the Numeric Data Variable (Y). Click OK.
2.66
Run Chart - Std. Pearson Residual
-1.34
-3.34
-5.34
-7.34
0
1
2
3
4
5
6
7
8
109
11
12
13
14
15
16
17
18
19
20
21
22
234
256
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
432
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
809
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
1099
2
2
Any Standardized Pearson Residual value that is less than -3 or greater than +3 is considered
extreme and should be investigated. There are 3 such outliers here: rows 24, 67, and 77 in the
residuals table. The +3.4 value indicates that the predicted event probability was low (.08) but
the actual result was a 1. The -6.6 value indicates that the predicted event probability was high
(.98) but the actual result was a 0. The large negative residuals have high Responsive to Calls
and Ease of Communications but dissatisfied customers. The reasons for these discrepancies
should be explored further but we will not do so here.
14. Reselect the Binary Logistic sheet. Scroll over to display the Event Probability calculator:
377
SigmaXL: Analyze Phase Tools
This calculator provides a predicted Event Probability for given values of X (in this case the
probability of a satisfied customer). Enter the values 3,3; 4,4; 5,5 as shown:
If Responsive to Calls and Ease of Communications are both equal to 3, the probability of a
satisfied customer is only .07 (7%); if Responsive to Calls and Ease of Communications are both
equal to 5, the probability of a satisfied customer is .9995 (99.95%)
15. Note that if the calculator includes predictors that are categorical (discrete), enter a 0 or 1 to
denote the selected level as shown below (using the original analysis which included Customer
Type):
If we wanted to select Customer Type 1, enter a 0 for both Customer Types 2 and 3. Customer
Type 1 is the hidden reference value.
378
SigmaXL: Analyze Phase Tools
We will analyze the Customer Satisfaction data using Y=Loyalty –Likely to Recommend score which
contains ordinal integer values from 1 to 5, where 5 indicates that the customer is very likely to
recommend us and 1 indicates that they are very likely to not recommend us.
1. Open Customer Data.xlsx. Click Sheet 1 Tab (or press F4 to activate last worksheet). Click
SigmaXL > Statistical Tools > Regression > Ordinal Logistic Regression. If necessary, click Use
Entire Data Table, click Next.
2. Select Loyalty – Likely to Recommend, click Numeric Ordinal Response (Y) >>; select Responsive
to Calls and Ease of Communications, click Continuous Predictors (X) >>.
379
SigmaXL: Analyze Phase Tools
4. The Likelihood Ratio P-Value < .05 tells us that the model is significant. The low P-Values for
the coefficients confirm that Responsive to Calls and Ease of Communications are significant.
5. The Odds Ratios tell us that for every one-unit increase in Responsive to Calls the chance of a
Loyalty score of 1 versus 2 (or 2 versus 3, etc.) is reduced by a multiple of 0.36. This is not very
intuitive but will be easy to see when we use the Response Outcome Probability calculator.
6. McFadden’s Pseudo R-Square value is 0.185 indicating that this is a weak (but close to
moderate) degree of association. This is also confirmed with the Percent Correctly Predicted
value of 48%.
7. The Pearson and Deviance Goodness of Fit (GOF) tests are used to confirm if the ordinal logistic
model fits the data well. P-Values < .05 would indicate a significant lack of fit. Given that the
GOF P-Values are greater than .05, we conclude that there is no significant lack of fit.
380
SigmaXL: Analyze Phase Tools
8. Scroll across to the Response Outcome Probability calculator. This calculator provides
predicted outcome (event) probabilities for given values of X (in this case the probability of a
satisfied customer). Enter the values 3,3; 4,4; 5,5 as shown:
9. Referring to the Predicted Probability for each Level, if Responsive to Calls and Ease of
Communications are both equal to 3, we would expect to see typical loyalty scores of 3 (57%)
with some at 2 (20%) and 4 (20%); if Responsive to Calls and Ease of Communications are both
equal to 5, we would expect typical loyalty scores of 4 (53 %) with some at 3 (34%) and 5 (9%).
381
SigmaXL: Improve Phase Tools: Design of Experiments
(DOE)
The DOE templates are similar to the other SigmaXL templates: simply enter the inputs and
resulting outputs are produced immediately. The DOE templates provide common 2-level designs
for 2 to 5 factors. These basic templates are ideal for training, but use SigmaXL > Design of
Experiments > 2-Level Factorial/Screening Designs to accommodate up to 19 factors with
randomization, replication and blocking.
Click SigmaXL > Design of Experiments > Basic DOE Templates to access these templates:
After entering the template data, main effects and interaction plots may be created by clicking
SigmaXL > Basic DOE Templates > Main Effects & Interaction Plots. The DOE template must be
the active worksheet.
DOE Templates are protected worksheets by default, but this may be modified by clicking SigmaXL
> Help > Unprotect Worksheet.
Advanced analysis is available, but this requires that you unprotect the DOE worksheet. The
following example shows how to use Excel’s Equation Solver and SigmaXL’s Multiple Regression in
conjunction with a DOE template. Caution: If you unprotect the worksheet, do not change the
worksheet title (e.g. Three-Factor, Two-Level, 8-Run, Full-Factorial Design of Experiments). This
title is used by the Main Effects & Interaction Plots to determine appropriate analysis. Also, do not
modify any cells with formulas.
385
SigmaXL: Improve Phase Tools: DOE
1. Open the file DOE Example - Robust Cake.xlsx. This is a Robust Cake Experiment adapted from
the Video Designing Industrial Experiments, by Box, Bisgaard and Fung.
2. The response is Taste Score (on a scale of 1-7 where 1 is "awful" and 7 is "delicious").
3. The five Outer Array Reps have different Cooking Time and Temperature Conditions.
4. The goal is to Maximize Mean and Minimize StDev of the Taste Score.
5. The X factors are Flour, Butter, and Egg. Actual low and high settings are not given in the video,
so we will use coded -1 and +1 values. We are looking for a combination of Flour, Butter, and
Egg that will not only taste good, but consistently taste good over a wide range of Cooking Time
and Temperature conditions.
6. Scroll down to view the Pareto of Abs. Coefficients for Average (Y).
7. The BC (Butter * Egg) interaction is clearly the dominant factor. The bars above the 95%
confidence blue line indicate the factors that are statistically significant; in this case only BC is
significant. Keep in mind that this is an initial analysis. Later, we will show how to do a more
powerful Multiple Regression analysis on this data. (Also, the Rule of Hierarchy states that if an
interaction is significant, we must include the main effects in the model used.)
8. The significant BC interaction is also highlighted in red in the table of Effects and Coefficients:
386
SigmaXL: Improve Phase Tools: DOE
Constant A B C AB AC BC ABC
Avg(Avg(Y)) @ +1: 4.865 4.79 5.125 4.74 4.755 4.06 4.62
Avg(Avg(Y)) @ -1: 4.5 4.575 4.24 4.625 4.61 5.305 4.745
Effect (Delta): 0.365 0.215 0.885 0.115 0.145 -1.245 -0.125
Coefficient (Delta/2): 4.6825 0.1825 0.1075 0.4425 0.0575 0.0725 -0.6225 -0.0625
SE Coefficient: 0.2325 0.2325 0.2325 0.2325 0.2325 0.2325 0.2325 0.2325
T-value 20.1397849 0.7849 0.4624 1.9032 0.24731 0.31183 -2.6774 -0.2688
P-value 9.6928E-20 0.4383 0.6469 0.066 0.80625 0.75719 0.01161 0.7898
9. The R-Square value is given as 27%. This is very poor for a Designed Experiment. Typically, we
would like to see a minimum of 50%, with > 80% desirable.
R-Square: 27.03%
R-Square Adj.: 11.06%
S 1.4705
The reason for the poor R-square value is the wide range of values over the Cooking
Temperature and Time conditions. In a robust experiment like this, it is more appropriate
to analyze the mean response as an individual value rather than as five replicate values.
The Standard Deviation as a separate response will also be of interest.
10. If the Responses are replicated, SigmaXL draws the blue line on the Pareto Chart using an
estimate of experimental error from the replicates. If there are no replicates, an estimate
called Lenth’s Pseudo Standard Error is used.
11. If the 95% Confidence line for coefficients were to be drawn using Lenth’s method, the value
would be 0.409 as given in the table:
387
SigmaXL: Improve Phase Tools: DOE
13. The A (Flour) main effect is clearly the dominant factor, but it does not initially appear to be
statistically significant (based on Lenth’s method). Later, we will show how to do a more
powerful Regression analysis on this data.
14. The Pareto chart is a powerful tool to display the relative importance of the main effects and
interactions, but it does not tell us about the direction of influence. To see this, we must look
at the main effects and interaction plots. Click SigmaXL > Design of Experiments > Basic DOE
Templates > Main Effects & Interaction Plots. The resulting plots are shown below:
388
SigmaXL: Improve Phase Tools: DOE
15. The Butter*Egg two-factor interaction is very prominent here. Looking at only the Main Effects
plots would lead us to conclude that the optimum settings to maximize the average taste score
would be Butter = +1, and Egg = +1, but the interaction plot tells a very different story. The
correct optimum settings to maximize the taste score is Butter = -1 and Egg = +1.
16. Since Flour was the most prominent factor in the Standard Deviation Pareto, looking at the
Main Effects plots for StDev, we would set Flour = +1 to minimize the variability in taste scores.
The significance of this result will be demonstrated using Regression analysis.
389
SigmaXL: Improve Phase Tools: DOE
17. Click on the Sheet Three-Factor 8-Run DOE. At the Predicted Output for Y, enter Flour = 1,
Butter = -1, Egg = 1 as shown:
The predicted average (Y-hat) taste score is 5.9 with a predicted standard deviation (S-hat) of
0.68. Note that this prediction equation includes all main effects, two-way interaction, and the
three-way interaction.
19. In the Coded Design Matrix, highlight columns A to ABC, and the calculated responses as
shown:
20. Click SigmaXL > Statistical Tools > Regression > Multiple Regression. Click Next.
390
SigmaXL: Improve Phase Tools: DOE
21. Select Average (Y), click Numeric Response (Y) >>; holding the CTRL key, select B, C, and BC;
click Continuous Predictors (X) >> as shown:
Note that the R-square value of 92.85% is much higher than the earlier result of 27%. This is
due to our modeling the mean response value rather than considering all data in the outer
array. Note also that the C main effect now appears as significant.
391
SigmaXL: Improve Phase Tools: DOE
24. With the Coded Design Matrix highlighted as before, click SigmaXL > Statistical Tools >
Regression > Multiple Regression. Click Next.
25. Select StDev (Y), click Numeric Response (Y) >>; select A, click Continuous Predictors (X) >> as
shown:
Note that Factor A (Flour) now shows as a statistically significant factor affecting the Standard
Deviation of Taste Score.
27. Now we will use Excel’s Equation Solver to verify the optimum settings determined using the
Main Effects and Interaction Plots.
392
SigmaXL: Improve Phase Tools: DOE
28. Click on the Sheet Three-Factor 8-Run DOE. At the Predicted Output for Y, enter 1 for Flour.
We are setting this as a constraint, because Flour = +1 minimizes the Standard Deviation. Reset
the Butter and Egg to 0 as shown:
29. In Excel 2007 click Office Button | Excel Options | Add-Ins > Manage Excel Add-Ins, click Go; in
Excel 2010: File | Options. Ensure that the Solver Add-in is checked. If the Solver Add-in does
not appear in the Add-ins available list, you will need to re-install Excel to include all add-ins.
30. Click OK. Click Data | Analysis | Solver). Set the Solver Parameters as shown:
Cell J11 is the Y-hat, predicted average taste score. Solver will try to maximize this value. Cells
H11 to H13 are the Actual Factor Settings to be changed. Cells I11 to I13 are the Coded Factor
393
SigmaXL: Improve Phase Tools: DOE
settings where the following constraints are given: I11=1; I12 >= -1; I12 <= 1; I13 >= -1; I13 <=1.
31. Click Solve. The solver results are given in the Predicted Output for Y as Butter = -1 and Egg =
1.
394
SigmaXL: Improve Phase Tools: DOE
4. The available designs are then given as: 4-Run, 2**(3-1), 1/2 Fraction, Res III and 8-Run, 2**3,
Full-Factorial. If we had more than 5 factors, a Resolution III or Plackett-Burman Screening
design would typically be used. Here we will choose the 8-Run, 2**3, Full-Factorial design.
395
SigmaXL: Improve Phase Tools: DOE
5. Next we will review the Power Information to help us determine the number of replicates. This
design currently shows the following:
Note: the power calculations require an estimate of experimental error, hence an assumption
of 3 center points is used.
We would like to have a reasonable chance (medium power) to detect an Effect = 2*StDev.
Change the Number of Replicates to 2. The Power Information is now:
Very Low Power to detect Effect = 1*StDev (1-Beta < 0.5);
Medium Power to detect Effect = 2*StDev (0.8 <= 1-Beta < 0.95);
Very High Power to detect Effect = 3*StDev (1-Beta >= 0.99).
We will therefore choose two replicates. The number of replicates will always be a tradeoff
between the desired power and the cost of the experimental runs.
Specify 2 or more blocks if there are constraints such as the number of runs per day or some
other known external “nuisance” variable (like 2 different catapults or 2 operators). Here we
will keep Blocks = 1 (i.e. no Blocking).
Center Points are useful to provide an estimate of experimental error with unreplicated
designs, and allow detection of curvature. Typically 3 to 5 center points are used. Here we will
not use center points because we have replicated the design twice and do not expect significant
396
SigmaXL: Improve Phase Tools: DOE
curvature in the distance response. Furthermore, center points could not be set for Pin Height
and Stop Pin (without drilling additional holes!).
6. Complete the Factor Names, Level Settings and Response Name as shown:
8. You can enter information about the experiment in the fields provided. If you have access to a
catapult, perform the experimental runs in the given randomized sequence, and enter the
distance values in the Distance column.
9. If you are not able to perform the catapult experiment, open the file Catapult DOE V6.xlsx.
397
SigmaXL: Improve Phase Tools: DOE
10. Before we begin the regression analysis we will have a quick look at the Main Effects and
Interaction Plots. Click SigmaXL > Design of Experiments > 2 Level Factorial/Screening > Main
Effects & Interaction Plots. The resulting plots are shown below:
Pull Back Angle is the dominant factor having the steepest slope. We can also see that the
interaction terms are weak with the almost parallel lines.
398
SigmaXL: Improve Phase Tools: DOE
11. Click SigmaXL > Design of Experiments > 2-Level Factorial/Screening > Analyze 2-Level
Factorial/Screening Design.
12. We will use the default analyze settings (all terms in the model) to start. Click OK. The
resulting Analysis report is shown:
399
SigmaXL: Improve Phase Tools: DOE
400
SigmaXL: Improve Phase Tools: DOE
13. The model looks very good with an R-Square value of 99.9%! The standard deviation
(experimental error) is only 1.03 inches. Clearly Pull Back Angle is the most important predictor
(X factor), but all the main effects and two way interaction are significant. However, the three-
way interaction is not significant, so it should be removed from the model.
401
SigmaXL: Improve Phase Tools: DOE
402
SigmaXL: Improve Phase Tools: DOE
17. All the terms in the model are now significant, and there is no evidence of lack of fit (P-Value
for lack-of-fit is 0.128 which is > .05).
1.60
5
1.10
4
0.60
Frequency
NSCORE
3 0.10
-0.40
2
-0.90
1
-1.40
0 -1.90
21
62
03
44
00
00
00
00
4
0
.4
.0
.6
.2
.0
.0
.0
0.
0.
1.
1.
0.
1.
2.
3.
-1
-1
-0
-0
-3
-2
-1
Regular Residuals Regular Residuals
Regular Residuals vs Predicted Values for: Distance Regular Residuals vs Data Order for: Distance
2.000 2.00
1.500 1.50
1.000 1.00
Regular Residuals
0.000
0.00
-0.500
-0.50
-1.000
-1.00
-1.500
-1.50
-2.000
-2.00
00
00
00
00
0
0
.0
.0
.0
0.
0.
0.
0.
40
60
80
9
10
12
14
16
11
13
15
17
Fitted Values Observation Order
19. Scroll up to the Predicted Response Calculator. Enter the predicted values shown. These initial
settings were determined by trial and error.
403
SigmaXL: Improve Phase Tools: DOE
20. Excel’s Solver may also be used to get a more exact solution:
21. The model prediction must then be confirmed with actual experimental runs at the given
settings of Pull Back Angle = 179.5, Stop Pin = 2, and Pin Height = 2.
22. Alternative settings to achieve the target distance may be obtained with Contour/Surface Plots.
Click SigmaXL > Design of Experiments > 2-Level Factorial/Screening > Contour/Surface Plots.
Set the Pin Height to 2 as shown (after clicking OK, you can use Recall SigmaXL Dialog to create
another Contour/Surface plot with Pin Height set to 3) :
404
SigmaXL: Improve Phase Tools: DOE
23. Click OK. The following Contour and Surface Plots are displayed (with Pin Height = 2). Note the
contour line with Catapult target distance = 100 inches. Although pin settings are discrete, they
appear as continuous, so this will be a constraint in our selection of alternative settings. In
addition to Pull Back Angle = 179.5, Stop Pin = 2, Pin Height = 2, we see that Pull Back Angle =
171.4, Stop Pin = 3, Pin Height = 2 is also a valid setting. Alternative setting options are valuable
in a designed experiment because they allow you to select lowest cost optimum settings, or
settings that are easier to control.
Alternative
setting to
achieve a target
distance of 100
inches
Tip: Use the contour/surface in conjunction with the predicted response calculator to
determine optimal settings.
405
SigmaXL: Improve Phase Tools: DOE
1. We will illustrate the use of response surface methods using a layer cake baking experiment.
The response variable is Taste Score (on a scale of 1-7 where 1 is "awful" and 7 is "delicious").
Average scores for a panel of tasters have been recorded. The X Factors are A: Bake Time (20
to 40 minutes) and B: Oven Temperature (350 to 400 F). The experiment goal is to find the
settings that maximize taste score. Other factors such as pan size need to be taken into
consideration. In this experiment, these are held constant.
2. Click SigmaXL > Design of Experiments > Response Surface > Response Surface Designs.
4. The available designs are sorted by number of runs. Increasing the number of runs allows for
uniform precision or blocking, but we will select the design with fewest runs, the 10-Run,
Central Composite Design (2 Ctr Pts).
6. Change the Alpha Axial Value option to Face Centered (Alpha = 1.0). This simplifies the design
to a 3 level design, rather than a 5 level design with alpha = 1.414. (The trade-off is that we lose
the desirable statistical property of rotatability for prediction).
Tip: This alpha is the distance from the center point to the end of the axial (star) point. Do not
confuse this with Alpha for Pareto Chart which is the P-Value used to determine statistical
significance. Unfortunately the term “alpha” has been chosen by statisticians to define two
completely different things.
7. Enter Factor Names and Level Settings and Response Name as shown:
406
SigmaXL: Improve Phase Tools: DOE
9. Open the file RSM Example – Cake Bake to obtain response values.
407
SigmaXL: Improve Phase Tools: DOE
10. Click SigmaXL > Design of Experiments > Response Surface > Analyze Response Surface
Design.
11. We will use the default analyze settings (all terms in the model, including the block term) to
start. Click OK. The resulting Analysis report is shown:
408
SigmaXL: Improve Phase Tools: DOE
12. The model looks very good with an R-Square value of 99.09%. The standard deviation
(experimental error) is only 0.19 on a 1 to 7 taste scale. All of the model terms are statistically
significant (P < .05), but the Block term is not, so it should be removed from the model. Note
that AA and BB denote the quadratic model terms.
409
SigmaXL: Improve Phase Tools: DOE
410
SigmaXL: Improve Phase Tools: DOE
16. To create a contour and surface plot, click SigmaXL > Design of Experiments > Response
Surface > Contour/Surface Plots.
17. Click OK. The following Contour and Surface Plots are displayed.
Approximate
Maximum
Taste Score
411
SigmaXL: Improve Phase Tools: DOE
18. Go to Analyze – 2 Factor RSM tab. Scroll across to the Predicted Response Calculator. Enter
the predicted values shown. These initial settings were determined from the contour plot as an
estimate to yield a maximum taste score.
19. Excel’s Solver may also be used to get a more exact solution:
20. Although the model is predicting values that exceed the maximum taste score of 7, this is
expected to give us the best possible settings for cook time and bake temperature. To quote
the eminent statistician George Box, “All models are wrong, some are useful”. Additional
experimental runs carried out at Time = 23.4 minutes and Temperature = 367.7 F. confirm that
these are ideal settings.
412
SigmaXL: Control Phase Tools: Statistical Process Control
(SPC) Charts
Tip: See Part E – Control Chart Selection Tool. The Control Chart Selection Tool makes it easy for
you to select the correct statistical process control chart depending on data type and
subgroup/sample type and size.
This template is also located at SigmaXL > Control Charts > Basic Control Chart Templates >
Individuals Chart.
See Measure Phase Part B – Templates and Calculators for Individuals Chart template
example:
415
SigmaXL: Control Phase Tools: SPC
Individuals Chart
1. Open Customer Data.xlsx, click on Sheet 1. Click SigmaXL>Control Charts>Individuals. Ensure
that entire data table is selected. If not, check Use Entire Data Table. Click Next.
2. Select Overall Satisfaction, click Numeric Data Variable (Y) >>. Ensure that Calculate Limits is
selected.
4. We have seen this data earlier as a run chart. The Control Chart adds calculated control limits.
Note that the Upper Control Limit exceeds the survey upper limit of 5. Here it would be
appropriate to change the UCL to 5.0. Click Recall SigmaXL Dialog menu or press F3 to recall
last dialog.
5. Select Overall Satisfaction, click Numeric Data Variable (Y) >>, change UCL to 5.
416
SigmaXL: Control Phase Tools: SPC
Tip: You should not change the calculated control limits unless you have a legitimate reason
such as a boundary condition. Do not replace the control limits with specification limits – the
control chart will lose its statistical ability to detect assignable causes. We will redo the
Individuals chart for Overall Satisfaction later using the Individuals Nonnormal tool.
7. Open Delivery Times.xlsx. Click Sheet 1 Tab. This data set contains room service delivery time
deviations in minutes. The Critical Customer Requirement is target time +/- 10 minutes.
8. Click SigmaXL > Control Charts > Individuals. Ensure that entire data table is selected. If not,
check Use Entire Data Table. Click Next.
9. Select Delivery Time Deviation, click Numeric Data Variable (Y) >>, check Tests for Special
Causes, check Sigma Zone Lines. Ensure that Calculate Limits is selected.
10. Select Advanced Options, specify LSL = -10, Target = 0, USL = 10 as shown:
417
SigmaXL: Control Phase Tools: SPC
418
SigmaXL: Control Phase Tools: SPC
Some tests for special causes are indicated on the chart. If more than one test fails, the number
corresponds to the first failed test.
12. There are no points that exceed the +/- 3 sigma limits on this chart, but we see some indication
of instability with tests for special causes. The Tests for Special Causes report below the chart
provides detailed information about each observation identified as a special cause. Note that
the control chart also shows the +/- 1 sigma and +/- 2 sigma lines to aid in viewing these tests.
419
SigmaXL: Control Phase Tools: SPC
13. These tests for special causes can have defaults set to apply any or all of Tests 1-8. Test 2 can
be set to 7, 8, or 9 points in a row on same side of CL. Test 3 can be set to 6 or 7 points in a row
all increasing or decreasing. Test 7 can be set to 14 or 15 points in a row within 1 standard
deviation from CL. Click Sheet 1 Tab. Click SigmaXL > Control Charts > “Tests for Special
Causes” Defaults to run selected tests for special causes:
Note that these defaults will apply to Individuals and X-bar charts. Test 1 to 4 settings will be
applied to Attribute Charts.
15. If you prefer to create a separate worksheet for each Tests for Special Causes report, choose
Create a new sheet option. The default is to display Tests for Special Causes on the same sheet
as the Control Chart. Note that this report will be overwritten when you add data or
recalculate control limits.
420
SigmaXL: Control Phase Tools: SPC
17. Click on Indiv Proc Cap Tab to view the Process Capability Report, which includes potential
(short term) capability indices Cp and Cpk:
18. While this process demonstrated some slight instability on the control charts, the bigger issue
was being late 6 minutes on average and having a Standard Deviation of 7.2 minutes! One
improvement implemented was rescheduling the service elevators so that Room Service and
Maintenance were not both trying to use them during peak times.
421
SigmaXL: Control Phase Tools: SPC
19. Click on the Indiv sheet. With 725 data points, you may want to have a closer look at the most
recent data. To do this, click SigmaXL Chart Tools > Show Last 30 Points. (If this menu item
does not appear, click on any cell adjacent to the chart.) The resulting chart is shown:
20. To reset the chart, click SigmaXL Chart Tools > Show All Data Points.
21. To enable scrolling, click SigmaXL Chart Tools > Enable Scrolling. A warning message is given:
Scrolling will clear user custom formats, but does not affect Tests for Special Causes.
22. Click OK. A scroll dialog appears allowing you to specify the Start Subgroup and Window
Width:
23. At any point, you can click Restore/Show All Data Points or Freeze Chart. Freezing the chart
will remove the scroll and unload the dialog. The scroll dialog will also unload if you change
worksheets. To restore the dialog, click SigmaXL Chart Tools > Enable Scrolling.
422
SigmaXL: Control Phase Tools: SPC
24. Click OK. The control chart appears with the scroll bar beneath it. You can also change the
Start Subgroup and Window Width and Update.
423
SigmaXL: Control Phase Tools: SPC
2. Click SigmaXL > Control Charts > Individuals. Ensure that entire data table is selected. If not,
check Use Entire Data Table. Click Next.
3. Select Distance, click Numeric Data Variable (Y) >>; select Operator, click Optional X-Axis
Labels >>.
4. Click Advanced Options. Select Specify Historical Groups. Select Before_After, click Historical
Groups >>.
Note: Process Capability analysis is not permitted when Historical Groups are used.
424
SigmaXL: Control Phase Tools: SPC
5. Click OK. The resulting Individuals Control Chart with split limits based on historical groups is
shown, demonstrating a clear process improvement:
425
SigmaXL: Control Phase Tools: SPC
2. Select Overall Satisfaction, click Numeric Data Variable (Y) >>. Ensure that Calculate Limits is
selected.
426
SigmaXL: Control Phase Tools: SPC
1. Open the file Nonnormal Cycle Time2.xlsx. This contains continuous nonnormal data of
process cycle times. We performed a Process Capability study with this data earlier in the
Measure Phase, Part H.
2. Initially, we will ignore the nonnormality in the data and construct an Individuals Chart. Click
SigmaXL > Control Charts > Individuals. Ensure that entire data table is selected. If not, check
Use Entire Data Table. Click Next.
3. Select Cycle Time (Minutes), click Numeric Data Variable (Y) >>. Select Calculate Limits. Check
Tests for Special Causes. Click OK.
427
SigmaXL: Control Phase Tools: SPC
This chart clearly shows that the process is “out-of-control”. But is it really? Nonnormality can
cause serious errors in the calculation of Individuals Chart control limits, triggering false alarms
(Type I errors) or misses (Type II errors).
5. We will now construct Individuals Control Charts for nonnormal data. Select Sheet 1 Tab (or
press F4). Click SigmaXL > Control Charts > Nonnormal > Individuals Nonnormal. Ensure that
the entire data table is selected. If not, check Use Entire Data Table. Click Next.
6. Select Cycle Time (Minutes), click Numeric Data Variable (Y) >>. We will use the default
selection for Transformation/Distribution Options: Box-Cox Transformation with Rounded
Lambda. Check Tests for Special Causes as shown:
428
SigmaXL: Control Phase Tools: SPC
Note that there are no out-of-control signals on the control charts, so the signals observed earlier
when normality was assumed were false alarms.
429
SigmaXL: Control Phase Tools: SPC
The Individuals – Original Data chart displays the untransformed data with control limits calculated
as:
CL = 50th percentile
The benefit of displaying this chart is that one can observe the original untransformed data. Since
the control limits are based on percentiles, this represents the overall, long term variation rather
than the typical short term variation. The limits will likely be nonsymmetrical.
The Individuals/Moving Range – Normalized Data chart displays the transformed z-values with
control limits calculated using the standard Shewhart formulas for Individuals and Moving Range
charts. The benefit of using this chart is that tests for special causes can be applied and the control
limits are based on short term variation. The disadvantage is that one is observing transformed
data on the chart rather than the original data.
430
SigmaXL: Control Phase Tools: SPC
Tip: See Part E – Control Chart Selection Tool. The Control Chart Selection Tool makes it easy for
you to select the correct statistical process control chart depending on data type and
subgroup/sample type and size.
2. Select B2:F22 (if not already selected); here, we will only use the first 20 subgroups to
determine the control limits.
3. Select SigmaXL > Control Charts > X-Bar & R. Do not check Use Entire Data Table!
4. Click Next. Select Subgroups across Rows, select Shot 1, Shot 2, Shot 3, click Numeric Data
Variables (Y) >>; select Operator, click Optional X-Axis Labels >>. Check Tests for Special
Causes as shown:
431
SigmaXL: Control Phase Tools: SPC
432
SigmaXL: Control Phase Tools: SPC
6. This is currently a stable catapult process. Subgroups 21 to 25 were added afterwards. To add
the additional data to this chart, click SigmaXL Chart Tools > Add Data to this Control Chart as
shown. In Excel 2007+ click:
433
SigmaXL: Control Phase Tools: SPC
7. Note that the Add Data button does NOT recalculate the control limits. Once control limits are
established, they should only be recalculated when a deliberate process change or
improvement is introduced. (Control Limits can be recalculated using SigmaXL Chart Tools >
Recalculate Control Limits, but a warning message Are you sure that you want to recalculate
control limits? is given.)
8. The Tests for Special Causes report gives us more detail on the recent instability:
The X-bar chart and Tests for Special Causes report clearly shows that this process is now out of
control with an unstable mean. The process must be stopped, and the Out-of-Control Action
Plan must be followed to determine and fix the root cause. In this case, the assignable cause
was a change of rubber band requiring a reset of the pull back angle. The use of tests for
special causes gave us an early warning of this at observation number 22.
Note that the Range chart is in-control even though the X-Bar chart is out-of-control.
9. The tests for special causes can have defaults set to apply any or all of Tests 1-8. Test 2 can be
set to 7, 8, or 9 points in a row on same side of CL. Click SigmaXL > Control Charts > “Tests for
Special Causes” Defaults to run selected tests for special causes. (Note that these defaults will
apply to Individuals and X-bar charts. Test 1 to 4 settings will be applied to Attribute Charts.)
10. To add a comment to a data point, select SigmaXL Chart Tools > Add Data Label. Select Text
Label. Enter a comment as shown. Click on the data point to add the comment. Click Done.
11. Now we will look at Process Capability Indices for this process. Click on Sheet 1 (or press F4 to
activate last worksheet). Click SigmaXL > Control Charts > X-Bar & R. Check Use Entire Data
Table. Click Next. (Alternatively select B2:F27, press F3.)
12. Select Shots 1-3, click Numeric Data Variables (Y) >>.
13. Click Advanced Options. Enter LSL = 92, Target = 100, USL = 108.
434
SigmaXL: Control Phase Tools: SPC
435
SigmaXL: Control Phase Tools: SPC
15. Click OK. Click X-Bar & R – Proc Cap sheet for the Process Capability report:
Note the difference between Pp and Cp; Ppk and Cpk. This is due to the process instability. If
the process was stable, the actual performance indices Pp and Ppk would be closer to the Cp
and Cpk values.
436
SigmaXL: Control Phase Tools: SPC
16. Click on Sheet 1 (or press F4 to activate last worksheet). Click SigmaXL > Control Charts > X-Bar
& R. Check Use Entire Data Table. Click Next.
17. Select Shots 1-3, click Numeric Data Variables (Y) >>. Ensure that Calculate Limits is selected.
Click OK.
The control limits here were calculated including subgroups 21 to 25 which have a known
assignable cause.
437
SigmaXL: Control Phase Tools: SPC
19. To calculate the control limits excluding subgroups 21 to 25, click SigmaXL Chart Tools >
Exclude Subgroups. Select Show Highlighted Points for Excluded Subgroups. Enter
21,22,23,24,25 as shown:
20. Click Exclude Subgroups. The control chart limits are recalculated and the excluded points are
highlighted:
Tip: You can also choose to show gaps for excluded subgroups or delete excluded subgroups
from the charts.
438
SigmaXL: Control Phase Tools: SPC
2. Select B2:F22; here, we will only use the first 20 subgroups to determine the control limits.
5. Click Next. Select Subgroups across Rows, select Shot 1, Shot 2, Shot 3, click Numeric Data
Variables (Y) >>; select Operator, click Optional X-Axis Labels >>. Check Tests for Special
Causes as shown:
439
SigmaXL: Control Phase Tools: SPC
7. This is currently a stable catapult process. Subgroups 21 to 25 were added afterwards. To add
the additional data to this chart, click SigmaXL Chart Tools > Add Data to this Control Chart as
shown. In Excel 2007+ click:
440
SigmaXL: Control Phase Tools: SPC
Tip: See Part E – Control Chart Selection Tool. The Control Chart Selection Tool makes it easy for
you to select the correct statistical process control chart depending on data type and
subgroup/sample type and size.
P-Charts
1. Open New York Daily Cycle Time – Discrete.xlsx. This is data from the Sigma Savings and Loans
Company, New York location. Each day, the cycle time (in days) for completed loans and leases
was recorded. N indicates the number of loans counted. A Fail was recorded if the cycle time
exceeded the critical customer requirement of 8 days. Note that we are not recommending
that continuous data be converted to discrete data in this manner, but rather using this data to
illustrate the use of P charts for Discrete or Attribute data. P Charts (for Defectives) can have
fixed or varying subgroup sizes.
2. Select SigmaXL > Control Charts > Attribute Charts > P. Ensure that B3:E23 are selected, click
Next.
3. Select Fail, click Numeric Data Variable (Y) >>; select N, click Subgroup Column or Size >>. If
we had a fixed subgroup size, the numerical value of the subgroup size could be entered
instead of Column N. Ensure that Tests for Special Causes is unchecked.
Note: Tests for Special Causes cannot be used with varying subgroup sizes.
441
SigmaXL: Control Phase Tools: SPC
The moving limits are due to the varying subgroup sizes. While this P-chart shows stability, a
much bigger concern is the average 41% failure rate to deliver the loans/leases in 8 days or less!
2. Click SigmaXL > Control Charts > Attribute Charts > P. Ensure that the entire data table is
selected. If not, check Use Entire Data Table. Click Next.
3. Select Fail, click Numeric Data Variable (Y) >>; select N, click Subgroup Column or Size >>.
Ensure that Tests for Special Causes is unchecked.
4. Click Advanced Options. Select Specify Historical Group Column. Select Before_After, click
Historical Groups >>.
442
SigmaXL: Control Phase Tools: SPC
5. Click OK. The resulting P-Chart with split limits based on historical groups is shown,
demonstrating a clear process improvement:
443
SigmaXL: Control Phase Tools: SPC
NP-Charts
1. Open Attribute Data NP Chart Defectives.xlsx. The subgroup size is constant at 50. NP Charts
(for Defectives) require a fixed subgroup size.
2. Select SigmaXL > Control Charts > Attribute Charts > NP. Ensure that the entire data table is
selected. If not, check Use Entire Data Table. Click Next.
3. Select Defectives, click Numeric Data Variable (Y) >>; enter 50 for Enter Subgroup Size. Check
Tests for Special Causes.
Note: Tests for Special Causes can always be used with NP charts because the subgroup size is
fixed, which results in constant control limits.
444
SigmaXL: Control Phase Tools: SPC
C-Charts
1. Open Attribute Data C Chart Defects.xlsx. C Charts (for Defects) assume a fixed subgroup size.
2. Select SigmaXL > Control Charts > Attribute Charts > C. Ensure that the entire data table is
selected. If not, check Use Entire Data Table. Click Next.
3. Select Defects, click Numeric Data Variable (Y) >>. Check Tests for Special Causes.
Note: Tests for Special Causes can always be used with C charts because the subgroup size is
assumed to be fixed, which results in constant control limits.
445
SigmaXL: Control Phase Tools: SPC
U-Charts
1. Open Attribute Data U Chart Defects.xlsx. U Charts (for Defects) can have fixed or varying
subgroup sizes.
2. Select SigmaXL > Control Charts > Attribute Charts > U. Ensure that the entire data table is
selected. If not, check Use Entire Data Table. Click Next.
3. Select Defects, click Numeric Data Variable (Y) >>. Select N, click Subgroup Column or Size >>.
Ensure that Tests for Special Causes is unchecked (Tests for Special Causes cannot be used
with varying subgroup sizes).
446
SigmaXL: Control Phase Tools: SPC
P' and U' (Laney) Control Charts are attribute control charts that should be used when the
subgroup/sample size is large and assumptions are not met. Typically you will see that the control
limits do not “look right,” being very tight with many data points appearing to be out-of-control.
This problem is also referred to as overdispersion. This occurs when the assumption of a Binomial
distribution for defectives or Poisson distribution for defects is not valid. Individuals charts are
often recommended in these cases, but Laney’s P’ and U’ charts are a preferred alternative.
Tip: See Part E – Control Chart Selection Tool. The Control Chart Selection Tool makes it easy for
you to select the correct statistical process control chart depending on data type and
subgroup/sample type and size.
References:
[1] Crossley, Mark L.(2007), The Desk Reference of Statistical Quality Methods, Second Edition,
Milwaukee, WI, ASQ Quality Press, pp. 352 – 355.
[2] Laney, David B., “P-Charts and U-Charts Work (But Only Sometimes),” Quality Digest,
http://www.qualitydigest.com/sept07/departments/what_works.shtml.
[3] Laney, David B. (2002), “Improved Control Charts for Attribute Data,” Quality Engineering
14:531–7.
[4] M. A. Mohammed and D. Laney (2006), “Overdispersion in health care performance data:
Laney’s approach,” Qual. Saf. Health Care 15; 383-384.
P’-Charts
1. Open Laney – Quality Digest – Defectives.xlsx. This data is used with permission from David
Laney and was used in the example given in reference [2].
2. We will begin with the creation of a regular P-Chart for this data. Select SigmaXL > Control
Charts > Attribute Charts > P. Ensure that the entire data table is selected. If not, check Use
Entire Data Table. Click Next.
3. Select Defectives as the Numeric Data Variable (Y), N as the Subgroup Column or (Size).
Ensure that Tests for Special Causes is unchecked (Tests for Special Causes cannot be used
with varying subgroup sizes). Click OK. The resulting P-Chart is shown:
447
SigmaXL: Control Phase Tools: SPC
At first glance, this chart suggests that the process is out of control. The problem here is
actually due to the Binomial assumption not being valid (also known as overdispersion). This
problem becomes evident when the sample sizes are large.
4. Select Sheet Defectives (or press F4). Click SigmaXL > Control Charts > Attribute Charts > P’
(Laney). Click Next.
5. Select Defectives, click Numeric Data Variable (Y) >>; select N, click Subgroup Column or Size
>>. (If we had a fixed subgroup size, the numerical value of the subgroup size could be entered
instead of Column N.)
448
SigmaXL: Control Phase Tools: SPC
7. Now we see that the process is actually “in-control.” Laney’s Sigma (Z) is a measure of the
overdispersion. See referenced articles for further details.
U’-Charts
1. Open Attribute Data – U’ Defects - Crossley.xlsx. This data is used with permission from Mark
Crossley and was used in the example given in reference [1]. Note that there are multiple
defects per unit.
2. We will begin with the creation of a regular U-Chart for this data. Select SigmaXL > Control
Charts > Attribute Charts > U. Ensure that the entire data table is selected. If not, check Use
Entire Data Table. Click Next.
3. Select Defects as the Numeric Data Variable (Y), N as the Subgroup Column (Size). Click OK.
The resulting U-Chart is shown:
This chart suggests that the process is out of control, but the problem here is actually due to
the Poisson assumption not being valid (also known as overdispersion). This problem becomes
evident when the sample sizes are large.
4. Select Sheet U’ Chart Defects - Crossley (or press F4). Click SigmaXL > Control Charts >
Attribute Charts > U’ (Laney). Click Next.
449
SigmaXL: Control Phase Tools: SPC
5. Select Defects, click Numeric Data Variable (Y) >>; select N, click Subgroup Column or Size >>.
(If we had a fixed subgroup size, the numerical value of the subgroup size could be entered
instead of Column N.) Ensure that Tests for Special Causes is unchecked (Tests for Special
Causes cannot be used with varying subgroup sizes).
7. Now we see that the process is actually “in-control.” Laney’s Sigma (Z) is a measure of the
overdispersion. See reference [1] for further details.
450
SigmaXL: Control Phase Tools: SPC
The Control Chart Selection Tool makes it easy for you to select the correct statistical process
control chart depending on data type and subgroup/sample type and size.
Defect: Any specific nonconformity to customer requirements. There can be more than one
defect per unit or area of opportunity, such as the entry errors described above.
Subgroup/Sample: Data for a subgroup are usually collected within a short period of time to
ensure homogeneous conditions within the subgroup (common cause variation), in order to
detect differences between subgroups (special cause variation).
Subgroup/Sample Size: The number of observations within your sample, not the number of
samples. Subgroup sizes of 3 to 5 are common for continuous measures in parts
manufacturing, while individual measurements are common in chemical processes
(temperature, pH) and transactional areas (financial). Subgroup size for discrete data should be
a minimum of 50.
Subgroup/Sample Size is constant: The number of observations within your sample remains
fixed over time.
Subgroup/Sample Size varies: The number of observations within your sample varies over
time.
451
SigmaXL: Control Phase Tools: SPC
Subgroup/Sample Size is very large and assumptions not met: This applies to discrete data
when the subgroup sizes are approximately 5,000 or higher and the control limits do not “look
right,” being very tight with many data points appearing to be out-of-control. This problem is
also referred to as “overdispersion.” This occurs when the assumption of a Binomial
distribution for defectives or Poisson distribution for defects is not valid. (Note: If the problem
of overdispersion is apparent with your continuous data, use SigmaXL > Control Charts >
Advanced Charts > I-MR-R or I-MR-S).
2. We would like to create a control chart of the Overall Satisfaction data. Since this can be
considered as continuous data, the data type is Continuous/Variable Data. The
subgroup/sample size is 1 (i.e. there is no subgrouping), so select Individuals (subgroup/sample
size = 1). At this point, we can choose Individuals or Individuals and Moving Range. We will
keep the simpler Individuals selection as shown. (Note that the above data types and
definitions can be viewed by clicking the Data Types and Definitions tab):
3. Click OK. This starts up the Individuals Chart dialog (see Part A – Individuals Charts for
continuation).
452
SigmaXL: Control Phase Tools: SPC
2. Click SigmaXL > Control Charts > Control Chart Selection Tool.
3. Since catapult shot distance measurement is continuous, we keep the default selection
Continuous/Variable Data. The catapult shot data are in subgroups, so select Subgroups
(subgroup/sample size > 1). The subgroup/sample size is small (3), so we will use the X-Bar &
Range Chart as shown:
4. Click OK. This starts up the X-Bar & Range dialog (see Part B – X-Bar & Range Charts for
continuation).
453
SigmaXL: Control Phase Tools: SPC
2. Click SigmaXL > Control Charts > Control Chart Selection Tool.
3. Since this data is discrete, select Discrete/Attribute Data. We are looking at Defectives data
since each loan is a pass or fail, so select Defectives (unit is good/bad, pass/fail). The
subgroup/sample size varies day to day so Subgroup/Sample Size varies is selected as shown.
The recommended chart is the P-Chart (proportion defective):
4. Click OK. This starts up the P-Chart dialog (see Part C – P-Charts for continuation).
454
SigmaXL: Control Phase Tools: SPC
I-MR-R Charts
If the within-subgroup variability is much smaller than between subgroup, the classical X-bar & R
(or S) chart will not work, producing numerous (false) alarms. The correct chart to use, in this case,
is the I-MR-R (or S) chart. The subgroup averages are treated as individual values (I-MR) and the
within subgroup ranges are plotted on the Range chart.
1. Open Multi-Vari Data.xlsx. Select Sheet Between. We saw this data previously using Multi-Vari
charts. First, we will incorrectly use the X-bar & R chart, and then apply the correct I-MR-R
chart.
2. Click SigmaXL > Control Charts > X-bar & R. Check Use Entire Data Table.
3. Click Next. Select Stacked Column Format. Select Measurement, click Numeric Data Variable
(Y) >>; select unit, click Subgroup Column or Size >>.
455
SigmaXL: Control Phase Tools: SPC
Almost all of the data points in the X-bar chart are out-of-control! This is due to the small
within-subgroup variability (the control limits are calculated from the within-subgroup
variability).
5. Click Sheet Between. Select SigmaXL > Control Charts > Advanced Charts > I-MR-R.
6. Click Next. Select Stacked Column Format. Select Measurement, click Numeric Data Variable
(Y) >>; select unit, click Subgroup Column or Size >>.
456
SigmaXL: Control Phase Tools: SPC
This chart is much cleaner, showing a stable Individuals and Range chart. The MR chart may be
trending up, but we would want to collect more data before making this conclusion. Typically,
you want at least 20 (30 preferred) subgroups before calculating final control limits.
457
SigmaXL Appendix
SigmaXL: Appendix
460
SigmaXL: Appendix
461
SigmaXL: Appendix
462
SigmaXL: Appendix
463
SigmaXL: Appendix
For sample sizes greater than 5000, Kolmogorov-Smirnov-Lilliefors (KSL) is used. Lilliefors [4]
modified the Kolmogorov-Smirnov test for the case of unknown mean and standard deviation.
This test is best suited for data with ties, i.e. “chunky” data. Anderson-Darling and Shapiro-Wilk are
severely affected by ties in the data and will trigger a low P-Value even if the data are normal.
A tolerance interval is the interval, described by a lower and upper bound, that a specified
minimum proportion of the sample’s values will lie with a given confidence probability. Thus, as an
example, the 95% tolerance interval for the range in which 80 percent of the sample will fall will
consist of a lower bound and an upper bound.
Notation:
464
SigmaXL: Appendix
v χ p2 ,1,λ
G ( z , n, p, k ) = 1 − Fv 2
k
where: χ p ,1,λ is the p percentile of the non-central χ 2 distribution with 1 degree of freedom and
2
non centrality parameter λ = z 2 /n , and where Fv is the cdf of the χ 2 distribution with v degrees
of freedom. Then k is the root such that:
∞
2∫ G ( z, n, p, k )φ ( z )dz = 1 − α
0
where φ (z ) is the pdf of the standard normal distribution. SigmaXL uses a 40 point Gauss-Legendre
quadrature to evaluate the integral. See Krishnamoorthy and Mathew [7]. The approximation
method by Howe [8] is used to obtain initial values.
465
SigmaXL: Appendix
Let r and s be the exact order statistics for confidence interval. Then
−1
r (1 − p )( Bn , p (r ) − α/2)
λl = 1 +
( n − r ) p (α /2 − B ( r − 1))
n, p
−1
s (1 − p)( Bn , p ( s ) − (1 − α/2))
λu = 1 +
( n − s ) p (1 − α /2 − B ( s − 1))
n, p
where Bn , p ( k ) is the cdf of the binomial distribution with parameters n and p . See Nyblom [9]
which is an extension of the interpolation method of Hettmansperger and Sheather [10] for median
using a double exponential distribution.
466
SigmaXL: Appendix
v1 = λxr + (1 − λ ) xr +1
v2 = λxs + (1 − λ ) xs −1
If xs − v1 < v2 − xr then L = v1 and U = xs ; else L = xr and U = v2 . See Young and Mathew [11].
467
SigmaXL: Appendix
The maximum likelihood estimates are then calculated using the Newton-Raphson method. This is
an iterative process that uses both the first and second derivatives to move to a point at which no
further improvement in the likelihood is possible.
The standard errors of the parameter estimates are derived from the Hessian matrix. This matrix,
which describes the curvature of a function, is the square matrix of second-order partial derivatives
of the function.
For some data sets, the likelihood function for threshold models is unbounded, and the maximum
likelihood methodology fails. In this context, a threshold is estimated using a bias correction
method. This is an iterative methodology that evaluates the threshold based on the difference
between the minimum value of the variate and the prediction for the minimum value, conditional
on the current values of the parameters.
468
SigmaXL: Appendix
Distributions
Beta Distribution
PDF
CDF
Range 0 ≤ x ≤ 1.
Γ(β) is the Gamma function and is described below under Gamma distribution.
469
SigmaXL: Appendix
CDF
Range 0 ≤ x ≤ 1.
Lower threshold θ 1
Upper threshold θ 2
Notes: Estimation of the 4 parameter Beta distribution is undertaken in two parts. In the first part,
initial parameter estimates are derived using the method of moments. The threshold parameters
are then held using these values, and the shape parameters are then estimated using maximum
likelihood.
Wang, J.Z. (2005). "A note on Estimation in the Four Parameter Beta Distribution", Comm in
Stats Simulation and computation, Vol. 34 pp. 495- 501.
470
SigmaXL: Appendix
CDF
Shape parameter λ.
Notes: The concentrated likelihood is used in the ML estimation. This implies that the location and
scale parameters are not estimated freely, but are derived as the mean and standard deviation of
the BoxCox transformed variate. The estimated parameters λ and θ are then used in the Box-Cox
(Power) transformation. See Transformations below.
471
SigmaXL: Appendix
Exponential Distribution
PDF
CDF
Range 0 ≤ x < ∞.
CDF
Range 0 ≤ x < ∞.
472
SigmaXL: Appendix
Gamma Distribution
PDF
CDF
Range 0 ≤ x < ∞.
473
SigmaXL: Appendix
CDF
Range 0 ≤ x - θ < ∞.
CDF
Range 0 ≤ x < ∞.
Chou, C., & H. Liu, (1998). "Properties of the half-normal distribution and its application to quality
control", Journal of Industrial Technology Vol. 14(3) pp. 4-7
474
SigmaXL: Appendix
CDF
Logistic Distribution
PDF
CDF
475
SigmaXL: Appendix
Loglogistic Distribution
PDF
CDF
CDF
476
SigmaXL: Appendix
Lognormal Distribution
PDF
CDF
CDF
477
SigmaXL: Appendix
Normal Distribution
PDF
CDF
Note: For consistency with other reports in SigmaXL such as Descriptive Statistics, the standard
deviation is estimated as the sample standard deviation using n-1 (rather than n).
478
SigmaXL: Appendix
CDF
Weibull Distribution
PDF
CDF
Range 0 ≤ x < ∞.
479
SigmaXL: Appendix
CDF
Range 0 ≤ x < ∞.
Lockhart, R.A. and M.A. Stephens (1994)."Estimation and Tests of Fit for the Three-parameter
Weibull Distribution", Journal of the Royal Statistical Society, Vol.56(3), pp. 491-500.
480
SigmaXL: Appendix
Transformations
Box-Cox (Power) Transformation
λ. ≠ 0
λ. = 0
Note: The optimum shape parameter,λ, is derived using a grid search in which the criteria function is the
standard deviation of the standardized transformed variable.
λ. ≠ 0
λ. = 0
481
SigmaXL: Appendix
Shape parameter λ.
Note: The parameters λ and θ are estimated using MLE as described above in Box-Cox Distribution
with Threshold.
Johnson Transformation
The Johnson Transformation selects one of the three families of distribution: SB (bounded), SL
(lognormal), and SU (unbounded) and the associated parameters so as to transform the data to be
normally distributed. The methodology follows Chou et al (1998) and uses the Anderson Darling P-
Value as the normality criteria.
Johnson Transformation - SB
z is N (0, 1)
482
SigmaXL: Appendix
Johnson Transformation - SL
z is N (0, 1)
Range x > ε
Johnson Transformation - SU
z is N (0, 1)
483
SigmaXL: Appendix
Since AD P-Values are not available for distributions with thresholds (other than Weibull), an
estimate is obtained by transforming the data to normality and then using a modified Anderson
Darling Normality test on the transformed data. The transformed z-values are obtained by using
the inverse cdf of the normal distribution on the cdf of the nonnormal distribution. The Anderson
Darling Normality test assumes a mean = 0 and standard deviation = 1.
This approach, unique to SigmaXL, is an extension of the Chou methodology used in Johnson
Transformations and allows a goodness of fit comparison across all distributions and
transformations.
Chou,Y., A.M. Polansky, and R.L. Mason (1998). "Transforming Nonnormal Data to Normality in
Statistical Process Control," Journal of Quality Technology, Vol. 30(2), pp. 133-141.
D'Agostino, R.B. and Stephens, M.A. (1986). Goodness-of-Fit Techniques, Marcel Dekker.
http://en.wikipedia.org/wiki/Akaike_information_criterion
484
SigmaXL: Appendix
The Z-Score method for computing process capability indices utilizes the same z-value
transformation to normality described above in Automatic Best Fit. The transformed z-values are
obtained by using the inverse cdf of the normal distribution on the cdf of the nonnormal
distribution. Normal based capability indices are then applied to the transformed z-values. This
approach offers two key advantages: the relationship between the capability indices and calculated
defects per million is consistent across the normal and all nonnormal distributions, and short term
capability indices Cp and Cpk can be estimated using the standard deviation from control chart
methods on the transformed z-values. The Z-Score method was initially developed by Davis Bothe
and expanded on by Andrew Sleeper. For further details, see Sleeper, Six Sigma Distribution
Modeling.
The Percentile method to calculate process capability indices uses the following formulas:
485
SigmaXL: Appendix
The Individuals – Original Data chart displays the untransformed data with control limits calculated
as:
CL = 50th percentile
The benefit of displaying this chart is that one can observe the original untransformed data. Since
the control limits are based on percentiles, this represents the overall, long term variation rather
than the typical short term variation. The limits will likely be nonsymmetrical.
The Individuals/Moving Range – Normalized Data chart displays the transformed z-values with
control limits calculated using the standard Shewhart formulas for Individuals and Moving Range
charts. The benefit of using this chart is that tests for special causes can be applied and the control
limits are based on short term variation. The disadvantage is that one is observing transformed
data on the chart rather than the original data.
SigmaXL’s default setting is to display the two charts: Individuals – Original Data, Individuals &
Moving Range – Normalized Data (with tests for special causes unchecked).
486
SigmaXL: Appendix
Introduction
Nonparametric tests are popular because they do not assume that the sample data are normally
distributed, but they do, however, assume that the test statistic follows a normal or chi-square
distribution when computing the P-Value using “large sample” or “asymptotic” theory. Similarly,
the Chi-Square test for independence used in r x c contingency tables assumes that the test statistic
is distributed as chi-square. When the sample size is small or sparse, these assumptions may be
invalid and can result in substantial error in the computed P-Value. For example, Cochrane’s rule
[5] for r x c contingency tables is a commonly used guideline: no more than 20% of the cells should
have an expected value less than 5 and none should have an expected value less than 1.
The solution to the small sample problem is to compute exact P-Values using the true permutation
distribution of the test statistic. However, explicit full enumeration of all possible outcomes quickly
becomes computationally infeasible. For example, a 5x6 contingency table results in 1.6 billion
possible outcomes.
Network algorithms by Mehta & Patel [6, 7] use techniques from operations research such as
backward induction and forward probing to implicitly and efficiently solve the exact P-Value for
contingency tables. These algorithms have also been extended to work with continuous data. For
example, the nonparametric Mann-Whitney can be converted to a weighted singly ordered
contingency table and the exact P-Value solved using the network algorithm.
Some data sets are too large even for the network algorithm to solve in a reasonable time period,
but still do not satisfy Cochrane’s rule. In these cases, Monte Carlo simulation is used to sample
from the full enumeration of possible outcomes. 10,000 replications are typically used and 99%
confidence intervals provided for the Monte Carlo P-Values. If the P-Value is low, 1 million
replications are recommended.
Full Enumeration
To deal with a situation where the size of the data set does not justify the large sample asymptotic
assumption, we will first consider full enumeration. To illustrate, consider the problem of
evaluating the probability of getting a total of 3 or less when tossing two dice.
Evaluated for each of the n = 36 possible outcomes the sum of those situations (m) for which the
sum or the two dice is 3 or less. Since m = 3 {1,1}, {1,2}, {2,1}, the probability is m/n = 3/36 =1/12.
487
SigmaXL: Appendix
Full enumeration is 100% accurate, and always produces the exact P-Value. It works fine provided
n is not too large. However, for 20 dice, n, the number of possible outcomes is 3656 trillion, which
makes full enumeration infeasible. As mentioned above, even a relatively small problem such as a
5x6 contingency table results in 1.6 billion possible outcomes, again making full enumeration
infeasible.
Exact
Instead of evaluating each of the n possible outcomes, shortcuts can be derived which reduces the
number of outcomes that need to be evaluated. Starting with the first dice showing unity, once the
second dice exceeds 2 the remaining four outcomes can be counted as a block, since the statistic
for each case will exceed the test statistic. Thus rather than evaluating each as one would with full
enumeration, the block is evaluated. When the first dice shows two, and once the second dice
exceeds 1, the remaining 5 outcomes can again be counted as a block. Once the first dice shows 3,
all the remaining outcomes can be evaluated as a block, since again every possible outcome will
then exceed 3. Hence instead of enumerating 36 outcomes, one enumerates 6 outcomes {1,1},
{1,2}, {1, 3+}. {2,1}, {2,2+}, {3+,1+}. This type of model is referred to as Enhanced Enumeration. It
uses full enumeration of all possible outcomes node by node, and evaluates the statistic as a block
if the largest statistic possible from the current position exceeds the critical value (in which case the
probability of this path is added to the total probability), or if the smallest statistic possible from
the current position is less than the critical value, implying no augmentation of the total probability
[15].
A second methodology involves creating a network of a set of nodes in a number of stages. The
distances between nodes are defined so that the total distance of a path through the network is
the corresponding value of the test statistic. At each node, the algorithm computes the shortest
and longest path distances for all the paths that pass through that node. This works easily for those
statistics that can be evaluated as the linear sum of partial statistics evaluated at each node. This
methodology is commonly referred to in the literature as the Network Model, after the pioneering
work of Mehta and Patel [1, 6, 7].
The Network Model has advantages over enhanced enumeration in that the problem is broken
down to nodes. For example, if we had 4 dice, then there would be 5 nodes. The advantage of the
network model is that the length (and hence probability) at each node for the remaining nodes can
be estimated using a backward step through the network. Since networks tend to be sparse, the
set of all paths through the network which have the same probability up to the current node can be
retained. Thus, for any set of paths which have the same length up to the current node, one can
quickly read off the longest and shortest remaining sub-paths to the terminal node. Specifically, as
in enhanced enumeration, if every possible outcome does not contribute to the P-Value, the entire
488
SigmaXL: Appendix
set can be eliminated, while if every possible outcome contributes, then the probability of the
entire set can be evaluated as a closed form.
Thus an exact evaluation of a probability is identical to full enumeration, but significantly faster.
Again, as the size of the problem increases, the exact method also becomes infeasible since the
number of outcomes that do have to be evaluated becomes too large. The methodology used to
evaluate each of the exact nonparametric tests is described below.
Monte Carlo
Simulate the throw of the two dice nrep times, and count the number of times (m) that the sum of
the two dice is 3 or less. The Monte Carlo P-Value is then defined as m/nrep; this is an estimate of
the exact P-Value, and can be made more accurate by increasing nrep. Thus, with 20 dice, one
could choose a value of nrep = 100,000, and carry out the simulation to ascertain an unbiased
estimate of the P-Value. Thus the beauty of Monte Carlo is that it can be used on any problem, no
matter how large the total possible outcomes; however it is only asymptotically exact.
summed over all possible combinations, where statcrit is the statistic evaluated for each possibility,
and mu is the expected value of the statistic. For symmetric distributions, 2*1-Sided and 2-Sided
will be identical. For non-symmetric distributions, such as the Runs Test and the Mann-Whitney
Test, there will be a difference between 2*1-Sided and 2-Sided.
489
SigmaXL: Appendix
The following sample size guidelines are based on our recommendations and those given in Siegel
[2] and Gibbons [3] for when exact nonparametric tests should be used rather than asymptotic
(Normal or Chi-Square approximation):
Sign Test: N <= 50 (exact is automatically utilized in the regular Sign Test)
These are minimum sample size guidelines. It is acceptable to use an exact test for larger sample
sizes, but computation time can become an issue especially for tests with two or more samples. In
those cases, one can always use a Monte Carlo P-Value with 99% confidence interval.
The maximum total sample size (total of all samples) that can be used for Nonparametric Exact or
Monte Carlo is 1000. The Runs Test for Randomness and Wilcoxon Test automatically switch to
large sample asymptotic if N > 1000.
Use Chi-Square Exact/Monte Carlo or Fisher’s Exact/Monte Carlo if more than 20% of the cells have
expected counts less than 5 or if any of the cells have an expected count less than 1 [4,5].
The maximum number of cells in a table is limited to 1000. For a One-Way Chi-Square the
maximum sum of cell counts is 1000. For a Contingency Table, the maximum sum of cell counts is
5000.
Power
It is important to note that while exact P-Values are “correct,” they do not increase (or decrease)
the power of a small sample test, so they are not a solution to the problem of failure to detect a
change due to inadequate sample size.
490
SigmaXL: Appendix
Monte Carlo P-Values are validated using 1e6 replications and compared against exact. Repeated
simulations are used to validate the confidence intervals.
1 Sample Tests
Sign Test: The number of sign changes in one direction is distributed binomial, and so the binomial
distribution function can be used directly to evaluate the tail probability [10].
Runs Test (Above/Below): Since the number of runs cannot exceed the length of the sample, it is
easy to evaluate the probability of r runs when there is a binary variable with n1 and n2 elements
[11].
Runs Test (Up/Down): Given the sample size and the number of up/down runs, the probability is
provided by published tables [12].
Wilcoxon Test: Given the signed rank statistic, look at all combinations of ranks that could
contribute to the observed value of the Wilcoxon critical value. For each of these cases, the
number of combinations that actually contribute is evaluated. These counts are then converted to
probabilities [13].
2 Sample Tests
Mann-Whitney Test: The continuous data is first converted to a weighted singly ordered
contingency table and the exact P-Value is solved using the network algorithm.
Each possible combination of ranks is systematically evaluated, but the process is cut short for any
given path if the largest statistic possible from the current position is less than the critical value (in
which case all paths from the present are added to the feasible combinations), or if the smallest
statistic possible from the current position is greater than the critical value (in which none of the
paths from the present are added to the feasible combinations).
K Sample Tests
Kruskal-Wallis Test: Full enumeration.
491
SigmaXL: Appendix
Mood’s Median Test: This test is evaluated using the Network Model. In each case, a chi-square
statistic is evaluated. The number of combinations at the end of each path that has a chi-square
statistic greater than the critical value is summed over all the enumerations, and is then converted
to a probability.
Contingency Tests
One Way Chi-Square Test: For a one dimension vector of frequencies, and expected probabilities,
the Chi Square test is evaluated using the Enhanced Enumeration model.
Two Way Fisher Exact Test: The exact probability for a two dimension r x c contingency table is
evaluated using the Network model [1]. This is a one tailed test, and includes the probabilities of
all the equivalent tables (those with same row and column counts) that have a lower or equal
probability as the sample table. As in the one dimensional case, tests are made to see if the
shortest path from the current position exceeds the path length of the observed table (in which
case the probability of all paths from the current position are added to the probability, or whether
the longest path from the current position is less than the path length of the current table, implying
no augmentation of the probability.
492
SigmaXL: Appendix
[9] Narayanan, A. and Watts, D. “Exact Methods in the NPAR1WAY Procedure,” SAS Institute
Inc., Cary, NC.
[10] Myles Hollander and Douglas A. Wolfe (1973), Nonparametric Statistical Methods. New
York: John Wiley & Sons.
[11] https://onlinecourses.science.psu.edu/stat414/node/330
[12] Eugene S. Edgington, (1961). “Probability Table for Number of Runs of Signs of First
Differences in Ordered Series,” Journal of the American Statistical Association, Vol. 56, pp.
156-159.
[13] David F. Bauer (1972), “Constructing confidence sets using rank statistics,” Journal of the
American Statistical Association Vol 67, pp. 687–690.
[14] Hongsuk Jorn & Jerome Klotz (2002). “Exact Distribution of the K Sample Mood and Brown
Median Test,” Journal of Nonparametric Statistics, 14:3, 249-257.
[15] Valz, P. D. and Thompson, M. E. (1994), “Exact Inference for Kendall’s S and Spearman’s Rho
with Extensions to Fisher’s Exact Test in rxc Contingency Tables,” Journal of Computational
and Graphical Statistics, Vol 3(4), pp. 459- 472.
493
SigmaXL: Appendix
The Hypothesis Test Assumptions Report is an optional text report for the 1 Sample t-Test, Paired t-
Test, 2 Sample t-Test, One-Way ANOVA and Welch’s ANOVA. The cells are color highlighted: Green
(OK), Yellow (Warning) and Red (Serious Violation).
Normality
Each sample is tested for Normality using the Anderson-Darling (AD) test. If the AD P-Value is less
than 0.05, the cell is highlighted as yellow (i.e., warning – proceed with caution). The Skewness and
Kurtosis are reported and a note added, “See robustness and outliers.” An example report is
shown:
If the AD P-Value is greater than or equal to 0.05, the cell is highlighted as green.
Robustness
A minimum sample size for robustness to nonnormality is determined using the regression
equations in Basic Statistical Templates – Minimum Sample Size for Robust t-Tests and ANOVA.
These were derived from extensive Monte Carlo simulations to determine a minimum sample size
required for a test to be robust, given a specified sample Skewness and Kurtosis.
When there are 2 or more samples, a minimum sample size is calculated for each sample, and the
largest value is (conservatively) considered as the minimum sample size required for the test.
494
SigmaXL: Appendix
If the AD Normality test for every sample is greater than or equal to 0.05, the following message is
shown for robustness:
If each sample size is greater than or equal to the minimum for robustness, the minimum sample
size value is reported and the test is considered to be robust to the degree of nonnormality present
in the sample data:
If any sample size is less than the minimum for robustness, the minimum sample size value is
reported and a suitable Nonparametric Test is recommended. The cell is highlighted in red:
1 Sample Sign
1 Sample Wilcoxon
The 1 Sample Wilcoxon is more powerful that the 1 Sample Sign, but assumes that the data are
symmetrical. Wilcoxon is therefore the default recommendation, but if the sample is moderately
skewed then the Sign test is recommended.
Robustness to symmetry equations were developed in a manner similar to those used for
robustness to nonnormality above. Monte Carlo simulations were run using sample sizes (N) of 20
to 2000. Observed alpha values were determined empirically from the P-Values of 100,000
replicate Wilcoxon tests. For a given sample size, nonnormal data with Skew = 0.01 to +1.0 was
generated using the DiscoverSim Pearson Family function (see SigmaXL’s DiscoverSim Workbook
for details). The population test median (H0) used for the Wilcoxon test was estimated empirically
by averaging the 100,000 sample medians. The test was considered robust when the simulated
495
SigmaXL: Appendix
observed alpha was between .04 and .06, .i.e. within +/- 20% of the specified alpha of .05.
“Maximum Skewness” occurred when the observed alpha exceeded .04 or .06.
Other alpha values were not considered. Power was not considered.
This simulation was performed for Ha = “Not Equal To” (NE), “Greater Than” (GT), and “Less Than”
(LT). A regression analysis was performed on the Log10 (Maximum Skewness) versus Log10(N),
resulting in the following equations (with R-Square reported):
Note that if Skewness is negative, equation 2 is used for LT and equation 3 is used for GT.
For example, given a sample size of 30 and Ha = “Not Equal To,” solving for Maximum Skewness NE
in equation 1, the value is 0.626. Therefore, if the sample Skewness is less than or equal to 0.626,
Wilcoxon is recommended. If the sample Skewness is greater than 0.626, the Sign Test is
recommended.
If the Sign Test is recommended, Exact is not explicitly recommended since it is already “built in”
the Sign Test for N <= 50.
If the Wilcoxon Test is appropriate and N <= 15, then Wilcoxon – Exact is recommended (using the
rules of thumb in Sample Sizes for Exact).
2 Sample Mann-Whitney
If each sample N <= 10, then Mann-Whitney – Exact is recommended (using the rules of thumb in
Sample Sizes for Exact).
The following Nonparametric Tests may be recommended for three or more samples:
Kruskal-Wallis
Kruskal-Wallis - Exact
496
SigmaXL: Appendix
Kruskal-Wallis is more powerful than Mood’s Median Test, but Mood’s Median is robust to outliers.
Kruskal-Wallis is therefore the default recommendation, but if any sample has outliers that are
Likely (2.2*IQR) or Extreme (3*IQR), then Mood’s Median is recommended. See Outliers (Boxplot
Rules) below for further details.
If Mood’s Median is appropriate and each sample N <= 10, then Mood’s Median – Exact is
recommended.
497
SigmaXL: Appendix
The Anderson Darling normality test is applied to the sample data with outliers excluded. If this
results in an AD P-Value that is greater than 0.1, a notice is given, “Excluding the outliers, data are
inherently normal." The cell remains highlighted as yellow or red.
Randomness (Independence)
Each sample is tested for randomness (serial independence) using the Exact Nonparametric Runs
Test. If the sample data is not random, a warning is given and recommendation to review the data
with a Run Chart or Control Chart. See Nonparametric Runs Test for Randomness - Exact for
further details. Note that for this report, the option Values Equal to Median is set to both Counted
as Below and Counted as Above, with the reported P-Value being the larger of the two. This
results in a more conservative test to minimize false alarms.
If the Exact Nonparametric Runs Test P-Value is greater than or equal to 0.05, the cell is highlighted
as green:
If the Exact Nonparametric Runs Test P-Value is less than 0.05, but greater than or equal to 0.01,
the cell is highlighted as yellow:
498
SigmaXL: Appendix
If the Exact Nonparametric Runs Test P-Value is less than 0.01, the cell is highlighted as red:
Equal Variance
The test for Equal Variances is applicable for two or more samples. If all sample data are normal,
the F-Test (2 sample) or Bartlett’s Test (3 or more samples) is utilized. If any samples are not
normal, i.e., have an AD P-Value < .05, Levene’s test is used. If the variances are unequal and the
test being used is the equal variance option, then a warning is given and Unequal Variance (2
sample) or Welch’s Test (3 or more samples) is recommended.
If the test for Equal Variances P-Value is >= .05, the cell is highlighted as green:
If the test for Equal Variances P-Value is >= .05, but the Assume Equal Variances is unchecked (2
sample) or Welch’s ANOVA (3 or more samples) is used, the cell is highlighted as yellow:
If the test for Equal Variances P-Value is < .05, and the Assume Equal Variances is checked (2
sample) or regular One-Way ANOVA (3 or more samples) is used, the cell is highlighted as red:
499
SigmaXL: Appendix
If the test for Equal Variances P-Value is < .05, and the Assume Equal Variances is unchecked (2
sample) or Welch’s ANOVA (3 or more samples) is used, the cell is highlighted as green:
500
SigmaXL: Appendix
Kappa
Kappa ( κ ) measures the degree of agreement between appraisers when they asses the same
sample. Cohen’s Kappa, used in the Attribute MSA template, is only defined for two raters. Fleiss’
Kappa can be used for multiple assessors [4]. Thus Kappa provides a mechanism for evaluating the
consistency of assessors.
Let P be the measure of agreement between the assessors, and Pe be the measure of agreement
that would occur by chance. Then:
P − Pe
κ=
1 − Pe
The denominator provides the measure of agreement above chance, while the numerator provides
the actual measure of agreement above chance. Thus κ is constrained between -1 and +1, with +1
implying complete consistency or perfect agreement between assessors, zero implying no more
consistency between assessors than would be expected by chance and -1 implying perfect
disagreement.
Kappa: >= 0.75 or so signifies excellent agreement, for most purposes, and <= 0.40 or so
signifies poor agreement.
In Six Sigma process improvement applications, a more rigorous level of agreement is commonly
used. Futrell [6] recommends:
501
SigmaXL: Appendix
The lower limit for an acceptable Kappa value (or any other reliability coefficient) varies
depending on many factors, but as a general rule, if it is lower than 0.7, the measurement
system needs attention. The problems are almost always caused by either an ambiguous
operational definition or a poorly trained rater. Reliability coefficients above 0.9 are
considered excellent, and there is rarely a need to try to improve beyond this level.
SigmaXL uses the guidelines given by Futrell and color codes Kappa as follows: >= 0.9 is green, 0.7
to < 0.9 is yellow and < 0.7 is red. This is supported by the following relationship to Spearman Rank
correlation and Percent Effectiveness/Agreement (applicable when the response is binary with an
equal proportion of good and bad parts):
Kappa = 0.7; Spearman Rank Correlation = 0.7; Percent Effectiveness = 85%; Percent
Agreement = 85% (two trials)
Kappa = 0.9; Spearman Rank Correlation = 0.9; Percent Effectiveness = 95%; Percent
Agreement = 95% (two trials)
(Note that these relationships do not hold if there are more than two response levels or the
reference proportion is different than 0.5).
Concordance
Let two assessors rank n subjects, such that x is the rankings of assessor 1, and y is the rankings
assessor 2. If for any pair of observations (i, j ) , the ranking of the two assessors agree - that is
xi > x j and yi > y j , or xi < x j and yi < y j , then the pair of observations are said to be
concordant. Pairs that are not concordant are said to be discordant. If x were a measurement of
exercise, and y a measurement of fitness, then we would expect high levels of concordance.
(k − 1) ρ s + 1
W =
k
where k is the number of trials (within) or trials*appraisers (between) among which Spearman’s
correlation coefficients are computed.
502
SigmaXL: Appendix
Confidence limits for Kendall’s Concordance cannot be solved analytically, so are estimated using
bootstrapping. Ruscio [8] demonstrates the bootstrap for Spearman’s correlation and we apply
this method to Kendall’s Concordance. The data are row wise randomly sampled with replacement
to provide the bootstrap sample (N = 2000). W can be derived immediately from the mean value of
the Spearman’s correlation matrix from the bootstrap sample. The bias corrected values of α and
1 − α are then used to determine the confidence interval from the sorted bootstrap vector W. In
Efron’s bias-corrected and accelerated bootstrap, the value of α is adjusted to account for skew
and/or bias. This has been determined to be a superior bootstrap method compared to the simple
percentile method [9]. A fixed seed is used so that the report results are consistent.
SigmaXL uses the following “rule-of-thumb” interpretation guidelines: >= 0.9 very good agreement
(color coded green); 0.7 to < 0.9 marginally acceptable, improvement should be considered
(yellow); and < 0.7 unacceptable (red).
This is consistent with Kappa and is supported by the relationship to Spearman’s correlation.
Note however that in the case of Within Appraiser agreement with only two trials, then the rules
should be adjusted: very good agreement is >= 0.95 and unacceptable agreement is < 0.85.
Let two assessors rank n subjects, such that x is the rankings of assessor 1, and y is the rankings of
assessor 2. If for any pair of observations (i, j ) , the ranking of the two assessors agree - that is
xi > x j and yi > y j , or xi < x j and yi < y j , then the pair of observations are said to be
concordant. Pairs that are not concordant are said to be discordant. If x were a measurement of
exercise, and y a measurement of fitness, then we would expect high levels of concordance.
Define:
503
SigmaXL: Appendix
Then the number of concordant ( nc ) and discordant ( nd ) pairs are derived as:
nc = 0.5P where P = ∑xij Cij
i, j
Dc = n 2 − ∑C 2j
j
504
SigmaXL: Appendix
These values were determined using Monte Carlo simulation with correlated integer uniform
distributions. They correspond approximately to Spearman 0.7 and 0.9 when there are 5 ordinal
response levels (1 to 5). With 3 response levels, the rule should be modified to 0.65 and 0.9.
505
SigmaXL: Appendix
This section describes a number of tests and measures of association that are derived from a
contingency table. These tables depict the observed counts of two variables, shown in r rows and
c columns, with a row sum vector R and a column sum vector C . The total number of cells is n .
In a contingency table, xij represents the number of elements that correspond to row attribute i
and column attribute j .
aij = sij / (1 − ni /n)(1 − n j /n) , the adjusted (standardized) residual expressed as a z score.
506
SigmaXL: Appendix
McNemar-Bowker Symmetry
For square tables, the McNemar statistic [11] tests the hypothesis that the contingency table is
symmetric. It is defined as:
( xij − x ji ) 2
M =∑
i< j xij + x ji
for all xij + x ji > 0 . Under the null, M is distributed χ 2 with n(n − 1)/2 degrees of freedom.
Perason’s Phi
The Phi ( φ ) statistic [1,9] is defined as:
φ = χ 2 /n
where χ 2 is the Pearson Chi-squared statistic. For a 2x2 table, φ is a measure of association for two
binary variables, and equivalent to Pearson Correlation, measures the percent of concentration of
cases on the diagonal. See Cohen [15} for interpretation rules-of-thumb.
Cramer’s V
The Cramer’s V statistic [3, 9] measures the association between two variables as a percentage of
their maximum possible variation. It has a range of 0 ≤ V ≤ 1 . It is defined as:
V= χ 2 /(n(m − 1))
where χ 2 is the Pearson Chi-squared statistic, and m = min (r , c)
Contingency Coefficient
The contingency statistic [3] is defined as:
z = χ 2 /(n + χ 2 )
where χ 2 is the Pearson Chi-squared statistic. Unlike the other statistics, it does not have a
specified range.
507
SigmaXL: Appendix
Cohen’s Kappa
For square tables, Cohen’s κ statistic [3,5] tests the null hypothesis that there is no association
between the row and column variables. It is calculated from the observed and expected
frequencies on the diagonal of a square contingency table. It is frequently used to provide a
mechanism for evaluating the consistency of assessors. It is defined as:
n∑xii − ∑Ri Ci
κ= i i
n 2 − ∑Ri Ci
i
Goodman-Kruskal Lambda
Lambda [3] is defined as the proportional reduction in error of prediction of the dependent variable
as a consequence of using the information in the independent variable.
Goodman-Kruskal Tau
Tau [9] is the same as lambda, except the probabilities are specified by marginal proportions.
508
SigmaXL: Appendix
n∑xij2 /C j − ∑Ri2
i, j i
τc =
n − ∑R
2
i
2
Thiel’s Uncertainty
Uncertainty [12] is defined as the proportional reduction in the uncertainty (entropy) of the
dependent variable as a consequence of using the information in the independent variable.
where:
U r = −∑Ri /n ln ( Ri /n)
i
U c = −∑C j /n ln (C j /n)
j
509
SigmaXL: Appendix
Define:
Cij = ∑∑xhk + ∑∑xhk
h <i k < j h >i k > j
Then the number of concordant ( nc ) and discordant ( nd ) pairs are derived as:
nc = 0.5 P where P = ∑xij Cij
i, j
The probability value returned for this test is given in Kendall’s Tau-b below.
ρ n−2
t=
1− ρ 2
510
SigmaXL: Appendix
Kendall’s Tau-b
The Kendall τ b coefficient [10], like gamma, is a non-parametric measure of association using rank
correlation, but makes a correction for ties. τ b is defined as:
2(nc − nd )
τb =
Dr Dc
where:
Dr = n 2 − ∑Ri2
i
Dc = n 2 − ∑C 2j
j
∑x (C
i, j
ij ij − Dij ) 2 − ( P − Q) 2 /n
s=2 under the null hypothesis,
Dr Dc
then τ b /s is distributed standard normal. This is used to provide a P-Value for the Concordant –
Discordant test.
Kendall-Stuart Tau-c
Stuart’s τ c coefficient [1] is a measure of association, but makes a correction for ties as well as table
size. τ c is defined as:
2m(nc − nd )
τc =
n 2 (m − 1)
where m = min (r , c) . τ c ranges from +1 to -1. A value of zero implies independence.
Goodman-Kruskal Gamma
The Goodman-Kruskal’s gamma statistic [8] provides a measure of association using rank
correlation, and is based on the number of concordant and discordant pairs. It is defined as:
nc − nd
γ=
nc + nd
Somer’s D
Somer’s D statistic [1,14] is also a measure of association, similar to τ b , but differs in respect to a
consideration as to whether the row variable is considered the independent variable (and the
column variable dependent), or vice versa.
Column as dependent, row as independent:
511
SigmaXL: Appendix
S r = 2( nc − nd )/Dr
Row as dependent, column as independent:
S c = 2( nc − nd )/Dc
Symmetric version:
S s = 4( nc − nd )/( Dr + Dc )
512
SigmaXL: Appendix
[15] Cohen, J. (1988), Statistical Power Analysis for the Behavioral Sciences (Second Ed.),
Lawrence Erlbaum Associates, pp. 215 - 227.
513
SigmaXL: Appendix
This set of tests is undertaken after an ANOVA or Equal Variance test, and are often referred to as
“post hoc tests”. After the null of equal means or variances has been rejected, the following tests
are computed for each of the .5k (k − 1) sample pairs. Thus this permits the researcher to ascertain
exactly which group(s) are responsible for the rejection of the null.
One-Way ANOVA
Fisher
Fisher’s least significant difference test is a set of individual t-tests on each of the sample pairs, but
uses the pooled standard deviation from all groups, which increases the power. The test statistic is:
yi − y j
t=
s 1/ni + 1/n j
where s is the pooled standard error for all groups, and is assumed distributed Student t.
There is no adjustment for family-wise error rate, so should only be used for k = 3.
Tukey
Tukey’s test is a set of individual t-tests on each of the sample pairs, but uses the pooled standard
deviation from all group. The test statistic is:
2 ( yi − y j )
t=
s 1/ni + 1/n j
Dunnett
The Dunnett test compares each group to a given control group. For each group, a t statistic is
calculated using the Fisher statistic. The distribution of this statistic is a k − 1 multivariate t, with all
k − 1 arguments being the same Fisher t statistic, and with correlation matrix with off diagonal
elements 2 for the balanced case, and weighted by sample size for the unbalanced case [1].
514
SigmaXL: Appendix
This is the pairwise two sample t-test for unequal variances. Like Fisher, there is no adjustment for
family-wise error rate so should only be used for k = 3.
Games-Howell
The Games-Howell test does not assume equal variance for each group. The variance of each group
is used. The test statistic is:
2 | yi − y j |
t=
si2 /ni + s 2j /n j
The standard F test statistic for testing the null of equal variances of two groups is:
si2
F=
s 2j
2
where si is the variance of group i , and is distributed as Fn −1,n .
i j −1
There is no adjustment for family-wise error rate so should only be used for k = 3.
The pairwise Levene test statistic for testing the null of equal variances of two groups is a simple F
statistic:
515
SigmaXL: Appendix
SSA/(k − 1)
F=
SSE/(n − k )
where n = ni + n j , and SSA and SSE are the sample sum of squares and the error sum of squares of
the transformed variable respectively:
z =| x − median( x) |
There is no adjustment for family-wise error rate, so should only be used for k = 3.
Tukey’s method is applied to the ADM (Absolute Deviations from Median). Conceptually, this is like
an extension of the ANOMV-LEV chart, Analysis of Means version of the Levene Test. See [5]
Nelson, Wludyka and Copeland (2005), page 65 for discussion on robustness.
516
SigmaXL: Appendix
A comparison of treatment means to determine which are significantly different from the overall
average is commonly required. For example an educational professional might ask which schools in
the system have a higher or lower rate of graduation compared to the average for all schools.
ANOM (Analysis of Means) addresses this problem for a fixed effects model. ANOM identifies the
means that are significantly different from the overall mean. Analysis of means is often used in
process control, since it has a graphical representation that is similar to a Shewhart chart. Typically,
a central line represents the overall average. The treatment means, plotted as deviations from the
overall average are compared with upper and lower decision limits to identify which are
significantly different from the overall mean. The underlying assumptions are that the response
belongs to a given distribution, and that the design is one way or two way.
The lower and upper decision levels (LDL, UDL) are used to ascertain whether the means at the
factor level are different from the grand mean. A factor level mean that lies outside the UDL and
LDL is statistically different from the grand mean at the specified confidence level.
Let n = ∑ jn j is the total number of observations, xij be the i th observation at factor level j , and
the mean for factor level j be:
nj
∑x ij
xj = i =1
nj
The grand mean is:
∑x
i, j
ij
x=
n
UDL j = x + hα s p w j
517
SigmaXL: Appendix
k
s 2p = ∑(n j − 1) s 2j /(n − k )
j =1
2
where s j is the variance of factor j .
w j = (n − n j )/(nn j )
where hα is the absolute value of the inverse CDF of the students t distribution at probability α *
with df = n − k degrees of freedom, or is determined from lookup tables [4] (Table B.1 for
balanced, Table B.3 for unbalanced) at the given probability level ( α ),
and where α * = (1 − (1 − α )1/k )/2 if k > 2 and α * = α/2 if k = 2 .
UDL fj = x + hα MSE w fj
where LDL fj and UDL fj are the lower/upper bounds for factor f at level j respectively,
where MSE is derived from an ANOVA analysis with a two factor model with cross effects,
where the weighting factor for each group j is:
w fj = (k f − 1)/(k f n j )
which reduces to w = (k − 1)/n in the balanced case.
where hα is the absolute value of the inverse CDF of the students t distribution at probability α *
with df = k1k 2 (n − 1) degrees of freedom, or is determined from Nelson’s lookup tables at the
given probability level ( α ),
1/k f
and where α * = (1 − (1 − α ) )/2 if k > 2 and α * = α/2 if k = 2 .
518
SigmaXL: Appendix
BY method
A straight forward method of analyzing a two factor model with k1 and k 2 groups for factor 1 and
2 respectively is to simply evaluate the one factor model with factor 1 for each value of factor 2.
Thus a total of k 2 one way models runs are undertaken.
SLICE method
This method [Wludyka 2013, 2015] is the same as the BY method, but uses the pooled MSE and
degrees of freedom from the full two factor model. The SLICE factor with minimum number of
levels is automatically selected.
A binomial distribution can be used for data that consists of proportions. Thus p j = ∑i j xij /n j .
n
For a single factor with k factor levels, and N = ∑ jn j observations, and with the response
assumed binomial, the LDL and UDL are given by:
LDL j = p − hα s p w j
UDL j = p + hα s p w j
519
SigmaXL: Appendix
and where is the absolute value of the inverse CDF of the normal distribution at probability
hα
α = 1 − α/(2k ) if k > 2 , and α = 1 − α/2 if k = 2 , or is determined from Nelson’s lookup tables at
* *
UDL = p + hα s p w
and where is the absolute value of the inverse CDF of the normal distribution at probability
hα
α = 1 − α/(2k ) if k > 2 , and α = 1 − α/2 if k1 = 2 , or is determined from Nelson’s lookup tables at
* *
The significance of the interaction effect is evaluated using a likelihood ratio test:
q = −2(llfu − llf r )
520
SigmaXL: Appendix
where n is the vector of sample size, p is the vector of probabilities, and p̂ is the predicted
probability given by exp( xβ )/(1 + exp( xβ )) . x is the factor matrix, and β is the maximum likelihood
coefficient vector.
q is distributed χ 2 with degrees of freedom equal to the number of restrictions.
For a single factor with k factor levels, and N = ∑ jn j observations, and with the response
assumed distributed Poisson, the LDL and UDL are given by:
LDL j = c − hα s p w j
UDL j = c + hα s p w j
c=
N
and where hα is the absolute value of the inverse CDF of the normal distribution at probability
α = 1 − α/(2k ) if k > 2 , and α = 1 − α/2 if k = 2 , or is determined from Nelson’s lookup tables at
* *
521
SigmaXL: Appendix
The significance of the interaction effect is evaluated using a likelihood ratio test:
q = −2(llfu − llf r )
The rank ( r ) of each element in the data set is derived. Then the rank is transformed using inverse
normal scores:
0.5 + r
x = Φ −1 ( )
2N −1
where N is the number of observations, and Φ −1 is the inverse of the standard normal
distribution function. x is then distributed normally, and the usual ANOM procedure can be
applied.
ANOM Variances
ANOMV can be used to test whether the variance for any sample is significantly different from the
pooled varaiance - it is thus a test of homogeneity. Typically, consistent output is required from a
set of production lines, and ANOMV can be used to identify which line has variability significantly
different from the average.
Let N = ∑ jn j be the total number of observations, xij be the i th observation at factor level j , and
2
the sample variance for factor level j be s j .
522
SigmaXL: Appendix
CL = MS e
UDL = U α ,k ,n −1k MSe
and where Lα ,k ,n −1 and U α ,k ,n −1 are determined from Nelson’s lookup tables at the given
probability level ( α ) and k and n − 1 degrees of freedom.
CL j = MSe
N −k
UDL j = U α* ,k ,n MS e
j n j −1
∑( n
j =1
j − 1) s 2j
MS e =
N −k
L*α ,k ,n = β −1 (α * , θ j , φ j )
j
U α* ,k ,n = β −1 (1 − α * , θ j , φ j )
j
and where α * = (1 − α )/(2k ) , θ j = (n j − 1)/2 , φ j = ∑i ≠ jθ i , and β −1 is the inverse of the cdf of the
beta function.
523
SigmaXL: Appendix
CL = MS e
UDL = MSe + hα ,k ,∞ MSe w
where hα ,k ,∞ is determined from Nelson’s lookup tables at the given probability level α and k
and ∞ degrees of freedom.
and where γ is an estimate of the common kurtosis - for samples distributed normal, this will be
zero..
A similar formula is used for the large sample unbalanced case - see [4] for details.
524
SigmaXL: Appendix
[4] Nelson, P.R., Wludyka, P.S. and Karen A. F. Copeland, K.A. (2005) The Analysis of Means: A
Graphical Method for Comparing Means, Rates, and Proportions ASA-SIAM Series on
Statistics and Applied Probability, SIAM, Philadelphia, ASA, Alexandria, VA.
[5] Wludyka, P.S., (2013). “Analyzing Multiway Models with ANOM Slicing” SESUG Working
Paper SD-06 21st Southeast SAS users Group Conference.
[6] Wludyka, P.S., (2015). “Using ANOM Slicing for Multi-Way Models with Significant
Interaction”, Journal of Quality Technology 47(2), pp. 193-203.
[7] Wludyka, P.S. and Noguera, J.G. (2016, November). “Using Anom Slicing For Multiway
Models With Binomial Or Poisson Data,” Poster session presented at INFORMS Annual
Meeting, Nashville, TN.
[8] Wludyka, P.S., Bergen, A. & Liu, X. (2002) “A Monte Carlo Study of Two Normal Based
Homogeneity of Variance Tests for Unbalanced Designs”, Joint Statistical Meetings -
Statistical Computing Section, pp. 3758-3759.
[9] Wludyka, P.S. & Nelson, P.R. (1997). “An Analysis-of-Means-Type Test for Variances from
Normal Populations”, Technometrics, 39(3), pp. 274-285
[10] Wludyka, P.S. & Nelson, P.R. (1999) “Two non-parametric, analysis-of-means-type tests for
homogeneity of variances”, Journal of Applied Statistics, 26:2, pp. 243-256,
525
About SigmaXL, Inc.
Established in 1998, SigmaXL Inc. is a leading provider of user friendly Excel Add-ins for Lean Six
Sigma graphical and statistical tools and Monte Carlo simulation.
Our flagship product, SigmaXL®, was designed from the ground up to be a cost-effective, powerful,
but easy to use tool that enables users to measure, analyze, improve and control their service,
transactional, and manufacturing processes. As an add-in to the already familiar Microsoft Excel,
SigmaXL® is ideal for Lean Six Sigma training and application, or use in a college statistics course.
DiscoverSim™ enables you to quantify your risk through Monte Carlo simulation and minimize your
risk with global optimization. Business decisions are often based on assumptions with a single point
value estimate or an average, resulting in unexpected outcomes. DiscoverSim™ allows you to
model the uncertainty in your inputs so that you know what to expect in your outputs.
Contact Information:
1-519-579-5877
527