Untitled
Untitled
Untitled
by
Joachim Melcher
Dissertation, Karlsruher Institut für Technologie
Fakultät für Wirtschaftswissenschaften
Tag der mündlichen Prüfung: 8. Februar 2011
Referenten: Prof. Dr. Detlef Seese, Prof. Dr. Andreas Oberweis
Impressum
ISBN 978-3-86644-789-9
Process Measurement in
Business Process Management
Theoretical Framework
and
Analysis of Several Aspects
genehmigte
DISSERTATION
von
This thesis was written during my occupation as research assistant at the Institut
für Angewandte Informatik und Formale Beschreibungsverfahren (Institute for Applied
Informatics and Formal Description Methods) at the Universität Karlsruhe (TH)
and the Karlsruher Institut für Technologie (KIT) (Karlsruhe Institute of Technology
[KIT])1 . Many people have contributed to the success of the thesis.
First of all, I want to thank Professor Detlef Seese for the supervision of this
thesis and his willingness to let me choose this topic. So, I was able to transfer
knowledge acquired during my studies of computer science at the Universität
Karlsruhe (TH) in the area of empirical software engineering (Professor Walter F.
Tichy).
Furthermore, I am indebted to the following persons and organizations for
their contribution to the thesis:
Agnes Koschmider for the possibility to conduct experiment 1 in the “Workflow
Management” lecture (Chapter 6), Markus Kress for many fruitful discussions
about different aspects of the thesis, Roland Küstermann for his technical assis-
tance with Apache Tomcat (Chapter 7), Christine Melcher for procuring some
publications from libraries in Heidelberg and Mannheim, Jan Mendling for his
cooperation in experiment 2 (Chapter 6), Sanaz Mostaghim for her fruitful re-
marks on heatmaps (Chapter 5), Kerstin Schmidt for her technical assistance
with the online questionnaire for experiment 2 (Chapter 6), Oliver Schöll for
his administrative support, Hajo A. Reijers for his cooperation in experiment 2
(Chapter 6), Jana Weiner for her administrative support, the students from the
Humboldt-Universität zu Berlin, the Technische Universiteit Eindhoven and the Univer-
sität Karlsruhe (TH) who participated in the experiments (Chapters 6 and 7), the
anonymous reviewers for their fruitful remarks on the papers which this thesis is
partially based upon and—finally—the Deutsche Forschungsgemeinschaft (German
Research Foundation) for the financial support of my SYNASC 2008 conference
trip to Timişoara (Romania).
Additionally, my thanks go to all my colleagues from the research group (Časlav
Božić, Hagen Buchwald, Tobias Dietrich, Matthes Elstermann, Jörg Janning,
Markus Kress, Roland Küstermann, Andreas Mitschele, Amir Safari, Oliver Schöll
and Andreas Vogel) for the pleasant and cooperative working atmosphere, our
weekly lunches and the exchange of experiences and information about writing a
PhD thesis.
I also want to thank the colleagues from the other research groups of our
institute, our administrative and technical staff, our tutors, Kerstin Schmidt as
well as Frederic Toussaint and the whole CIP pool team for their support in the
daily work as research assistant.
1 The KIT was founded on October 1, 2009 by a merger of the Universität Karlsruhe (TH) and the
Forschungszentrum Karlsruhe (Research Center Karlsruhe).
ix
Furthermore, I want to express my gratitude to the members of the university’s
debating club Debatte Karlsruhe e. V. and judo club. They offered me the possibility
for intellectual and physical diversion from the work on this thesis.
Last but not least, I want to thank my sister Christine Melcher for proof-reading
the manuscript of this thesis. However, for any errors which may remain in this
work the responsibility is entirely my own.
x
CONTENTS
1 introduction 1
1.1 Motivation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.2 Objective and Contribution . . . . . . . . . . . . . . . . . . . . . . . 1
1.3 Outline . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
1.4 Previous Publications . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
2 basics of business process management 7
2.1 Business Process Management . . . . . . . . . . . . . . . . . . . . . 7
2.1.1 Business Processes . . . . . . . . . . . . . . . . . . . . . . . . 7
2.1.2 What Is Business Process Management? . . . . . . . . . . . . 10
2.1.3 Business Process Management in Practice . . . . . . . . . . . 14
2.2 Workflow Management . . . . . . . . . . . . . . . . . . . . . . . . . . 16
2.2.1 Workflow Management and Terminology . . . . . . . . . . . 16
2.2.2 Workflow Management Systems . . . . . . . . . . . . . . . . 18
2.3 Process Modeling Languages . . . . . . . . . . . . . . . . . . . . . . 18
2.3.1 Event-Driven Process Chains . . . . . . . . . . . . . . . . . . 20
2.3.2 Product Data Models . . . . . . . . . . . . . . . . . . . . . . . 21
3 process measurement 27
3.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27
3.2 Related Work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28
3.3 Discussion of Related Work . . . . . . . . . . . . . . . . . . . . . . . 30
3.3.1 Process Model Complexity . . . . . . . . . . . . . . . . . . . 30
3.3.2 Process Model Quality and Performance . . . . . . . . . . . 43
3.4 Process Measurement Approach . . . . . . . . . . . . . . . . . . . . 44
3.4.1 Measurement and Prediction Systems . . . . . . . . . . . . . 44
3.4.2 Process Measurement Approach—An Adaptation from
Software Measurement . . . . . . . . . . . . . . . . . . . . . . 45
3.4.3 Validation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47
3.5 Application of Process Measurement . . . . . . . . . . . . . . . . . . 51
3.5.1 Selection of Metrics and Measures . . . . . . . . . . . . . . . 51
3.5.2 Different Measurement Purposes . . . . . . . . . . . . . . . . 51
3.6 Assessment of Existing Work . . . . . . . . . . . . . . . . . . . . . . 53
3.7 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 55
4 analysis of process model metric properties 57
4.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57
4.2 Approach for the Analysis of Process Model Metric Properties . . . 58
4.2.1 General Properties . . . . . . . . . . . . . . . . . . . . . . . . 58
4.2.2 Process Model Collection Specific Properties . . . . . . . . . 59
xi
Experimental Application . . . . . . . . . . . . . . . . . . . . . . .
4.3 . 61
4.3.1 Selected Process Model Metrics and Process Model
Collection . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 61
4.3.2 Results Concerning General Properties . . . . . . . . . . . . 62
4.3.3 Results Concerning Process Model Collection Specific
Properties . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 80
4.4 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 94
5 visualization and clustering of process model
collections 97
5.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 97
5.2 Visualization of High-Dimensional Data . . . . . . . . . . . . . . . . 97
5.2.1 Inadequate Visualization Techniques . . . . . . . . . . . . . . 98
5.2.2 Heatmaps . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 100
5.2.3 Principal Component Analysis Visualization . . . . . . . . . 100
5.3 Clustering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 101
5.3.1 Basics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 101
5.3.2 Hierarchical Clustering . . . . . . . . . . . . . . . . . . . . . . 102
5.3.3 Partitive Clustering: k-means . . . . . . . . . . . . . . . . . . 103
5.4 Approach for Visualization and Clustering of Process Model
Collections . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 105
5.4.1 Heatmap Visualization . . . . . . . . . . . . . . . . . . . . . . 105
5.4.2 Principal Component Analysis Visualization . . . . . . . . . 105
5.4.3 Clustering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 105
5.5 Experimental Application . . . . . . . . . . . . . . . . . . . . . . . . 105
5.5.1 Selected Process Model Metrics and Process Model
Collection . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 105
5.5.2 Results Concerning Heatmap Visualization . . . . . . . . . . 106
5.5.3 Results Concerning Principal Component Analysis
Visualization . . . . . . . . . . . . . . . . . . . . . . . . . . . . 107
5.5.4 Results Concerning Clustering . . . . . . . . . . . . . . . . . 108
5.6 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 109
6 measuring structural process model understandability 113
6.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 113
6.2 Related Work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 114
6.2.1 Existing Approaches . . . . . . . . . . . . . . . . . . . . . . . 115
6.2.2 Criticism on Existing Approaches . . . . . . . . . . . . . . . 116
6.3 Framework for Evaluating Modeling Technique Understanding . . 117
6.4 Approach for Measuring Structural Process Model
Understandability . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 118
6.4.1 Aspects of Structural Process Model Understandability . . . 119
6.4.2 Structural Process Model Understandability . . . . . . . . . 121
6.4.3 Partial Structural Process Model Understandability . . . . . 123
6.4.4 Virtual Subjects’ Structural Process Model
Understandability . . . . . . . . . . . . . . . . . . . . . . . . . 124
xii
6.5 Experimental Evaluation . . . . . . . . . . . . . . . . . . . . . . . . . 125
6.5.1 Experiment 1 . . . . . . . . . . . . . . . . . . . . . . . . . . . 125
6.5.2 Experiment 2 . . . . . . . . . . . . . . . . . . . . . . . . . . . 135
6.6 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 159
7 effects of process model granularity 165
7.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 165
7.2 Process Model Granularity Heuristic . . . . . . . . . . . . . . . . . . 166
7.2.1 Process Model Granularity Metric . . . . . . . . . . . . . . . 166
7.2.2 Process Model Granularity Heuristic . . . . . . . . . . . . . . 168
7.3 Experimentation System . . . . . . . . . . . . . . . . . . . . . . . . . 169
7.4 Experimental Evaluation . . . . . . . . . . . . . . . . . . . . . . . . . 170
7.4.1 Experiment Design . . . . . . . . . . . . . . . . . . . . . . . . 171
7.4.2 Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 176
7.4.3 Alternative Error Probability Model . . . . . . . . . . . . . . 178
7.4.4 Validity Evaluation . . . . . . . . . . . . . . . . . . . . . . . . 181
7.5 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 182
8 conclusion and outlook 183
8.1 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 183
8.2 Outlook . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 187
a measurement fundamentals 189
a.1 Definitions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 189
a.2 Hierarchy of scale types . . . . . . . . . . . . . . . . . . . . . . . . . 192
a.3 Measurement of Non-Physical Properties . . . . . . . . . . . . . . . 196
a.3.1 Conceptualization, Operationalization and Measurement . . 196
a.3.2 Criteria of Measurement Quality . . . . . . . . . . . . . . . . 198
b basics of empirical research 201
b.1 Motivation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 201
b.2 Empirical Approaches . . . . . . . . . . . . . . . . . . . . . . . . . . 202
b.2.1 Surveys . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 202
b.2.2 Case Studies . . . . . . . . . . . . . . . . . . . . . . . . . . . . 202
b.2.3 Controlled Experiments . . . . . . . . . . . . . . . . . . . . . 203
b.2.4 Comparison of the Approaches . . . . . . . . . . . . . . . . . 203
b.3 Controlled Experiments . . . . . . . . . . . . . . . . . . . . . . . . . 205
b.3.1 About Controlled Experiments . . . . . . . . . . . . . . . . . 205
b.3.2 Basic Terminology . . . . . . . . . . . . . . . . . . . . . . . . 206
b.3.3 Experiment Process . . . . . . . . . . . . . . . . . . . . . . . . 207
c measuring correlations 219
c.1 Pearson’s Product-Moment Correlation Coefficient . . . . . . . . . 219
c.2 Spearman’s Rank Correlation Coefficient . . . . . . . . . . . . . . . 225
bibliography 229
xiii
ACRONYMS
xiv
INTRODUCTION
1
1.1 motivation
tions between different process model metrics and finding (structurally) similar
process models among a process model collection.
The approach is again applied to the same 33 process model metrics and 515
process models in order to make the findings comparable.
Next, the thesis deals with the measurement of structural process model under-
standability as an example of a very important external process model attribute.
Understandability influences other quality aspects of process models like error-
proneness and maintainability. Even though the importance of understandability
is undoubted, Mendling et al. state that “we know surprisingly little about the act
of modeling and which factors contribute to a ‘good’ process model in terms of
human understandability” [101, p. 48].
Some published studies try to examine the dependencies between some influ-
encing factors and process model understandability: In [138, 139], Sarshar et al.
compare the understandability of different process modeling languages. Recker
and Dreiling examine whether somebody’s experience with one process modeling
language can be helpful for understanding process models based on another
modeling language he/she is not familiar with [125]. Mendling et al. search for
dependencies between personal and process model specific (structural) properties
and process model understandability [101]. In [102], Mendling and Strembeck
also examine the influence of content related factors on process model under-
standability. Reijers and Mendling test the effect of process model modularization
on process model understandability [126].
For examining structural process model understandability and validating ap-
propriate prediction systems (as done in the above mentioned studies), one first
has to quantify structural process model understandability. Thus, a proper mea-
sure for structural process model understandability which fulfills the reliability
and validity requirements has to be found. Looking at the few proposed measures
for structural process model understandability [101, 102, 126], serious doubts
about this necessary validation arise.
In this thesis, concrete and detailed definitions for measuring structural process
model understandability are given which exceed those in existing publications.
Using these definitions, hypotheses about effects of measuring structural process
model understandability are formulated which have to be considered in the
measuring process.
Finally, an experimental evaluation comprising two experiments is conducted to
examine these hypotheses. The first experiment involves 18 students, the second
178 students. The results show the applicability and the behavior of the proposed
measures. Furthermore, they support most of the postulated hypotheses.
The last part of the thesis deals with the experimental evaluation of a postulated—
yet not validated—prediction system.
During the design phase of a process model, choosing the adequate size of
process activities (process model granularity) is a well-known problem. Vander-
feesten et al. have proposed a heuristic for this problem which is inspired by the
concepts of coupling and cohesion in software engineering [129, 168].
4 introduction
1.3 outline
Chapter 1: Introduction
Appendix
This thesis summarizes the results of several years of research. During this time,
several papers were published which present parts of the results. Some of the
chapters of this thesis partially base upon these previously published papers.
6 introduction
Definition
Weske states [175, p. 4]: “Business process management is based on the observa-
tion that each product that a company provides to the market is the outcome of
a number of activities performed. Business processes are the key instrument to
organizing these activities and to improving the understanding of their interrela-
tionships.”
Thus, one has to clarify what one means with “business process” before one
can think about BPM in detail.
Ko gives an overview of different definitions of this term [72, p. 12]. According
to him, most existing definitions can be traced back to a definition given by
Hammer and Champy in their seminal book about Business Process Reengineer-
ing (BPR) [56, p. 35].
“We define a business process as a collection of activities that takes
one or more kinds of input and creates an output that is of value to
the customer.”
The next definition which is cited by Ko was proposed by Davenport in his
seminal book about using information technology for process innovation [35, p. 5].
In this definition, an additional emphasis on the performed activities’ structure
and how work is done can be found.
8 basics of business process management
Finally, Ko quotes a definition by Ould [116, p. 26] which adds two important
elements still missing in the other definitions—the actors which perform the
activities and the collaboration between them.
Classification
In the literature, one can find several classification schemes for business processes.
van der aalst and van hee Van der Aalst and van Hee [162, pp. 9–10]
use the role of a process within a company for classification. They distinguish
production, support and managerial processes. Production processes produce a
company’s products or services and, thus, generate income for the company. Sup-
port processes support the production processes. This class comprises maintenance
processes for the production systems as well as personnel management processes
such as recruitment and selection, training and payment. The last class, the man-
agerial processes, directs and coordinates the production and support processes.
Here, the objectives and preconditions for the managers of the other processes are
formulated, required resources are allocated and contact is held with financiers
and other stakeholders.
2.1 business process management 9
high
collaborative production
business
value
ad hoc administrative
low
low high
repetition
Figure 2.1: Business process classification scheme by Leymann and Roller [77, p. 10].
weske In [175, pp. 17–21], Weske proposes a further classification scheme for
business processes. It consists of the three dimensions degree of automation,
degree of repetition and degree of structuring. The degree of automation measures
the rate of human interaction within a process. The degree of repetition—as in the
classification scheme by Leymann and Roller—declares how often a process is
executed in the same manner. Finally, the degree of structuring indicates whether
a process with all its execution rules (e. g., decision rules for a loan granting
process) can be fully described.
conclusion Independent from the chosen classification scheme, one can state
that the higher a process’s repetition rate, degree of structuring and business
value, the more applicable and useful is the modeling and execution with IT
support of this process.
Evaluation:
business activity monitoring
business process mining
Design:
business process
Enactment: identification and modeling
operation
monitoring Analysis:
maintenance validation
simulation
verification
Configuration:
system selection
implementation
test and deployment
design and analysis The first phase of the BPM lifecycle consists of two
steps—design and analysis.
There are two possible initial situations: (1) The business process is totally new
in the company and has not yet been executed or (2) it is already performed
manually without explicit BPM support. Depending on the situation, there either
2 This definition restricts BPM to operational processes, i. e., processes at the strategic level or
processes that cannot be made explicit are excluded [163, pp. 4–5].
12 basics of business process management
evaluation In the evaluation phase, the available log data is used to evaluate
and improve the process model and its implementation. For these purposes,
business activity monitoring and business process mining techniques are used.
Business activity monitoring can help to identify, for example, bottlenecks of
process model implementations caused by a shortage of required resources.
Business process mining—a rather recently developing field of research [160]—
can be used as a starting point for the development of process models from
log files created by traditional information systems (instead of dedicated BPM
systems). These traditional systems support the execution of processes even
2.1 business process management 13
• Better understanding
The explicit representation of business processes can help a company to get
a better understanding of the operations it performs and the dependencies
(possible “side effects”) between the different processes [175, p. 21] [10,
p. 7].
• Improved communication
The explicit representation of business processes as well as using BPM
terminology can improve the communication between the different stake-
holders by making it more efficient and effective. Through this, also the
collaborative analysis and identification of potential improvements becomes
easier. [175, p. 21]
• More flexibility
BPM can improve the flexibility of business processes for a faster adaptation
to changing market situations and customer requirements. The explicit
modeling of the process and their IT support are important factors to reach
this goal. [175, p. 21] [10, p. 9]
• Benchmarking
The explicit representation and the IT support enable to collect performance
measures at specific points of a process. That way, internal and external
benchmarking becomes possible. [10, p. 6]
• Enabling cooperation/outsourcing
When a process is modeled with all its performed operations and dependen-
cies, it also becomes easier to virtually span processes over the borders of
companies (cooperation) or to outsource parts of or whole processes which
are not the core competency of that company [10, p. 8].
After the definition of BPM and the presentation of the BPM lifecycle earlier in this
section, this subsection gives an overview of the BPM market with information
about its size and offered software tools as well as the actual usage of the BPM
approach in companies.
oracle survey In 2007, an Oracle survey of more than 200 Oracle Business
Process Management Suite customers was conducted worldwide. It revealed the
following areas where customers expect the greatest return on investment from
their investment in BPM (multiple responses possible) [114, p. 8]:
pentadoc and trovarit survey Another German survey with 157 partici-
pants conducted in 2010 by Pentadoc and Trovarit [28, 119] revealed the following
results:
Only 23% of the asked companies use special BPM software [119, p. 5]. Another
23% plan its introduction—yet, around half of them not within the next two years
[119, p. 8].
Both groups have similar goals for a BPM usage—more process control and
transparency, reduction of workload, process speedup, increase of competitiveness
and reduction of costs [119, pp. 4, 9]. Respondents using BPM could reduce lead
times by 37% and costs by 31% on average [119, pp. 4–5].
Participants which do not use BPM name as reasons: lack of available competent
resources (almost 50%), unclear or too small benefit (around 45%), expectation of
too much effort (43%) and expectation of too much costs (30%) [119, pp. 5–6].
Besides BPM, the term “Workflow Management (WfM)” can be found quite often
in literature. WfM as a concept is older than BPM (see [52] for a—meanwhile
“historic”—overview of WfM). While BPM is a process-oriented management
discipline (see Definition 2.2), WfM is a technological approach for the automation
of business processes.
According to van der Aalst et al. [163, p. 5], WfM can be seen as today’s tech-
nological basis of BPM and can be integrated into the BPM lifecycle—comprising
all phases except the evaluation phase.
Definition 2.3 (Workflow Management) Workflow Management deals with the au-
tomation of business processes, in whole or part, during which documents, information or
tasks are passed from one participant to another for action, according to a set of procedural
rules [176, p. 8].
There are manual activities, which require the execution by humans, and
automated activities, which can be done fully automatically by machines and/or
computers.
Each activity needs a resource (human or machine/computer) with specific
“skills” to be executed. The requirements for that resource are summarized within
a role.
Definition 2.6 (Role) A role comprises attributes as skills, location and authority
within an organizational structure which are either required by a resource to execute a
special activity or which a resource provides [176, pp. 54–55].
A business process (e. g., an insurance claim process) can be initiated and
executed indefinitely often—normally with different data each time.
The main goal of WfM is the software-supported execution of processes. For that
purpose, Workflow Management Systems (WfMSs) are used.
• “navigation” and “routing” between the activities (Which activity is executed next
and which branch is taken at branchings?)
The WfMC has created the Workflow Reference Model (see Figure 2.3) as a
generic architectural representation of a WfMS including its generic components
and most important system interfaces (using a workflow Application Program-
ming Interface (API) and standardized interchange formats) [58, pp. 20–27] [176,
p. 23]
In the area of BPM, numerous standards have emerged which are used to create
process models of business processes and to interchange and/or execute these
models.
2.3 process modeling languages 19
process definition
tools
administration
& monitoring tools workflow
engine(s) workflow
engine(s)
All these listed process modeling languages are activity-based, i. e., process
models modeled with these languages consist of the activities to be executed
and a network of arcs between them modeling the possible sequences of activity
execution. Independently from the chosen language, several so-called control
flow patterns [175, pp. 126–149] can be expressed in these languages—(linear)
sequence, parallel (AND) and alternative (XOR or OR) execution being the most
frequently used.
In the following two subsections, two process modeling languages which are
used in this thesis are presented in more detail—EPCs (Chapter 4 and 5) and
Product Data Models (Chapter 7).
20 basics of business process management
Event-driven process chains (EPCs) are a development by the Institut für Wirt-
schaftsinformatik (Institute for Information Systems) at Universität Saarbrücken. In
the 1990s, there was a project together with SAP to define a suitable business
process modeling language to document the processes of the SAP R/3 enterprise
resource planning system. This project produced two major results: the definition
of EPCs [70] and the documentation of the SAP system in the SAP Reference
Model [33, 71]. [99, p. 17]
The following formal definition is based on [175, p. 162]. Yet, it partially uses
some terminology from [99, pp. 22–23] instead.
• C is a set of connectors.
Functions (see Figure 2.4b for the graphical notation) represent the activities
of the modeled process—events (Figure 2.4a) represent their pre- and post-
conditions. Connectors are the third node type of an EPC. They are used for
modeling non-sequential control flows. Connectors can be divided into split (one
incoming and several outgoing arcs) and join (several incoming and one outgoing
arcs) connectors. Besides, each connector node has one of the three types AND
(Figure 2.4c), OR (Figure 2.4d) or XOR (Figure 2.4e). At AND split connectors, all
2.3 process modeling languages 21
event function V
name name V XOR
(a) Event. (b) Function. (c) AND connec- (d) OR connec- (e) XOR connec-
tor. tor. tor.
Product data models are the process modeling language used in Product-Based
Workflow Design (PBWD)—a modeling methodology proposed by van der Aalst,
Reijers and Limam in [128, 159]. It is based on former work by van der Aalst
[157, 158] and is further examined by Reijers in [127].
The motivation of the PBWD methodology is the area of manufacturing where
the interaction between the design of a product and the process to manufacture
this product is studied in detail [128, p. 229] [18, pp. 469–471].
There, a so-called bill of materials (BOM) [115] [18, pp. 138–146] is used to
define the design of the product. A BOM has a tree structure with the final end
product as its root and raw materials and purchased products as leaves. The nodes
correspond to products (end products, raw materials, subassemblies). The edges
represent an is-part-of relation and have a cardinality to indicate the number of
products needed. The simplified BOM of a car is depicted in Figure 2.6. According
to that, a car is composed of an engine and a subassembly. The subassembly
consists of four wheels and one chassis. [128, p. 234]
In [157, pp. 397–398], van der Aalst et al. emphasize that there is also this kind of
dualism between product and process model in the information-intensive service
industry—even though, it is seldom made explicit. They give the example of
processing an insurance claim. There, the product is basically a decision: either the
claim is accepted (followed by a payment) or it is rejected. Different data elements
22 basics of business process management
event A
function B
XOR
event C event D
function E function F
XOR
event G
may play a role in making this decision. One can think of these data elements
as raw materials or subassemblies. The process model should manufacture the
decision.
Reijers et al. criticize that “[i]n contrast to manufacturing, the product and
the process have often diverged in actual workflows. Workflows found in banks
and insurance companies for products like credit, savings and mortgages, dam-
age and life insurance, and so on may well exist for decades. Since their first
release, those processes have undergone an organic evolution. [. . . ] Aside from the
evolutionary changes of the processes, the state of technology of some decades
ago has considerably influenced the structure of these workflows permanently.
[. . . ] So, the structure of an actual workflow may not be related to the product
characteristics any more.” [128, p. 230]
The PBWD methodology tries to reverse this divergence of product structure
and process model for information-intensive processes of the service industry
(e. g., bank, insurance and telecommunications companies). Instead of defining a
2.3 process modeling languages 23
car
engine subassembly
wheel chassis
process model at once, it is derived from the product structure. So, the PBWD
methodology provides the following steps [128, pp. 231–232]:
For the first step, a formalization for the product structure is necessary. Com-
pared to manufacturing, information-intensive processes have differences which
make the BOM not very useful [128, p. 234]:
Definition 2.11 (Product Data Model) A Product Data Model is a tuple (D, C, pre,
constr, cst, flow) for which holds:
• C is a set of constraints which can be any boolean function (including the boolean
value true).
• pre : D 7→ P(P(D)) is a function which gives for each data element the various
ways of determining a value for it based on the values of different sets of other data
elements so that holds:
– R := {(p, c) ∈ D × D|c ∈ es∈pre(p) es} is connected and acyclic, i. e., there
S
are no “dangling” data elements and a value of a data element does not depend
on itself.
– The top element cannot be used for determining the value of any other data
element:
∀(p, c) ∈ R : c 6= top
– If there is a data element which does not require any other data element, one
denotes for ease of analysis the set of required data elements as the empty set:
• flow : F 7→ N is a function which gives the time it takes to use a production rule.
A PDM is an acyclic graph. Its nodes are data elements which all have a value.
There is one special data element, the top data element, which represents the
final decision of the corresponding process. The edges of the graph represent
the relations between the data elements which are given by the pre function. It
yields for each data element d zero or more variants to determine a value for
d. If one supposes for the data elements d, e, f ∈ D that {e, f} ∈ pre(d), then the
value of d can be determined using the values of e and f—one says that (d, {e, f})
is a production rule of d. Each production rule has a corresponding constraint
(a boolean function) and can only be executed if its constraint evaluates to true.
Furthermore, the execution of each production rule causes a specific amount of
costs and needs a specific time. There are special data elements whose values do
not depend on those of others. They are called leaves.
An example of a PDM is depicted in Figure 2.7. The corresponding process
checks whether a person is suitable as a helicopter pilot. On the one hand, there
2.3 process modeling languages 25
suitability as
helicopter pilot
latest suitability
psychological
physical fitness result (max. 2
fitness
years old)
quality of quality of
reflexes eye-sight
Figure 2.7: Example of a PDM for a process checking the suitability as a helicopter pilot
[159, p. 399].
are production rules (e. g., from quality of reflexes and quality of eye-sight to
physical fitness) which need more than one data element to compute the value
of another one. On the other hand, there are other production rules (e. g., latest
suitability result to suitability as helicopter pilot) which only need exactly one
data element to compute the value of another one. Furthermore, there exist three
variants how to compute the final decision (suitability as helicopter pilot). Based
on the evaluation of their constraints, one of them can be selected: If the quality
of eye-sight is too bad, the final decision can immediately be made: the examinee
is not suitable. If the latest suitability result is not older than two years, its result
is adopted. Only otherwise, a more complicated check has to be performed. The
quality of eye-sight is an example of a data element whose value can be used for
several production rules.
In the next step of the PBWD methodology, one or several alternative process
models have to be derived from the PDM. For that, each production rule is
transformed into an adequate activity which performs the corresponding produc-
tion rule. The control flow between these activities has to preserve the relation
induced by the pre function of the PDM. So, the values of the data elements can
be computed in a correct sequence.
For the last step of the PBWD methodology, Reijers et al. propose some heuris-
tics how to select the optimal alternative process model based on the PDM’s cst
or flow function. As the details are unimportant for this thesis, the interested
reader is referred to [128, pp. 240–244].
PROCESS MEASUREMENT
3
3.1 introduction
them taken from [96] and [100], but also new ones). Once again, he uses logistic
regression in order to identify possible predictors for faulty process models.
In [101], Mendling et al. present an experiment for identifying influencing
personal (theoretical knowledge and practical experience) and structural factors
on process model understandability. They use the SCORE measure—the sum of
correct answers to eight closed and one open question on a process model—to
measure understandability.
In [167], Vanderfeesten et al. introduce the cross-connectivity metric. It measures
the average strength of connection between all pairs of process model nodes.
They empirically evaluate the metric using data of [101].
Mendling and Strembeck present a second experiment for identifying influ-
encing factors on process model understandability [102]. This time, also content
related factors (task labels) are analyzed.
In [135], Sanchez et al. emphasize the importance of process measurement
for the determination of the maturity level according to the Business Process
Maturity Model (BPMM) [113].
theory and Kolmogorov complexity, which are presented in the remainder of this
subsection), mathematical definitions are available. Generally, only “philosophical
definitions” exist. The Merriam-Webster’s Collegiate Dictionary, for example,
defines the adjective “complex” as “hard to separate, analyze, or solve” [106,
p. 235] (as quoted in [74, p. 4]).
The remainder of this subsection is structured as follows: First, statements on
process model complexity found during the literature review are reproduced.
Afterwards, several aspects of complexity which are research objects in different
research disciplines are presented. Next, some thoughts on the meaningfulness
of measuring complexity are given. Finally, consequences regarding measuring
process model complexity are drawn.
Aspects of Complexity
In several fields of research, different aspects of complexity are examined. For
some of them, formal definitions exist—yet, for most of them, only informal
textual descriptions of the concept are available.
Here, a short overview of some important aspects of complexity is given.
Time and Space Complexity If one knows that a problem is computable, the
next interesting step is to look at its runtime and space (memory) requirements.
In doing so, one has to notice for some computable problems that they are
intractable in practice.
Before time and space complexity are defined, three asymptotic notations [32,
pp. 41–50] are introduced which are used later.
Definition 3.1 (Ω-notation) For a given function g(n), one denotes by Ω(g(n)) the
set of functions
Definition 3.2 (Θ-notation) For a given function g(n), one denotes by Θ(g(n)) the
set of functions
f(n)
c2g(n) cg(n)
f(n) f(n)
cg(n)
c1g(n)
n n n
n0 n0 n0
Figure 3.1: Graphical examples of the (a) Ω-, (b) Θ- and (c) O-notations. In each subfigure,
the shown value of n0 is the minimum possible value. [32, p. 43]
The Θ-notation states that a function lies between an asymptotic lower and
upper bound.
Definition 3.3 (O-notation) For a given function g(n), one denotes by O(g(n)) the
set of functions
O(g(n)) := {f(n)|∃c, n0 > 0 ∀n > n0 : 0 6 f(n) 6 cg(n)} .
Table 3.1: Runtime depending on input size n for several “algorithms” with different
number of computation steps (under assumption: 1,000,000 computation steps
per second).
More details about the definition of the problem itself and possible approaches
to deal with it can be found, for example, in [79] (structural complexity manage-
ment, product complexity) and [13] (complexity management in supply chains).
networks
Besides modeling, important research questions in this area are network re-
silience (against random or intentional failures of parts of a network), epidemics
on networks (spread of diseases or computer viruses) and dynamical systems on
networks.
A good introduction into networks is the textbook by Newman [111].
Even though no real complexity measure is defined in this area, networks—and
especially their dynamics—are an important aspect of complexity in the real
world.
Network Models Over the years, several (theoretical) network models were
proposed and analyzed in order to get better insights into the creation process
and the properties of real world networks.
Random graphs (see [65] for details) are probably the oldest of these models.
There are two main “construction methods” for getting such a graph with n
nodes: (1) randomly choosing one of all graphs with n nodes and (2) starting with
n nodes and no edges and than adding each possible edge between the nodes
with a predefined probability. In recent years, it was discovered that random
graphs do not show and explain some important characteristics which one can
find in real world networks. Thus, other network models were suggested.
The development of the next network model, small world networks, was triggered
by a meanwhile famous experiment conducted by Milgram [104, 154]. He asked
296 randomly chosen persons in the USA to send a letter to a target person in
Massachusetts who they did not know. Yet, they were not allowed to send it
directly to the target person but were told to send the letters only to a person
they personally knew and ask him/her to proceed accordingly with the final goal
to reach the target person. 64 letters (21.6%) finally reached the target person
with an average of 5.2 intermediaries.
The reason for this small world phenomenon is a special network structure
with many local clusters (almost everybody in a cluster knows each other as,
e. g., relatives, friends or colleagues) and some so-called weak links between
these clusters (one member of one cluster knows somebody of another one). This
structure is depicted in Figure 3.2a.
Random graphs do not show—and thus—explain these properties. In [172],
Watts and Strogatz propose an algorithm which produces networks with these
small world properties.
Yet, also small world networks do not perfectly represent the entire character-
istics of real work networks. Most node degree distributions of real networks
follow a so-called power law, i. e., these networks have many nodes with few con-
nected “neighbors” and only few nodes (so-called “hubs”) with a large number
of “neighbors”. One can take, for example, airports and the flight connections
between them as an example. There are many small airports which only offer
flights to a few larger airports. On the other hand, there is a relative small number
of large airports (hubs as, e. g., London Heathrow) which are connected by flights
to many other hubs worldwide. This situation is depicted in Figure 3.2b. The
resulting networks are called scale-free networks.
3.3 discussion of related work 37
number of nodes
●
●
●
●
●
●
●
● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
node degree
(a) Local clusters and weak links (dashed lines) (b) Node degree distribution of a scale-free net-
in a small world network. work.
Figure 3.2: Illustrations of properties of (a) small world and (b) scale-free networks.
In [7, 8], Barabási et al. suggest and analyze an algorithm which is able to
produce such free-scale networks. The main principle of this algorithm is that a
new edge is inserted between a new node and and an already existing node by a
probability proportional to the current node degree of the existing node.
models the size of a population over time. This population size is a number
between 0 and 1. Parameter a is a growth factor.
For small values of parameter a (e. g., a = 2), the population size converges
to one value independently from the start value x0 (see Figure 3.3a). For larger
values of a (e. g., a = 3.3), the long-term population size alternates between two
values (see Figure 3.3c). If one further increases a, instead of alternation between
38 process measurement
two values, an alternation between four, then eight, etc. different values occurs.
For a = 4, finally, a non-periodic behavior of the population size can be observed.
Furthermore, the time series becomes totally different in the long term even for
very small differences of the start value x0 (see Figure 3.3b and 3.3d).
This effect is the reason for the term “deterministic chaos”. Even though the
underlying equation of the iteration is known and all iteration steps can be
deterministically computed if one knows the start value, the long-term behavior
of the iteration is “chaotic” as one cannot determine the start value without any
small error in practice.
The long-term behavior of the Verhulst equation for the different values of a
can be visualized using the Feigenbaum diagram (see Figure 3.4). Here, only the
long-term values of the iteration are displayed depending on the value of a. One
can observe the behavior described above: First one fix point, then two, four,
eight, etc. alternating values and finally the chaotic behavior for large values of a.
Control Theory Control theory deals with the question whether it is possible
to bring a dynamical system into a certain state. If it is possible, the identification
of the necessary control inputs is the second issue. An overview of control theory
can be found in [14]. An analysis of the computational complexity of different
control problems is given in [145].
Using different experimentation systems (see, for example, [47, pp. 188–204] for
an overview), several influencing factors (e. g., [45, pp. 20–21] [47, pp. 204–206]
3.3 discussion of related work 39
1.0
1.0
0.8
0.8
0.6
0.6
x
x
0.4
0.4
0.2
0.2
0.0
0.0
0 5 10 15 20 0 5 10 15 20
iteration iteration
(a) Time series for a = 2 and x0 = 0.11 (b) Time series for a = 4 and x0 = 0.11
(squares) or x0 = 0.79 (triangles) respec- (squares) or x0 = 0.11001 (triangles) respec-
tively. tively.
1.0
1.0
●
0.8
0.8
●
0.6
0.6
difference
●
x
●
0.4
0.4
●
0.2
0.2
●
●
●
●
● ●
0.0
0.0
● ● ● ● ● ● ● ● ● ●
0 5 10 15 20 0 5 10 15 20
iteration iteration
(c) Time series for a = 3.3 and x0 = 0.11 (d) Absolute value of difference between iter-
(squares) or x0 = 0.79 (triangles) respec- ation steps of the two start values x0 of
tively. Subfigure (b).
[48, pp. 250–251]) were examined in studies (see [48, pp. 251–260] for a review of
studies according to the examined factor).
There are two main types of conducted studies: studies using computer-
simulated mircoworlds [15] as well as studies using finite state automata or
linear equation systems [17, pp. 42–53]. For the first type, a system (e. g., a city)
is modeled in the computer. The test subject (e. g., acting as the city mayor) can
influence some variables of the simulated system (e. g., the tax rate) in order to
achieve a certain goal.
40 process measurement
1.0
●
●●
●●●● ●
●●●●●●●●
● ●
● ●●●
●●●
● ● ●●●● ●
●● ●● ● ●● ●●
●● ●●●
●●●
●● ● ● ●●● ●● ●
●●●●
●● ●●●●●●● ● ● ●●●●● ● ●
●●●●
●●
● ●● ●●●●
● ● ●●
●●● ●●● ● ● ● ● ● ●● ● ●● ●
●●● ●● ●
●●
● ● ●● ●●
● ● ●
●● ● ●
● ●
● ●● ●● ●●● ● ● ● ●● ●
●● ●●●●●●
● ●●● ● ● ● ●●● ● ●● ● ● ●
●●●
●●●●●●●
●● ●● ● ● ● ● ●
●●
●●●●●●● ●
● ● ● ● ● ●●● ●●
● ●●●
● ● ●
●
●●● ●●●●●● ● ●● ● ●●● ●
●● ●
●●● ● ● ● ●● ● ● ●● ● ●● ● ● ● ● ● ● ● ● ●●
●●●
●●●●● ● ●● ●
● ●● ● ●● ● ●
●●●●
●●●●
●●●●
●●●●●● ●● ● ● ● ●
● ●● ●
●●●
● ● ●●● ● ● ●● ●
● ● ●● ● ● ●●●●●●● ● ● ● ●●
●●
●● ●●
●
●
●●●
●●
● ●●●
● ● ●
●●● ●
●● ●● ● ●●●● ● ● ● ●●
● ● ●● ● ● ● ● ●
●
●●●●●●● ●
●●
●●● ●●● ●● ●
●● ● ● ● ● ● ●● ●●● ●
●● ●●●●●●
●● ●
●●●●● ●● ●● ● ●●● ●
●●
●●● ●
● ●●● ●● ●● ● ●●
●●
● ● ● ● ● ● ●● ● ● ● ● ●●
●●●●● ● ●● ● ●
●●
●●●
●● ●●● ● ● ●● ● ●
●●●
● ● ●●
●● ● ● ●
●●●● ●● ● ●● ●● ●● ● ●●●● ● ● ●
●●●●● ● ●●●● ● ●●●●● ● ●●●
●● ● ●● ● ●●
● ●● ●●
●●
●●●●● ●●● ● ●● ● ●● ● ●● ●● ● ●
●● ● ●
●
●
●●
●● ●● ●●●● ● ●●
●
●● ●●● ●
●
● ●● ●
●● ● ● ●● ●
● ●
● ●
●●● ● ●● ●● ●● ●● ● ● ● ● ●● ●● ● ● ●●●
●●
●● ● ●● ●● ● ● ●● ● ●●●● ●
●●● ● ● ●● ● ●●● ●● ● ●
●●●●
● ● ●●
●●●
●
●●
●● ●●● ● ● ● ●● ● ●● ● ● ● ● ● ● ●●●● ● ●● ● ●
●●● ●
●● ● ● ● ● ●● ● ●● ●
● ●● ●● ● ● ● ●●● ●
●●●● ● ● ●● ● ● ● ●● ● ●● ● ● ● ● ●● ●
●● ●● ●
●●● ● ●●● ●
● ● ●● ● ● ● ●● ●●
●●
●● ●
●●●●
● ●●● ● ●●●● ● ● ●● ● ● ● ●● ● ●●
●●●●
●●● ● ● ●
●● ● ● ●●● ● ● ● ●●● ●● ● ● ● ●
●●● ●● ●● ● ● ● ● ●
●●●●
●● ● ● ●● ● ●●
● ●●
●● ●
● ●● ●
●● ● ●
● ● ● ● ● ● ● ●
●●●●● ●● ● ● ● ● ●●● ● ●●● ● ●●● ●●● ●●●●● ● ●● ● ● ● ●● ● ●● ●
●●
●●● ●●
●●● ●● ● ● ●● ● ● ●● ● ●
● ● ●●●
●●●●
●●●●● ●
● ●●
● ● ●● ●● ● ● ●
● ●● ●●●● ●
●● ● ●● ● ●● ● ●
● ● ●
●
●
●●
● ●●● ● ● ● ● ●● ● ●● ●● ● ● ●● ●
●●
●●● ● ●●●● ●●● ● ●● ● ●● ● ●●
●●● ● ●● ● ● ● ●●
●●●● ●
●●●
● ●● ●
●
●● ● ●● ●●
● ●●●●
●● ● ● ●● ● ● ● ● ● ●●● ●
●● ●
●●● ●●●
●●● ●● ● ● ●● ● ●●● ● ●●●● ● ●● ●● ● ●● ●
●● ●● ●●● ● ● ● ●● ● ●● ● ● ●●●● ● ●
● ●● ● ●
●● ● ●
●●●●
●● ● ●
● ● ● ●● ●● ● ●● ● ● ● ● ●
● ●
●●● ●
●● ●●●●● ● ● ●
● ●●●● ●●
●
●●●● ●
● ● ● ● ●● ●● ● ●
●●●
●●
● ●●
● ●●
●● ● ● ●●● ● ● ●
● ● ● ●●
● ●
● ●● ●●● ●
● ● ●● ●● ● ●●
●●●● ●●●●
●●
● ● ●●● ● ●● ●●●● ●●● ● ●●
● ●
● ● ●●●
●●
● ● ●● ● ●●
●●
●● ●● ● ● ● ● ●
●
●●
●
●● ●●●● ● ●●● ●●●
● ● ● ●● ● ●● ●● ●● ● ● ● ● ● ●● ● ●● ● ● ●● ●● ● ● ●● ●
●●● ●
●
● ● ●●●●●
● ● ● ● ● ●● ● ● ● ● ● ●
●
●
● ● ● ●● ● ● ● ● ●
●
●
●● ●●● ● ●● ● ●● ● ●● ●●●
● ●● ● ● ● ● ● ● ● ● ● ●●
●●●● ● ●●● ●
●●●●●● ●●● ● ●● ●●●
● ●● ● ●● ●●
●● ● ●
●●●● ●●●
● ● ● ● ●● ●●● ● ● ●●
●● ● ●●●● ● ● ● ● ●● ● ● ●
● ●
●●●● ●●● ●● ●●
●● ●●●●●●● ● ●● ● ● ●●
●●● ● ● ●●●● ●●● ● ● ● ● ●
●●● ●● ●● ●
●● ● ● ●● ● ● ● ● ● ●● ● ● ● ● ● ●●
●●● ●● ●
●●
●●●●●
●● ●●
●●
● ●● ●●●
● ● ●●
● ●● ● ● ● ● ●● ● ● ● ●●● ●●● ● ●●
● ●
●●● ● ●● ● ●
●● ● ● ●●● ●● ●
● ● ● ● ● ● ● ●
●●● ●●
●●● ●●
● ●●● ●● ● ●● ●● ●●●
● ●● ● ● ●●● ● ●● ●
●● ●●
●● ●● ● ●●●
●● ● ● ● ● ●●● ●● ● ● ● ● ● ●● ●
●● ●●●
●●●●●●● ● ● ● ● ●● ● ● ●●
●● ● ●● ● ●
● ● ● ●●
●●● ●●
●● ●●● ● ●● ● ●
●● ● ● ● ● ● ● ● ● ● ● ●
●● ●● ●
●● ● ● ● ●●● ●● ● ●
● ● ● ● ●
● ●● ● ●●● ● ●
● ● ● ● ●
●
●
●
● ● ●● ● ● ● ●● ●● ● ●● ● ● ●● ● ● ● ● ●
● ●● ●●● ●● ●●●● ●● ●● ●● ● ● ●●●● ●● ● ●● ●●● ● ●●
●
●
● ● ●
● ●
●●
● ● ● ● ●
● ●● ● ● ●● ● ● ●●
● ● ● ●●● ● ● ● ● ● ●● ● ●
●
● ● ● ● ● ●● ● ●● ● ● ● ●● ● ●● ● ● ● ●● ●
●●●●● ● ● ●● ● ●● ● ● ● ●● ●● ● ● ●● ●●
0.8
● ● ● ●● ● ●● ● ●●
● ● ● ● ●
●●●●● ● ●● ●●●● ●●
● ●●● ● ●
●● ● ●
● ● ● ● ● ●
●●●●● ● ● ● ●
● ●●●●
● ● ● ●● ●
● ● ●● ●● ● ● ● ●
●●●●● ● ● ●● ● ●● ● ● ● ● ● ● ●● ● ● ● ● ●● ●
● ●● ● ● ● ● ● ● ●●
●●●●● ●● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ●
●●●●●●● ●● ● ● ● ● ●● ●● ● ●● ● ● ●
● ● ● ●● ● ● ●● ● ● ●● ●
●● ●● ● ● ● ● ● ● ●● ● ● ● ● ● ● ●● ● ● ● ●● ● ●● ● ●
●●●● ●● ●
●●● ● ● ● ●●● ● ● ●●● ● ● ● ● ● ● ●● ● ● ● ● ●● ●● ● ●●
●●●●● ●●● ●● ● ●●●
●● ● ● ●●●●● ● ● ● ●● ●● ● ● ●
●●●● ●●●● ● ●● ●
● ●● ● ● ●
● ● ● ●● ●● ● ● ●● ● ● ● ● ● ● ● ● ● ●●
●●●● ●●● ●● ● ● ●● ● ● ● ● ● ●● ● ● ● ● ●
● ● ● ●● ●
●●●● ●●●●● ● ●● ● ● ●● ● ● ● ●● ● ● ● ● ●● ●
●●●● ●●●●● ● ●● ● ● ●● ● ●●
● ● ● ●● ●●● ● ●
●●● ●●●●● ● ●● ●
●● ●● ●●● ● ● ● ● ●● ● ●● ●
● ●● ● ● ● ● ●●
●●●● ●● ● ●● ● ● ● ● ●
●● ● ● ● ● ●● ● ●● ● ● ●● ● ● ● ●
●●●● ●● ● ● ● ● ●
●● ●● ●● ● ●● ● ●
●●● ● ● ●● ● ●● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ●
●●●●● ●● ●
●●
●
● ●● ●
● ● ● ● ● ●● ●● ●●●● ●
●
●
● ●
●● ●
●●●● ●● ●● ● ● ● ●●● ● ● ● ●● ●
● ●●●
●
● ● ●
● ● ● ● ● ● ● ● ● ●● ● ● ●●
●●● ●●●●●●●●●● ● ● ● ● ●● ● ● ●
●●● ●●●●● ●●● ● ●● ●● ●●● ● ● ● ● ● ● ● ● ● ●●● ●
●●
●●● ● ● ●● ● ●● ●
●●● ●● ● ●●● ● ● ● ●
●●● ●●
● ●● ● ● ● ● ● ● ●
●●● ●●● ●● ● ● ●● ●● ●● ● ● ● ● ● ● ●● ● ● ●
●●● ● ●● ● ●●● ● ●● ● ● ● ●
● ●
●●● ●● ●
● ●● ● ●
●●● ● ● ●●● ● ● ●●
●● ●● ●● ●●●●● ● ●● ● ● ● ●● ●●● ● ● ● ● ●● ● ●●
●●● ● ●
●●● ● ● ● ● ●
●●
● ●●● ● ● ● ● ●
●●● ●● ●
● ●● ●●● ●●● ●● ● ●●
● ●● ● ● ●●
●● ●● ● ● ● ● ● ● ●
●● ● ●●●
●
● ●●
● ● ●●● ●
●
●● ●● ●
● ● ●●● ● ●●● ● ● ● ●
●●● ● ●
● ● ● ● ●●● ● ● ●●● ● ●●● ● ●●
● ● ● ● ● ● ●
●● ●● ● ●● ● ● ●● ● ●
● ● ● ●
●● ● ●●● ● ● ●
● ●
● ● ●● ● ●● ●●●●
● ● ● ● ●● ● ● ● ●
●● ● ● ●
●●● ●●●● ● ● ● ● ● ●● ●● ● ● ●●
●● ●
●
●● ● ●●● ● ● ● ●● ● ●
● ● ● ●● ● ●●
● ●
●● ●● ● ● ●●
● ● ●●● ● ● ● ●● ● ● ● ● ●●
●● ●●
● ●● ● ● ●● ●● ● ●●● ● ● ● ● ● ● ● ● ●● ● ●●
●● ● ●●●
●●● ● ●● ● ● ● ● ● ●● ● ● ● ● ●●
●● ● ● ●● ●
●● ● ● ●● ● ●● ● ● ●● ●● ● ● ● ●
●● ● ●●
● ● ● ●● ● ● ● ● ● ● ●●
●● ● ● ●
●● ● ● ●● ● ● ●
● ●
●●
●
●● ● ●● ● ● ●● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
●● ●●● ● ● ●● ● ●● ● ● ● ● ●● ●● ● ● ● ● ●● ●
●● ● ●● ● ●●● ● ●● ● ● ● ●
●●
● ● ● ● ●
● ● ● ● ● ● ●●
●● ●
●●●● ● ●● ● ●●● ●
● ● ●●● ● ●● ●● ● ● ● ●
● ●●
● ● ● ● ● ●●●● ●● ● ● ● ●
●● ●● ● ●●● ● ●● ●●● ● ● ●
● ● ●●●
●
● ● ● ●● ●● ● ● ● ● ●
●● ●
● ● ●●●
●●
● ●● ● ●● ●
● ●● ● ●● ●
● ●
●● ●● ● ● ● ● ●● ● ● ●
●
● ●●
● ● ●● ●● ● ● ●● ● ● ● ●● ● ● ●
●● ●●●● ● ●●
● ● ●●●
●●● ●●●
● ● ● ● ●● ● ● ●●
● ● ●● ●● ● ● ●●●● ● ● ●
● ●●●● ● ●
●● ● ●
●●
●
● ●
● ● ●●● ●● ● ●
●● ●
●●● ●
●
● ●
●●
●
● ●
● ● ●● ●● ● ● ● ● ●●
●
● ●
● ●
● ● ●● ● ●● ● ●
● ●● ● ●●●● ● ●● ●●
● ●
● ●● ● ● ● ●● ●● ●
● ●●●● ●● ● ●● ●
● ● ● ●●
● ● ● ● ●● ● ● ●●
●
● ●● ● ● ● ●●● ● ●
● ● ● ● ●●● ●● ● ● ● ●
●●●● ● ●● ● ●
● ●● ● ● ●●
●●●
● ● ●● ●
● ● ●● ● ●● ●● ●
●
● ● ●● ● ● ● ● ●●
●
● ●●● ● ●● ● ● ● ● ●●● ● ●● ● ● ● ● ● ●
●
●
● ● ● ● ●
●
●● ● ● ● ● ●● ● ●●● ● ●● ● ● ● ● ● ● ● ● ● ● ●●
●
●● ●● ● ●● ●● ● ● ● ● ● ●● ● ●
●
● ● ● ●
● ● ● ●● ● ● ● ● ● ●
●
● ● ●●● ● ●● ● ● ●● ● ●
● ●
●
● ● ● ●● ●● ● ●
●● ● ● ●● ● ●● ● ●● ● ● ● ●
●
● ● ●● ● ●
●● ● ●●● ● ● ● ● ● ●●
●● ●
● ● ● ● ● ● ●● ●
● ● ●
●
● ● ●● ● ● ● ● ● ● ● ● ● ● ● ●●
●● ● ● ● ● ● ●● ● ● ●
●
● ● ● ●●● ● ● ● ●● ●●
● ● ● ●● ●● ● ● ●
● ●●● ● ●●● ●●●● ● ● ● ● ● ●
● ● ● ● ●● ● ●● ● ● ●● ● ● ●●
● ● ●●
● ●● ● ● ● ● ● ●
● ● ● ●● ●● ●
● ●● ● ● ●
● ●● ● ● ● ● ●
● ●● ● ● ● ●
●
● ● ● ●● ● ● ● ● ●
● ● ●● ● ● ●● ● ● ●● ●
● ●● ● ●●
● ● ● ● ●● ● ● ● ●
● ● ● ● ● ●● ● ● ●
●
●●
●●
●●
●●
●
●● ● ●
●
●●●●
●●
●●
●●
●●
●● ● ●●
●● ● ● ● ●● ● ● ●●
● ●● ● ● ●
●●●●●● ● ● ● ● ● ● ● ● ● ● ●● ●●
●●●●●●●● ● ●● ●
●●
●
●● ● ●
●● ●● ● ●
●
●
●●
● ●
●
●●●●●● ● ● ● ● ● ●● ● ● ● ● ● ● ●
●●●●●● ● ● ● ● ● ●● ● ● ● ●● ●● ● ● ● ●
●●●●●●●● ● ● ● ●● ●● ● ● ●●
●●
●● ● ●● ● ● ●
● ● ● ● ● ●● ●●
●●●●●● ● ●
● ●● ●● ●● ● ● ●
●●●●●●●●●● ● ●
● ● ● ●● ● ● ● ●● ● ● ● ● ●
●
0.6 ●●●●●●
●●●●●●●●
●●●●●●
●●●●●●
●●●●●●
●●●●●●
●●●●●●
●●●●●●
●●●●●●
●●●●●●
●●●●●●
●●●●●●
●●●● ●●
●
●
●●
●
●
●●
●
●
●●
●
●
●●
●
●●
●
●●
●
●
● ●
●●
●
●●
●
●
● ●
●
● ● ●●
●●●
● ●
● ●●
● ●
●● ●
●
●
●
●
● ●
● ●
●
● ●
●
●
●
●
●
● ● ● ●● ● ●●
●
●
●
●
●● ●
●
●
●●
●●
● ●●● ●
●
●
●● ●●●●
●
●
●
● ●●
●
●● ●
●
●
● ●
●●●● ●●
●●●●
● ●
●●● ●
● ●●● ●
●
●●
●
●
●●● ●
●
●
●
●●
●
●
● ●● ●
●
●
●
● ●
●
●● ●
●
●●
● ●●
●
●
●
●
●●
● ●●
●
● ●
●
●● ●● ●
●
● ●
●●
●
●
●
● ●
●
●
●●
●●
●
● ●●
●
●
●
●
●●
●●
●●●
●
●● ●
●
● ●
●
●
●
●
●●●●●● ●●
● ●● ● ● ●
●● ● ● ●
● ● ● ●
●●●●●● ●● ● ● ● ● ● ● ●
●●●● ● ● ● ● ●
● ● ● ●● ● ● ●
●●●●●● ● ● ● ●● ● ● ●
● ●●
● ● ●
●●●●●● ● ● ● ● ● ● ●● ● ● ●● ● ●
●●●●●● ● ● ● ● ● ●● ● ● ● ●
●●●● ● ●● ● ●● ● ● ● ● ● ● ● ●●
●
●●●●●● ● ● ● ● ● ●●
● ● ● ●●●
●●●● ● ●
● ●●● ●● ● ● ● ● ●● ●
●●●●●● ● ●
● ● ● ● ●● ● ● ● ● ●
●●●● ● ● ● ● ● ●● ● ● ●
●●
● ●
●●●●●● ●●
● ● ●● ●● ● ●● ● ●
● ● ● ● ●
●●●● ● ● ● ● ●● ● ●
●
●●●●●● ● ● ● ● ● ●● ● ● ● ● ● ●●
●●●● ● ● ●● ●●●
●● ● ●● ● ● ● ● ● ●
●●●●●● ● ● ● ● ● ●
● ●
● ●
●●●● ● ● ● ● ●● ●● ● ● ●● ● ● ● ● ● ●
●●●● ● ● ● ●●● ●● ●●● ● ●
●● ●●● ●
●●●●●● ●● ●● ● ● ● ●● ●
● ● ● ●
●●●● ● ● ● ● ● ● ● ●
●●●● ● ● ● ●● ● ● ● ●
●●●●●●●● ●● ●● ●● ● ● ● ●
● ● ●
●
●
●
● ● ● ● ●
●● ● ●● ● ●● ● ●
●●●● ● ●● ● ● ● ●● ● ● ● ● ● ● ● ● ●
●●●● ●● ● ●● ● ●● ● ● ● ● ● ● ● ●
●●●● ●● ●● ● ● ● ● ● ● ● ● ● ●
●●●●●● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ●
●●●● ●● ● ● ●● ●●● ● ● ● ● ●
●●●● ●● ● ● ● ● ● ● ●● ●
●● ●●
●●●● ●●● ● ●● ● ● ●●
●●●● ●● ●●
● ●●●●● ● ●● ● ●● ●● ● ● ● ● ●● ●
● ● ● ●● ●
●●●● ●● ●● ● ● ●● ● ● ● ● ●
●● ● ●
●●●● ● ● ●● ● ● ●● ● ● ●● ● ● ● ● ●
●●●● ●● ● ● ● ● ● ● ●● ● ●● ● ● ●
●●●● ● ● ●●●● ● ● ● ● ● ● ● ● ● ● ●● ●
●●●● ●● ●● ● ● ● ● ●
● ●●●
● ● ● ● ● ●
●●●● ●● ● ●● ● ● ●● ● ● ● ●● ●●
●●●● ● ● ● ● ● ● ● ● ●● ● ● ●● ● ●
●●●● ●● ● ●● ●● ● ● ● ● ● ● ●●●● ● ● ●
●●●● ●● ● ●● ●● ●●
● ● ●●
●●●● ● ● ●●
●● ● ●●
● ● ● ●● ● ● ●● ●● ●●
●●●● ●● ● ●●● ● ●● ●●●● ● ● ●●
●
● ●●
● ● ● ● ● ●
●●●● ●●
●● ● ● ● ●
● ●● ●● ●● ● ●
●●●● ● ● ● ●● ● ●● ● ● ● ● ●● ● ●●●
●●●●●● ●●
● ●● ● ● ● ●
● ● ● ●●
● ●● ●
●● ●● ● ● ● ● ● ●
●●●● ●● ● ● ● ● ●●● ● ●● ● ●● ●● ● ●●
●●●● ●● ● ● ● ● ● ● ●● ●
● ●
● ●
●●●● ●● ● ● ● ●●
● ● ● ● ●● ● ● ● ● ● ● ● ● ●● ●
●●●● ●● ●● ●● ● ● ● ● ●● ● ●
●●●● ●●
●● ● ●●
● ● ●
● ●
●● ● ●●
● ●
●●●● ●● ● ● ● ● ● ● ● ●●
x
●● ●● ● ●● ● ● ● ● ● ●●
●●●● ●● ● ● ● ● ●● ● ● ●●
● ●● ● ● ● ●
●●●● ●● ●●
● ● ●
● ● ● ●● ● ●
●
● ● ●● ● ●
●
● ●
●
●
●●●●●● ●●
● ●● ● ● ●
●● ● ● ● ● ● ●
●
● ●
● ● ● ● ●
●●●● ●●●
●●
● ● ● ● ●● ● ● ● ● ●●●
● ● ●
●●●● ●● ●● ●
● ● ●●
● ●● ● ● ● ●● ● ●
● ●
● ●
●●●● ●●● ●● ●
● ●● ● ● ● ● ●● ●
●
●
● ● ● ● ● ● ●●● ● ● ● ●
●●●
●●●● ●● ●● ● ● ● ● ●● ● ● ● ●● ● ● ● ● ●● ●
●●●● ●●
●●● ● ● ● ●● ● ●● ● ● ●
●● ●● ●● ● ● ● ●●● ● ● ●
●● ●
● ● ● ● ● ●● ●
●●●● ●●
●●● ● ● ● ● ● ● ● ● ● ●
● ● ●
● ●
● ●
●●●● ●● ● ● ● ● ● ●
● ● ●● ● ●
● ●● ● ●
● ●●
●●●● ●●● ● ●
● ● ●
● ●
●
●●
● ●● ●
●●●● ●● ● ● ● ● ● ● ● ● ● ● ●
●● ●●● ● ● ● ●● ●● ●● ●
● ● ● ● ●● ● ● ● ●
●
●●●● ●●
●●● ●
● ●● ●● ● ● ● ● ●● ● ● ●●
●● ● ●
●● ● ● ● ●● ● ● ● ●●
●●●● ●●●
●●● ● ● ● ●● ● ● ●
●● ● ● ● ● ● ●
●● ● ● ● ● ●● ● ● ● ●●
●●●● ●●
●●● ●
●
● ● ● ●
●
● ● ● ● ●● ●● ●
● ● ●
●●●● ●●● ● ●● ● ● ● ● ● ● ● ●●● ● ●
●● ●●● ● ● ● ●● ● ● ● ●● ●
●● ● ●●
●●●● ●●● ● ●●● ● ● ● ●● ● ● ●●
● ● ● ● ● ●
●●●● ●●● ● ● ● ●
● ●
●
● ●●●
●
●●●
●
●
●
● ● ●
●●● ● ● ●
●●●● ●●● ● ● ● ● ● ●● ● ● ● ●
● ●
● ● ● ●● ● ●
●●
●●●● ●●● ● ● ●● ●
● ●●● ● ● ● ● ● ●● ● ● ●● ●●
●●● ● ● ● ●● ● ●
●●●● ●●● ● ●
● ● ●
● ●● ● ● ● ● ● ●
●●●● ●●●
●●●● ● ● ● ●● ● ●
● ● ●● ● ● ●● ● ●
0.4
●●●● ●●● ●
●
● ●● ● ● ●● ●● ● ● ●
●●●● ●●● ● ● ● ● ● ● ●
● ● ●● ●
●
●●●● ● ● ● ● ● ● ● ● ● ● ●●
●●●● ●● ● ● ● ● ●
● ● ● ●● ●
● ●●
●● ● ● ● ● ● ●●● ● ● ●●
● ●
● ●
●
●●●● ● ● ● ● ● ●● ●
● ● ● ● ●●● ● ●● ● ● ● ●● ● ●● ● ●
●●●● ●
●● ● ●
● ●
● ● ● ● ● ●● ●
●● ● ● ● ● ● ● ● ● ●
●●●● ● ●● ● ● ●● ● ●●● ●● ●●● ●
●● ● ●
●● ● ●● ● ● ●
● ● ●
● ● ● ● ● ● ● ● ●
●●●● ●● ● ● ● ● ● ● ● ● ●
● ● ●
●
●● ● ● ● ● ● ● ●● ● ● ● ●
●●●● ● ● ●● ●
●● ● ●
●●● ●
● ●
● ●
● ●
● ●
●●
● ● ●
●● ● ●●●● ●● ● ●● ●
●● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ●
●●●● ● ●
●●●
● ● ●● ● ● ● ● ● ●● ●
●● ● ●● ●● ● ● ●
●● ● ●
● ●●
● ● ●
● ●●● ●● ●
● ●●
●
● ● ●●
●●●● ●
● ●
●● ● ●
●●
● ●●
● ● ● ● ● ● ●●
●● ● ●● ● ● ● ●● ●
●● ● ●● ● ● ●●●
● ● ● ● ● ● ●● ● ●●
●
● ● ●
● ● ●●● ● ●
●● ● ● ● ● ● ● ●● ● ● ● ●
●●●● ● ● ●●
●● ● ● ●●
●● ● ● ● ● ●● ●
●● ● ●● ●● ● ● ●
●● ●● ● ● ● ●● ● ● ●● ● ● ● ● ● ● ● ●
●● ● ● ● ● ●● ●
● ● ● ● ● ● ● ●● ● ● ●
●● ● ● ● ● ● ● ● ● ● ●● ● ● ●
●●●● ●● ● ● ● ● ● ●● ● ● ●● ●●● ● ●● ●● ● ● ●
●●
●
●● ●● ●● ● ● ● ● ●● ● ●● ●● ● ● ● ● ●● ●
●● ● ● ●● ● ● ●
● ● ●
●● ●● ● ● ● ● ●● ●
●● ●● ●●● ●● ● ● ● ● ● ● ●● ●
●● ● ●● ●● ● ● ● ● ● ● ● ●●
●● ●● ●●● ●● ● ● ● ● ● ● ●● ● ● ● ● ●● ●
●● ●● ●●● ● ● ●● ● ● ●●
● ●
●● ●● ●● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ●
●● ●● ●● ● ●● ●● ● ● ● ● ●● ● ● ● ● ●
● ●
●
●● ●●● ● ● ● ● ● ● ●
●● ● ● ●●● ● ●
●
● ● ● ● ●
● ● ● ● ● ●● ● ● ●
●
● ● ● ●
●● ● ● ●● ● ● ● ● ● ●● ● ● ●
●● ● ●●
● ●● ●●● ● ● ●● ● ● ●
● ●
●● ● ● ● ●● ● ● ● ● ● ● ● ● ● ●
●● ● ● ● ● ● ● ●
●● ● ●● ● ● ● ● ● ● ●●
●● ●● ●●● ● ●●
● ●● ● ● ● ●● ●●
●● ●● ●
●● ●●● ● ● ● ●●
● ● ● ●● ●● ● ● ● ● ● ● ●●
●● ●
● ●●● ● ●●● ● ● ● ● ● ●● ● ● ● ● ●●●● ●
●● ● ●● ● ● ●● ● ●
●● ●●●●● ●
● ●● ● ●● ●● ● ● ● ● ● ●●
●● ●
● ● ● ● ● ● ● ● ● ●● ● ● ●
●● ● ●● ●● ● ● ●●
● ● ● ● ● ●●
●● ● ●
●● ● ● ●● ● ● ● ●
● ● ● ●
●●● ● ●● ● ● ●● ●
●● ●● ●● ● ● ● ● ● ●●
●● ● ● ●●
●● ●● ● ●● ● ● ● ● ● ●● ●
●
●● ●●●●● ● ●●● ● ● ● ● ● ● ● ●
●● ●● ●●●●
● ●● ● ● ● ● ●● ● ● ● ● ●
●● ● ●●● ● ● ●● ● ● ●● ● ● ● ●
●● ●●●● ●
● ●
●
● ● ● ●● ● ● ● ● ● ●
●
●
●
●● ● ● ● ● ● ● ● ● ●●
●● ●●● ●● ● ● ●● ● ● ● ● ● ● ● ●
●● ● ● ● ●●●● ●●● ●● ● ●● ●
● ● ●
●● ●●● ●● ●
●● ● ●● ●●
● ● ●
●● ● ●● ● ●
●● ●● ●● ● ● ●
●● ●● ● ●●
●● ● ● ● ●●● ● ● ● ● ● ● ● ● ●●● ● ● ●
● ●●
●●● ● ● ● ●● ● ● ●
● ●
●
●● ●●
●
●● ●● ● ● ● ● ● ●
● ●● ●● ●● ●
● ●
●● ● ●●● ● ●● ● ●● ● ● ● ● ●
●● ●●● ● ● ● ● ●
●● ● ●● ●
● ● ● ● ●●● ● ● ● ● ● ● ● ●●● ●
●● ●●● ●●
● ● ● ●● ● ● ● ●● ●
● ● ● ● ●
●● ●● ●● ● ● ●● ●
●● ● ● ● ● ● ●
●● ●●● ● ● ● ● ●●● ● ● ● ● ● ● ● ●
●● ●● ●●● ● ● ● ●●
● ● ● ● ● ●● ●
● ● ● ● ● ● ● ● ● ●● ●
●● ●● ● ● ● ● ●
●● ● ● ●● ● ●● ● ● ● ● ● ●
●● ● ● ● ● ● ●
●● ●●● ● ●● ● ● ● ●
●● ● ●●
● ● ● ● ● ●
●● ● ● ●● ● ● ● ●
●● ● ● ● ● ●● ● ● ●● ● ●
●● ●● ● ● ● ● ● ● ● ● ● ●
●● ● ● ●● ● ● ● ●● ● ● ● ● ●● ●
●
●
●● ●
● ● ● ●● ● ● ● ● ● ● ● ● ● ●
●●● ● ● ●● ● ● ● ● ●● ● ● ●●
●● ●● ● ● ● ●●
●● ●●● ●●● ● ● ● ● ● ●
●●
●● ● ● ● ●
●● ● ●●● ● ● ● ● ● ● ●
● ● ●● ● ● ● ●●● ●
●● ●● ● ● ● ● ●● ● ●●
●● ● ●● ● ● ● ●●
●● ●●
● ●● ● ●●● ●
● ●● ● ●
● ● ●● ●
● ●
0.2
●● ● ● ● ● ● ●● ● ● ●● ●●● ● ●
●● ● ●● ● ● ● ●
● ●● ● ● ●
● ●●
● ●●
●● ●● ●
●● ● ●● ● ● ● ● ● ● ● ● ● ●
●
●● ● ● ● ●● ● ● ● ● ●●
●● ●
● ● ● ●● ● ●● ● ● ●
●● ●● ● ● ● ● ● ●●● ● ●● ●
● ●
● ●● ●● ● ● ● ● ●●
●
●
●● ● ● ●
●● ●●●
●● ● ● ● ●●
● ● ● ● ●●
●●
●● ●● ● ●● ●● ● ● ●● ●
●● ● ● ●
●● ● ● ●● ● ● ● ● ●
●● ●● ●● ● ● ●● ● ●● ● ●
●
●●●
● ● ● ●●● ● ● ●
●● ●●● ●
●●●● ●● ● ● ● ●
●● ● ● ● ● ● ● ● ●
● ●
● ● ●● ● ● ● ●
●● ● ●● ● ● ● ● ● ● ●
●● ● ● ●●
● ●● ● ●
● ● ●
●● ● ● ●
●● ● ● ● ●
●● ● ● ● ●● ● ●
●● ● ● ●●● ● ● ●●●●
●● ● ●● ● ●
●●● ● ● ● ●
●● ● ●●● ●● ● ●● ● ●
●● ● ● ●●
●
●
●
●
●● ●
●
●● ● ●● ● ●● ●
●● ●● ●
●
●● ●● ●● ● ●●● ● ●
●
●● ● ● ● ●
● ● ● ● ● ●
● ●
●
●● ●●● ●● ●
● ●
●● ● ● ● ● ● ● ●
●● ● ● ● ● ● ● ●
●● ● ●
●
●●
●●●
● ● ●● ● ●
●
●● ● ●●● ●● ● ● ●●
●● ● ●● ● ● ●● ●
● ●● ● ●
● ●● ●
● ● ●● ● ●●● ●●
●● ● ●● ● ● ●
●●
● ●●●
● ● ●
●● ● ●● ● ● ● ● ● ●
● ●●● ●
● ●● ●
●● ● ● ● ● ● ●
● ●●
●
●● ●● ● ●●●● ● ● ●
● ●● ● ●
●● ● ● ● ● ● ● ●● ●●
●● ● ● ●
●● ●●● ● ● ● ● ●
●● ●●● ● ● ●
●● ● ●
● ●●● ●
● ●
● ●
●●
● ● ●
●● ● ●● ● ● ●
●●
●● ● ● ● ●● ● ●
●● ●● ●● ● ●● ●● ●●
● ● ●● ● ● ● ● ●
●● ●●●● ● ● ● ● ● ●
●● ● ● ●
●● ● ● ● ● ●
●●● ● ●● ●● ●● ● ●●
●
●● ●●
●● ● ●●● ● ●
●● ● ● ● ● ●
● ●●
● ●● ●● ●
●● ●● ● ● ●● ● ● ●
●● ● ●●
● ● ● ●
● ● ● ●● ● ● ●
●● ● ●
●● ● ● ● ●●
●● ● ● ● ●
●
●● ●●●●● ● ● ● ●● ●
●●● ●● ●● ● ●● ●●
●● ● ● ● ● ●
● ● ● ●
●● ●●●
● ● ●● ● ●
●● ●● ●● ● ● ●
● ●
●● ● ● ●● ● ●
● ● ●
● ● ●●
●
●● ● ● ●
●
●
●● ● ●● ● ●
●● ●● ● ● ● ● ●
●●● ●● ● ●
●● ● ●● ●
● ● ●● ● ● ●
●● ●●● ● ●
●
●● ●●
●
● ●● ●
● ● ● ● ● ●●
● ●● ● ●●
● ●● ● ● ●
●● ●
●●● ● ● ●
0.0
● ● ●
● ●●
● ● ●
●● ● ●● ●
●● ●●
●● ● ● ●● ●
●
● ● ●● ●
● ● ●●
●● ● ●
● ● ●
●● ●● ●
● ● ●
● ● ●
● ●
● ●● ●
● ● ●●
● ●
● ●
● ●
●
●
a
Figure 3.4: Feigenbaum diagram of the Verhulst equation.
Dörner and Wearing propose a systematic approach for problem solving [40,
pp. 67–70]: First, a concrete and measurable goal has to be formulated. Then, a
mental representation of the system has to be gained. Based on a prediction of
the system’s future development, own actions have to be planned. After these
actions are conducted, their effects have to be assessed in relation to the planned
goal. Depending on the outcome of this assessment, some of the previous steps
have to be redone in a modified manner.
Some important empirical results for real behavior of humans are presented by
Dörner and Wearing in [40, pp. 72–83]. It can be noticed that humans in general
have big problems when solving complex problems and that they make some
typical mistakes.
Dörner gives a rather informal overview of this field of research in [39].
goal,
mental image of
(set of possible situation
questions/actions)
questions, actions
problem/situation
information
(can be person,
correct, fuzzy, too late) group of persons,
internal parameters
machine or
agent
dynamical changes over time
with resources:
observable / not observable
time, tools, skills,
knowledge, organization,
influenceable / not influenceable
experience, ...
Figure 3.5: Complexity between problem/situation and agent [141, p. 30]. “Complexity
is in the eye of the beholder.”
studied exhaustively (as the only clue to a crime) may present a very
great quantity of significant detail.
Without further justification, I shall follow, in this paper, an interpre-
tation of ‘complexity’ that I have used and found suitable for about
ten years. I shall measure the degree of ‘Complexity’ by the quantity of
information required to describe the vital system. To the neurophysiologist
the brain, as a feltwork of fibers and a soup of enzymes, is certainly
complex; and equally the transmission of a detailed description of it
would require much time. To a butcher the brain is simple, for he has
to distinguish it from only about thirty other ‘meats’, so not more than
log2 30, i. e. about 5 bits, are involved. This method admittedly makes
a system’s complexity purely relative to a given observer; it rejects tha
[sic] attempt to measure an absolute, or intrinsic, complexity; but this
acceptance of complexity as something in the eye of the beholder is,
in my opinion, the only workable way of measuring complexity.”
The observer on the other hand has a set of resources (e. g., time, tools, skills,
knowledge, organization, experience) and of possible actions. The observer can
gather information about the problem/situation. This information can be correct
or incorrect, fuzzy or simply coming too late.
Based on the observer’s developing mental image of the situation, he/she can
try to reach his/her goal by conducting adequate actions.
According to this model, complexity is a somewhat “subjective” property of a
problem or situation—mainly influenced by the observer’s resources.
In his blog entry Complexity is in the Eye of the Beholder [173], Weber explicates
this thought that complexity depends on the observer’s resources:
“The question is, where is the boundary between complex and simple?
I think that this border only exists in the human mind. In fact the
border is movable and depends on the individual mind.
Nothing is complicated to Nature (the universe). Everything just is.
Confusion and complication only arise in the human mind while
trying to understand the universe.
Complexity is a function of our capacity to assimilate, store and
process data. If we were suddenly twice as smart, some complicated
problems would seemingly become simple problems. The problems
themselves did not change, only we changed.”
Consequences
Already some authors of process measurement literature stated that there are
different aspects of process model complexity. This statement was supported
by the presentation of different aspects of general complexity. These complexity
aspects can also be analyzed for process models:
• Product complexity
process portfolio of a company, interacting IT infrastructure
• Networks
process models seen as networks, communication network of persons in-
volved in a process
• Dynamical systems
changing state during process model execution
This fact suggests that it is at best possible to measure one of these complexity
aspects and not process model complexity in total.
Furthermore, it was shown that complexity can be seen as something which
“originates” through the connection between the problem and its observer instead
of an intrinsic property of the problem itself. Because of this, it is even hardly
possible to find a complexity metric for a special complexity aspect. The reason
that this works so well in computational complexity theory is the fact that here,
the “observer” with its “resources” can be mathematically defined by using
Turing machines. For most of the other presented complexity aspects—especially
those with human interaction involved—, such a stringent formal definition is
impossible—resulting in a rather vague complexity concept.
Definition 3.4 (Process model quality) For a customer, process model quality means
that the process model’s outcome (a product, a piece of information or a decision) is correct,
arrives within adequate time and to an adequate price.
For a company, all these factors also belong to quality—but additionally, the price
for the process model execution must be lower than the price which the customer is
willing to pay and, furthermore, the process model should be easily adaptable to changed
circumstances.
In this section, the measurement approach used throughout this thesis is intro-
duced.
The foundation of this approach is the observation made in the previous section
that it is impossible to find a stringent definition of process model complexity.
The reason is that this term has so many different aspects as has been shown in
Subsection 3.3.1.
Instead, a prediction system measurement approach which avoids concrete
definitions of process model complexity and quality is used. It is adapted from
software measurement where a similar problem (definition of software complex-
ity) exists.
Before this measurement approach can be presented, measurement and pre-
diction systems have to be explained in Subsection 3.4.1. Afterwards, the actual
measurement approach is introduced (Subsection 3.4.2). Finally, the necessary
validation steps for such a prediction system are shown in Subsection 3.4.3.
Definition 3.6 (Prediction system) A prediction system is used to predict some at-
tribute of a future entity, involving a mathematical model with associated prediction
procedures.
Besides the use for future entities, as stated in the definition of Fenton and
Pfleeger, prediction systems can also be used to predict some attribute of an
existing entity which is measurable only in a very laborious manner.
from Lewis Carroll’s novel Through the Looking-glass and What Alice Found There
[25, p. 114]. According to Weinberg and Weinberg, these authors simply define
“complexity in such a way that it does exactly what you want it to do” [174,
p. 313].
Therefore, an alternative measurement approach is suggested here. It is in-
spired from software measurement where a similar dilemma concerning the
definition of the term “software complexity” exists. There, a prediction system
46 process measurement
ha
has s
costs
PROCESS MODEL
duration
number of errors
internal external
possible reasons complexity possible implications changeability
attributes attributes
flexibility
pr
oc el
e od res understandability
m ss m m
et s s asu
ric od dependency e e
s e oc m
l pr lity
a
process model process model qu
metrics quality measures
values predictability values
measurement approach helped to overcome this problem (see, e. g., [43, pp. 74–
75]). The measurement approach can be adapted to process measurement as
described in the remainder of this subsection.
What is more important than process model complexity itself—especially for
economic reasons—are the implications of this complexity like costs, time, duration,
number of errors, changeability, flexibility, understandability, etc. (aspects of
process model quality according to Subsection 3.3.2). All these quantities have the
advantage to be quantifiable and measurable.8 The disadvantage is that they can
only be measured after the process model has been implemented and executed.
To overcome this problem, the adaptation of the prediction system measure-
ment approach from software measurement comes into play. It is sketchily de-
picted in Figure 3.6.
A process model has internal and external attributes.
Internal attributes can be measured purely in terms of the process model, sepa-
rately from its behavior [43, p. 74]. These attributes (e. g., structural properties like
the number of activities/tasks) could contribute to the process model complex-
ity. Numerous internal attributes are imaginable and appropriate metrics have
already been proposed (especially for structural properties) or can be defined.
Using these metrics, one gets corresponding metric values of the process model.
External attributes can be measured only with respect to how the process
model relates to its environment [43, p. 74]. The external attributes like costs,
time, number of errors, etc. are possibly affected by the process model complexity
and are measurable. External attributes are aspects of process model quality (and
performance respectively).
The last step of the proposed approach is also the most important one: One
has to show a dependency between the metric values and the measure values
of the external attribute. If such a dependency exists, the metric can be used as
a predictor for the external attribute at a much earlier time. Thus, a prediction
system is formed.
8 For, e. g., costs, time, duration, number of errors, this is trivial to see. But also attributes like
changeability, flexibility and understandability are measurable if one looks at the costs, time,
number of errors, etc. it takes to change or understand a process model. Fenton and Pfleeger give
some ideas for measuring maintainability in [43, pp. 354–355].
3.4 process measurement approach 47
3.4.3 Validation
9 Note that some metrics from Section 3.2 which are called “process quality metric” by their authors
(e. g., in [23]) are process model metrics in the nomenclature of this thesis.
48 process measurement
• validity: According to Kan [69, pp. 71–72], validity can be classified into
construct validity and content validity. The first checks whether the met-
ric/measure really represents the theoretical concept to be measured (e. g.,
is church attendance a good measure for religiousness?). The second checks
whether the metric/measure covers the range of meanings included in the
concept (e. g., a test of mathematical ability for elementary pupils cannot
be limited to addition but should also include subtraction, multiplication,
division and so forth).
The validation can be done either by using existing data (e. g., from log files) or
by conducting experiments (to get new data). Fenton and Pfleeger emphasize the
advantages of experiments as the level of control and the level of replication are
much higher [43, p. 120]. Basics about empirical investigations (e. g., experimental
design among other things) can be found in [43, pp. 117–152] and Section B.3.
As there can be different kinds of dependencies (e. g., positive linear, nega-
tive linear and many forms of non-linearity) [69, pp. 77–80] (see Figure 3.7 for
examples), scatter plots are a good method to visually search for any form of
dependency (also non-linear). The next step is to use a measure of correlation like
Spearman’s rank correlation coefficient (see Section C.2) or Pearson’s product-
moment correlation coefficient (see Section C.1). If a dependency is found, one
can also try to find an equation which mathematically describes the dependency
(e. g., using linear regression, multivariate regression, non-linear regression). [43,
pp. 199–200]
In the field of software measurement, IEEE Standard 1061 (IEEE Standard for a
Software Quality Metrics Methodology) gives a method for validating prediction
systems [60, pp. 10–13] which checks among correlation also additional properties
as tracking, consistency, predictability, discriminative power and reliability10 .
measurement dimensions Prediction systems are only valid for very spe-
cial conditions. According to Fenton and Pfleeger, “validation must take into
account the measurement’s purpose; a measure may be valid for some uses but
not for others” [43, p. 107].
Consequently, the conditions during validation and the later use of the predic-
tion system must be consistent. The following four “measurement dimensions”
are generally important conditions. For special cases, additional conditions may
exist.
• Subjects
Which persons are involved in the measurement? Possible persons are, e. g.,
process designers, process analysts, programmers and end-users (i. e., the
employees working in the process). As these persons have different skills
and different views of the process, the values of the same external process
model attribute (e. g., time, costs and number of errors) can differ a lot
depending on the involved persons (subjects).
10 In [60], “reliability” has another meaning than the homonymous requirement for valid measure-
ment systems presented earlier in this subsection.
50 process measurement
●● ●
50
50
●
●
● ●
●●
●●
●●
● ●
40
40
● ● ●
● ● ● ●
● ●
● ●
●
● ● ●
●
●
●
● ●
30
● ●
30
●
● ● ●
● ●
● ●
● ●
● ●
● ● ●
y
y
●
●
● ●
●
● ●
●
20
●
20
● ●
● ● ●
● ●
●
●
●● ●
●
● ● ●●
● ●
10
●
10
● ● ●
●
●
●●
●
●●
●
● ●
● ● ●
0
●
● ●
0
0 10 20 30 40 50 0 10 20 30 40 50
x x
(a) (b)
70
●● ● ●
● ●
●
● ● ● ●
60
● ● ●
● ●
●
60
● ● ●
●
● ● ●
●
●
●
● ●
50
●●
● ●● ●
50
●●
● ●
●
●
40
●
40
● ●●
●
●
●
● ●
y
● ●
● ●
30
30
● ● ●
●
● ●
● ●
● ●
20
●
20
●● ●
●
● ●
● ● ●
●
● ●
●●
10
10
● ●
● ● ●
●● ●
●
● ● ● ●
●
● ● ● ●
●● ●
●
0
● ● ●
0
0 10 20 30 40 50 0 10 20 30 40 50
x x
(c) (d)
Figure 3.7: Four possible types of dependencies between two variables [69, p. 79].
• Process phase
As in software engineering, a process life cycle consists of several process
phases: modeling, analysis, implementation, deployment, execution, main-
tenance and modification of the process model (cf. the BPM lifecycle in
Subsection 2.1.2).
In contrast to software engineering, process model execution is an addi-
tional phase in which problems can occur. After a software program is
implemented, no new errors are introduced by executing the program. But
as process models are executed (at least partially) by humans, additional
errors can occur while executing a process model.
3.5 application of process measurement 51
Having established valid measurement and prediction systems for process models,
the question arises what to do with these metrics.
In this section, several possible applications of process measurement are pre-
sented. It can be used both for process models that are newly implemented and
for finding and dealing with “those existing processes11 that are good candidates
for improvement and simplification, or even complete reengineering” [74, p. 3].
As there exist numerous metrics for process models, first, one has to select proper
metrics for the considered “problem”. Using all available or accidentally selected
metrics would just generate numerous numerical values without any purpose for
the considered “problem”.
Basili et al. propose an approach for the selection of metrics for software
measurement—the Goal Question Metric (GQM) approach [9]. This approach is also
applicable for process measurement and can be used both for selecting process
model metrics and process model quality measures.
The approach has three levels: conceptual level (goal), operational level (ques-
tion) and quantitative level (metric). At the first level, a precise goal is defined. A
set of questions for assessing and achieving the goal is established at the second
level. At the third level, a set of metrics is assigned to each question in order to
quantitatively answer the questions. The resulting GQM model has a hierarchical
structure with possibly several goals, multiple questions per goal and several
metrics per question. A metric can be assigned to multiple questions. Figure 3.8
shows an example for such a hierarchical GQM model structure.
Using this top-down approach, only useful metrics, measures and possibly
prediction systems for the current “problem” are selected and no unnecessary
metric/measure values are collected.
For the field of software measurement, Fenton and Pfleeger mention three differ-
ent measurement purposes [43, pp. 13–14] which can also be adopted to process
measurement.
goal 1 goal 2
Understand
For this first purpose, a process model is only measured using different selected
metrics and/or measures to get a better understanding about what happens
within this process model. Afterwards, no changes or concrete actions are con-
ducted. Through this, the process model can be compared while being modified
over time (modifications not caused by process measurement!) or it can be com-
pared with other process models within the same company.
For this purpose, only valid measurement systems are necessary.
Control
Here, a process model is also measured, but not with the general goal to change
it. Instead, the findings of the measurement are used to manage and control the
assignment of employees in testing and finding errors. Consequently, the limited
manpower can be deployed more intensively in “problematic” process models
or process model parts. Possible actions can comprise, for example, test cases or
inspections as in software engineering. A process model is only changed in order
to fix a found error.
For this purpose, valid prediction systems are necessary.
Improve
For the third purpose, a process model is measured. If a bad quality is measured
(for existing process models) or predicted (for new process models), the process
model will be changed in order to improve it. So, the goal is to reduce unnecessary
complexity within the process model. As complexity itself is not measured in
the process measurement approach used in this thesis (see Subsection 3.4.2),
reducing complexity means changing the process model’s internal attribute(s) in
such a way that the external attribute(s)—the process model quality—increases
according to the prediction system.
One has to consider that the complexity of a process model cannot be reduced
arbitrarily [19, p. 117]. Here, one must distinguish between the intrinsic complexity
3.6 assessment of existing work 53
of a process12 and the complexity of a process model. The chosen process model is
not independent of the overall problem. So, it has a “natural” minimal complexity.
This fact was already referred to by Fenton and Pfleeger for software measurement
[43, p. 267].
One can compare this with an example of an analogous problem—runtime
complexity of algorithms: The general problem of sorting has a (mathematically
proven) minimal complexity of Ω(n log n) [32, pp. 165–167]. The Heapsort sorting
algorithm, for example, has complexity O(n log n) [32, pp. 127–137]. But never-
theless, more inefficient sorting algorithms exist (e. g., Insertion Sort with O(n2 ),
see13 ).
But even if a reduction of complexity is possible and would probably cause
higher quality, one should first compare the costs for the process model change
with the expected increase of incomes with this process in order to decide whether
to actually implement the changes.
As the quality for the changed process model is predicted within this purpose,
valid prediction systems are necessary.
• Mendling et al.: Using process model metrics for predicting faulty EPCs
[96–100]
604 EPC process models of the SAP Reference Model were analyzed using
the verification tool WofYAWL. Through this, 34 faulty process models were
12 The use of the term “intrinsic complexity” may be confusing here as in Subsection 3.3.1, it
was stated that complexity is no intrinsic property of a problem—yet, it “originates” between
the problem and its observer. Here, “intrinsic complexity” is meant as the “natural” minimal
complexity in this “subjective” meaning.
13 According to [32, pp. 23–27], Insertion Sort has worst-case running time of θ(n2 ). Using Theo-
rem 3.1 in [32, p. 46], running time O(n2 ) follows.
54 process measurement
Thus, the assessment had the same result as Sánchez González et al. already
noticed [136, p. 124]: In the published literature, one can find a strong tendency to
create and propose new metrics and measures without any validation. In future
research, more attention should be paid on the empirical validation of existing
proposals instead of defining new ones.
3.7 conclusion
In this chapter, the state of the art in process measurement was presented and
the theoretical framework for process measurement used in this thesis was
introduced.
For this purpose, an overview of publications on process measurement was
given. Many proposed process model metrics are adapted from software metrics
and are claimed to measure process model complexity, quality and/or perfor-
mance. It could be observed that there are no concrete definitions of process
model complexity and process model quality in the literature. Often, both terms
are even used as synonyms.
Thus, a discussion of these terms followed. It could be shown that there does
not exist a single formal definition of complexity. Instead, numerous aspects of
complexity were identified and are analyzed in different research communities.
Consequently, it is problematic to say that a process model metric measures the
complexity of a process model.
The main contribution of this chapter is a theoretical framework for process
measurement in which the existing work can be integrated and which can help
to identify open research questions leading to new research directions in process
measurement.
For this, the more well-established concepts from software measurement were
adopted for process measurement: The result was a prediction system mea-
surement approach, which is based on measurement and prediction systems.
The measurement approach consists of process model metrics measuring (struc-
tural) internal attributes and process model quality measures measuring external
process model attributes. Through this, a concrete definition of process model
56 process measurement
Chapter 3 showed that numerous process model metrics have been proposed in
literature in the previous years. Nevertheless, it had to be noticed that only a
small minority of them are part of a validated prediction system.
A proper validation would require controlled experiments (see Appendix B),
which are very time and cost intensive. This fact—together with the possibility
of a negative outcome of the experimental validation—could explain the small
number of existing validated prediction systems.
In this chapter, an approach is proposed which shall help to ease this problem
by reducing the experimental effort for
• unsuccessful validations or
s ha
ha s
costs
PROCESS MODEL
duration
number of errors
internal external
possible reasons complexity possible implications changeability
attributes attributes
flexibility
pr
oc el
e od es understandability
m ss m m ur
et dependency ess eas
ric od oc m
s e l pr lity
a
process model process model qu
metrics quality measures
values predictability values
Figure 4.1: Analysis step visually located within the measurement approach of Subsec-
tion 3.4.2.
In this section, the general approach for analyzing process model metric proper-
ties is introduced independently from specific process model metrics and process
model collections.
In doing so, one has to distinguish between general and process model collec-
tion specific properties. General properties (Subsection 4.2.1) hold independently
from the considered process models just because of the metric’s definition. Process
model collection specific properties (Subsection 4.2.2) are true for the examined pro-
cess models—but not implicitly generalizable to other process model collections.
Scale Types
Next, the scale type (see Section A.2) of each process model metric is identified.
The following process model collection specific properties are examined by the
approach.
Descriptive Statistics
In this step, the value distributions of the different process model metrics are to
be analyzed.
In order to get a compact quantitative overview, the following descriptive
statistics are computed for each process model metric:
• minimum, 25% quantile [117, pp. 25–26], median [117, pp. 19–21], 75%
quantile [117, pp. 25–26] and maximum (for metrics on ordinal scale) as
well as mean [117, pp. 16–17] (for metrics on interval scale) as measures of
location,
• range (R) [117, p. 22], interquartile range (IQR)3 [156, p. 55], median absolute
deviation (MAD)4 (for metrics on ordinal scale), standard deviation (sd)
[117, p. 22, 35–36] (for metrics on interval scale) as well as coefficient of
variation (CV) [117, pp. 33–34] (for metrics on ratio scale) as measures of
dispersion.
2 In Table A.2, only those statistical operations which are used in the approach for analyzing
process model metric properties are listed.
3 difference between 75% and 25% quantiles
4 median of the set of absolute values of the differences between the values and the median of
these values
60 analysis of process model metric properties
As can be seen in the brackets, some statistics require that the metric has a
special scale type. Otherwise, the resulting number would be meaningless. See
Section A.2 for details.
If one is interested in more than just some compact quantitative numbers,
the value distributions of the process model metrics can be visualized using
histograms. Through this, one gets more detailed information.
Correlations
In the next step, possible correlations (see Appendix C) between process model
metrics are examined.
For that purpose, Spearman’s rank correlation coefficient (Section C.2) can be
computed. It is a measure of correlation between two variables assessing how well
an arbitrary monotonic function could describe the relationship. The two variables
must be measured at least on an ordinal scale. Their random distribution is
unimportant.
The second statistic, empirical Pearson’s product-moment correlation coefficient
(see Section C.1), measures the linear relationship between two variables. These
variables must be measured at least on an interval scale and must be nearly
normally distributed.
In order to identify also non-linear dependencies between process model
metrics, the scatter plot matrix can be analyzed. Yet, the resulting n × n matrix for
a larger number n of analyzed process model metrics soon becomes unhandily
large.
In this section, the abstract approach presented in the previous Section 4.2 is
applied to a set of selected process model metrics and a process model collection.
In a first step, a set of process model metrics and a process model collection have
to be selected.
As every process model metric is only applicable to some specific process
modeling languages (see Section 2.3), the following two requirements have to be
fulfilled by this selection:
So, one has to choose a process modeling language for which both numerous
process model metrics and a large process model collection exist. EPCs (see
Subsection 2.3.1) are such a language.
Table 4.1: Selected process model metrics for EPCs (Part 1 of 2).
name symbol reference definition
number start events SES [96, 99]
number internal events SEInt [96, 99]
number end events S EE [96, 99]
number events SE [99] SE (G) := |E| = SES (G) + SEInt (G) + SEE (G)
number functions SF [96, 99] SF (G) := |F|
number AND splits SSAND [96, 99]
number AND joins SJAND [96, 99]
number XOR splits SSXOR [96, 99]
number XOR joins SJXOR [96, 99]
number OR splits SSOR [96, 99]
number OR joins SJOR [96, 99]
number connectors SC [99] SC (G) := |C| = SSAND (G) + SJAND (G) + SSXOR (G) +
SJXOR (G) + SSOR (G) + SJOR (G)
number nodes SN [99, 100] SN (G) := |N| = SE (G) + SF (G) + SC (G)
number arcs SA [96, 99] SA (G) := |A|
diameter diam [99] length of the longest path (= number of arcs on this path) from
a start event to an end event
|A|
density (1) ∆ [99] ∆(G) := |N|·(|N|−1) : number of arcs divided by the maximum
number of arcs for the same number of nodes
density (2) D [97] D is a second and more complicated density metric with respect
to the EPC syntax constraints. See [97, pp. 3–4].
|A|
coefficient of connectivity CNC [74, 99] CNC(G) := |N|
|A|2
coefficient of network CNCK [74] CNCK (G) := |N|
complexity
cyclomatic number CN [74] CN := |A| − |N| + 1: number of linearly independent cycles
(arc directions are ignored)
In this subsection, the results concerning the general properties are presented.
6 no start event, no end event, a function with not exactly one predecessor and one successor node,
an event with more than one predecessor or successor node or several graph components
4.3 experimental application 63
Table 4.1: Selected process model metrics for EPCs (Part 2 of 2).
name symbol reference definition
1 P
avg. connector degree dC [99] dC (G) := |C| c∈C d(c) (see a )
a
max. connector degree dC [99] dC (G) := max{d(c)|c ∈ C} (see )
d d
|{n∈N|n is cut-vertex}|
separability Π [99, 100] Π(G) := |N|−2 : A cut-vertex is a node whose
deletion separates the process model into multiple compo-
nents.
|A∩((E∪F)×(E∪F))|
sequentiality Ξ [99, 100] Ξ(G) := |A| : number of arcs between non-
connector nodes divided by the number of arcs
depth Λ [99] Depth relates to the maximum nesting of structured blocks in
a process model. See [99, pp. 124–125].
mismatch MM [99] MM(G) :=
P P P
c∈Sl dout (c) − c∈Jl din (c) (see
l∈{AND,XOR,OR}
b ):
sum of mismatches for each connector type
P a ):
heterogeneity CH [99] CH(G) := − l∈{AND,XOR,OR} p(l) · log3 p(l) (see en-
tropy over the different connector types
|N |
C
cyclicity CYC [99, 100] CYCN (G) := |N| : number of nodes NC on a cycle (arc
directions are not ignored) divided by the number of nodes
P
token splits TS [99] T S(G) := c∈SAND ∪SOR (dout (c) − 1): number of newly
introduced tokens by split connectors
P P
control flow complexity CFC [21, 96, 99] CFC(G) := c∈SAND 1 + c∈SXOR dout (c) +
P
dout (c) − 1 : sum over all split connectors
c∈SOR 2
weighted by their number of possible states after the split
P P
join complexity JC [96] JC(G) := 1 + c∈JXOR din (c) +
P c∈JAND
2 din (c) − 1 : sum over all join connectors weighted
c∈JOR
by their number of possible states before the join
weighted coupling CP [166] CP measures the average coupling between all pairs of con-
nected events and functions—weighted according to the type
of connection. See [166, p. 42].
cross-connectivity CC [167] CC measures the average strength of connection between all
pairs of nodes. See [167, pp. 483–484].
a Metric value is 0 for |C| = 0 (source: personal communication with Jan Mendling).
b The original definition printed in [99, p. 125] is faulty (source: personal communication with Jan
Mendling).
Proof. The theoretical value ranges and the sets of numbers have to be proved.
As the determination of the set of numbers to which the process model metrics’
values belong is trivial when one looks at their mathematical definitions, no
detailed proof for this aspect is given here. Proofs are only presented for the value
ranges. In doing so, some results presented later in this subsection (paragraph
Behavior for Specific Process Models and Correlations Based on Process Model Metric
Definitions) are already used here.
Regarding “size metrics”: All “size metrics” (metric SES till SA in Table 4.1)
measure the number of objects of a specific type in an EPC. So, the values cannot
be negative. There is no maximal limit for the metrics. Only some metrics have a
64 analysis of process model metric properties
Table 4.2: Theoretic value ranges and scale types of the selected process model metrics.
• lower bound: The smallest process model SEQ-1 has the minimal longest
path from a start to an end event with diam value 2.
• upper bound: According to Table 4.3, the diam value becomes arbitrarily
large for SEQ-n process models. So, there is no maximal value.
Regarding metric ∆:
• lower bound: Metric ∆ measures a ratio between existing arcs and possible
arcs. As an EPC has at least two arcs, this ratio cannot be negative and
4.3 experimental application 65
cannot be 0. According to Table 4.3, the ∆ values for the SEQ-n converge to
0. So, 0 (excluded) is the minimal possible value.
• upper bound: The ∆ value of a SEQ-1 (with three nodes) is 13 . An EPC with
6
2− S 2− 56
N (G)
four nodes is impossible. According to (4.25), ∆(G) 6 SN (G)−1 6 5−1 = 0.2
1
for SN (G) > 5. So, 3 is the maximal value.
Regarding metric D:
• lower bound: According to its definition, metric D has value 0 for EPCs G
with SC (G) 6 1 (this includes the sequential process models SEQ-n). For
the remaining EPCs, the original definition in [97] can be transformed into
SA (G) − SN (G) + 1
D(G) := . (4.1)
cmax + SN (G) − 2SC (G) + 1
The nominator equals metric CN with CN(G) > 0 (see equation (4.12)). For
the denominator (shortly written here as D 0 (G)),
2
SC (G)
2
+ 1 + SN (G) − 2SC (G) + 1 for SN (C) even
2
D 0 (G) = SC (G)−1 S (G)−1
+1 + C 2 +1
2
for SN (C) odd
+SN (G) − 2SC (G) + 1
1 2
4 (SC (G)) − SC (G) + SN (G) + 2 for SN (C) even
= 1
(4.2)
2 − SC (G) + SN (G) + 1.75 for SN (C) odd
4 (SC (G))
1
(4.20)
4 (SC (G))
2 − ( 23 SN (G) − 2) + SN (G) + 2 for SN (C) even
> 1
4 (SC (G))
2− ( 23 SN (G) − 2) + SN (G) + 1.75 for SN (C) odd
1 2 1
4 (SC (G)) + 3 SN (G) + 4 for SN (C) even
= 1 1
2
4 (SC (G)) + 3 SN (G) + 3.75 for SN (C) odd
> 0
holds. So, the D values cannot be negative and value 0 is the minimal one.
1
SA (G)−SN (G)+1
2 −S (G)+S (G)+2 for SN (C) even
(SC (G)) C N
D(G) = 4
1
SA (G)−SN (G)+1
for SN (C) odd
2
4 (SC (G)) −SC (G)+SN (G)+1.75
(4.18) 1
(2SN (G)−6)−SN (G)+1
2 for SN (C) even
6 4 C (G)) −SC (G)+SN (G)+2
(S
1
(2SN (G)−6)−SN (G)+1
for SN (C) odd
2
4 (SC (G)) −SC (G)+SN (G)+1.75
SN (G)−5
for SN (C) even
1
2
(SC (G)) − SC (G) +SN (G)+2
4
| {z }
>−1 for SC (G)>2
=
1
SN (G)−5
for SN (C) odd
2
(SC (G)) − SC (G) +SN (G)+1.75
4
| {z }
>−0.75 for SC (G)>3
SN (G) − 5 6 SN (G)→∞
6 = 1− −→ 1 .
SN (G) + 1 SN (G) + 1
The D metric values for the process models AND-n converge to 1 (see
Table 4.3). So, 1 (excluded) is the maximal value.
• lower bound:
SA (G) (4.17) SN (G) − 1 1 1 2
CNC(G) = > = 1− > 1− =
SN (G) SN (G) SN (G) 3 3
| {z }
SN (G)>3
• upper bound: The CNC value of the smallest process model SEQ-1 (three
nodes) is 23 . An EPC with four nodes is impossible. For all remaining nodes,
• lower bound: According to the proof of (4.22), CNCK (G) > SN (G) − 2 +
1
SN (G) holds. The minimal value of the right side of this inequality is reached
for SN (G) = 3. In that case, the inequality can be further transformed into
CNCK (G) > 3 − 2 + 13 = 1 13 . SEQ-1 has this minimal CNCK value.
• upper bound: According to Table 4.3, the CNCK value becomes arbitrarily
large for SEQ-n process models. So, there is no maximal value.
4.3 experimental application 67
• upper bound: Following Table 4.3, the CN value becomes arbitrarily large
for AND-n process models. So, there is no maximal value.
• upper bound: Following Table 4.3, the values of both metrics become arbi-
trarily large for AND-n process models. So, there is no maximal value.
Regarding metric Π:
• lower bound: The ratio cannot become negative. In every EPC, a start event
must be connected with a non-event node as its only neighbor. Consequently,
this node is a cut-vertex and the ratio cannot become 0. According to
Table 4.3, the Π values of the AND-n process models converge to 0. So, 0
(excluded) is the minimal value.
• upper bound: As an EPC has at least one start and one end event and both
cannot be a cut-vertex, the ratio cannot become larger than 1. The sequential
process models SEQ-n have a value of exactly 1 (see Table 4.3).
Regarding metric Ξ:
• lower bound: As metric Ξ measures the ratio between the arcs connecting
non-connector nodes and all arcs, the value cannot become negative. The
AND-n process models have a value of 0 (see Table 4.3).
• upper bound: The ratio cannot become larger than 1. The sequential process
models SEQ-n have a value of 1.
• lower bound: Because of their definitions, both metrics cannot have negative
values. The sequential process models SEQ-n have value 0 for both metrics.
• upper bound: As there are no maximal nesting depth and no maximal sum
of mismatches, the values of both metrics can become arbitrarily large.
68 analysis of process model metric properties
∂f ln (1 − p1 − p2 ) − ln p1
= (4.5)
∂p1 ln 3
∂f ln (1 − p1 − p2 ) − ln p2
= . (4.6)
∂p2 ln 3
For extrema,
∂f ∂f !
= =0 (4.7)
∂p1 ∂p2
must hold. This is only true for p1 = p2 = p3 = 13 . Here, the function f—and
consequently metric CH—has its maximal value 1.
• upper bound: Imagine an EPC G with one start event followed by an XOR
join. The XOR join is following alternately by n functions and n events
until the XOR join is reached again (forming a cycle!). Between one of the
functions and events on the cycle, there is an XOR split inserted. The second
outgoing arc of this split is connected with an end event. So, there are 2n + 2
of the total 2n + 4 nodes on the cycle. Consequently,
2n + 2 n+1 1 n→∞
CYC(G) = = = 1− −→ 1
2n + 4 n+2 n+2
holds. As at least one start and one end node cannot lie on a cycle, value 1
is not reachable. The ratio cannot become larger than 1. So, 1 (excluded) is
the maximal value.
4.3 experimental application 69
• lower bound: Because of their definitions, negative metric values are im-
possible. The sequential process models SEQ-n have value 0 for all three
metrics (see Table 4.3).
• upper bound: According to Table 4.3, the values of all three metrics become
arbitrarily large for OR-n process models. So, there is no maximal value.
• lower bound: As the arc weights cannot be negative, the CP metric values
can also not become negative. Each EPC has at least two arcs and their
weights are larger than 0. The metric values of the sequential process models
SEQ-n converge to 0 (see Table 4.3), So, 0 (excluded) is the minimal value.
• upper bound: The sequential process model SEQ-1 has CP value 13 . This is
the smallest possible EPC. There is no EPC with four nodes. For SN (G) > 5,
the following considerations are true. An EPC G has at most 2SN (G) − 6 −
SC (G) pairs of non-connector nodes (maximal number of arcs of G minus
number of connectors as there is one arc per connector more than pairs of
non-connector nodes next to the connector) which each can have a maximal
weight of 1. Consequently,
21
holds. This fraction has its maximal value 52 ≈ 0.404 for SN (G) = 10.7
be negative. Consequently, the CC values are larger than 0. For the pro-
cess models AND-n, the metric values converge to 0 (see Table 4.3). So, 0
(excluded) is the minimal value.
Scale Types
The scale types of the selected process model metrics are listed in the third
column of Table 4.2.
Most metrics are ratio scales. But there are also some exceptions:
• For heterogeneity (CH), value differences and ratios are meaningless because
of taking the logarithm. So, the metric is an ordinal scale.
• For density (D), weighted coupling (CP) and cross-connectivity (CC), things are
a little unclear. To err on the side of conservatism, these metrics are listed to
be ordinal scales.
n functions
function 1 n functions
start event
V V end event
function n
• SEQ-n (n > 1): sequential EPC with n functions (see Figure 4.2a),
• AND-n (n > 2): EPC with n parallel (AND connectors) functions (see
Figure 4.2b),
• XOR-n (n > 2): EPC with n alternative (XOR connectors) functions (AND-n
with XOR instead of AND connectors) and
• OR-n (n > 2): EPC with n alternative (OR connectors) functions (AND-n
with OR instead of AND connectors).
The metric values for these four process model types are listed in Table 4.3. As
one can see, some process model metric values are constant, while others increase
infinitely by increasing n or converge to a limit value.
Besides the behavior of the selected process model metrics for sequential and
parallel process models, the special behavior of the metric heterogeneity (CH) is
examined here.
For this metric, only the relative (not absolute) number of the three connector
types p(AND), p(XOR), p(OR) ∈ [0, 1] is important. Consequently, all process
models can be represented as a point ~p := (p(AND), p(XOR), p(OR))T in the
3-dimensional space forming a triangular area on a 2-dimensional hyperplane
(as p(AND) + p(XOR) + p(OR) = 1,8 see Figure 4.3).
For the metric value, it is unimportant how the relative number of connector
types is actually mapped to the single connector types (e. g., a process model with
one AND and one XOR connector has the same CH metric value as a process
model with one OR and one XOR connector). This symmetry is represented by
the three dashed lines (axes of symmetry) in Figure 4.3.
The metric has its minimum value 0 for process models with only one connector
type (ratio 0 : 0 : 1, minimal connector type heterogeneity) or no connectors at all
8 The sole exception where this equation does not hold are process models with no connectors at
all.
72 analysis of process model metric properties
Table 4.3: Behavior of the selected process model metrics for sequential and parallel
process models (Part 1 of 2).
as well as its maximum 1 for process models with the same number of all three
connector types (ratio 13 : 13 : 13 , maximal connector type heterogeneity).
~µ := (µ1 , µ2 , . . . , µe )T (4.8)
with
0 if corresponding edge is not part of cycle
µi := 1 if cycle traverses corresp. edge in edge’s direction (4.9)
−1 if cycle traverses corresp. edge against its direction .
(1,0,0)
●
0.3
0.4
0.5
0.6
0.8
(1/2,0,1/2) ● ● (1/2,1/2,0)
(1/3,1/3,1/3)
●
0.9
0.4
0.4
0. 3
3 0.5 0.6 0.7 0.5 0.
● ● 0.6 ●
(0,0,1) (0,1/2,1/2) (0,1,0)
Figure 4.3: Values of the process model metric heterogeneity (CH) depending on the
probabilities of the three connector types AND, XOR or OR (2-dimensional
hyperplane in R3 transformed into R2 ).
It has to be noticed that a cycle can traverse edges against its direction according
to this notation. For process model metric CN, this definition of a cycle is used
(the arc directions are ignored). In contrast, all arcs have to be traversed in
their directions in order to form a cycle for process model metric CYC (the arc
directions are not ignored).
Using the above notation, linearly independent cycles can be defined [11, p. 15].
Definition 4.1 (Linearly independent cycles) The cycles ~µ1 , ~µ2 , . . . , ~µk are said to
be linearly independent if the equation
Lemma 4.1 Let G be a graph, n the number of its nodes, e the number of its edges and
p the number of its components. Then,
CN(G) = e − n + p (4.11)
Basis Let G be a graph with n nodes and no edges (e = 0). Consequently, each
single node is a component (p = n) and the graph has no cycle (CN(G) = 0). So,
CN(G) = e − n + p = 0 − n + n = 0 holds.
Induction step When an additional edge is added to the graph, two possible
cases can occur:
2. The new edge connects two nodes of one component. As a consequence, the
number of components stays unchanged (p 0 = p). As there was already a
path between the two nodes (they are in the same graph component), adding
the new edge between them results in a new cycle. As this new cycle contains
the new edge, which cannot be part of any already existing cycle, the new
cycle is linearly independent of all other cycles. Consequently, the number
of linearly independent cycles is increased by 1 (CN 0 (G) = CN(G) + 1).
Also here, CN 0 (G) = e 0 − n 0 + p 0 = (e + 1) − n + p = CN(G) + 1 holds.
With the help of Lemma 4.1, Theorem 4.1 can now be proved.
Theorem 4.1 The process model metric CN defined as CN(G) := |A| − |N| + 1 for any
EPC G counts the number of linearly independent cycles (the arc directions are ignored).
Proof. According to Definition 2.10, G is a connected graph. So, it has only one
component (p = 1). Consequently, equation (4.11) becomes CN(G) = |A| − |N| + 1
for the EPC.
CN(Gi )
cess model metric values of a set of EPCs Gi lie on a 2-dimensional hyperplane of the
R3 .
76 analysis of process model metric properties
1 1 −1
form a 2-dimensional hyperplane of the R3 which is spanned by two linearly
independent vectors.
In the remainder of this paragraph, the correlations between different process
model metrics are examined. To start with, no strong correlations in means of
Spearman’s rank correlation coefficient (see Section C.2) or Pearson’s product-
moment correlation coefficient (see Section C.1) just based on the metrics’ defi-
nitions could be proved. Yet, some inequalities between pairs of process model
metrics could be found which constrain the metric values in such a way that the
existence of strong (linear) correlations is highly possible.
Theorem 4.3 For every EPC G, the inequalities
0 6 CN(G) (4.12)
and
CN(G) 6 SN (G) − 5 , SN (G) > 5 (4.13)
hold.
Proof. The two inequalities are proved in separate steps.
Regarding (4.13): In order to add an additional cycle (arc directions are ignored)
to an EPC, there have to be two connectors (one split and one join connector) in
the EPC. Between these two connectors, an additional path to the “sequential one”
must exist. The smallest example for this construct is an AND-2 (see Figure 4.2b).
At least one event or function node lies on each of these paths between two
connectors. So, the number of cycles of an EPC is limited by the number of
nodes (at least one node per cycle). Yet, not all nodes lie on the mentioned paths
between two connectors. At least five nodes (one start and one end event, one
split and one join connector as well as one function) are used at other places
(think of an AND-2 as in Figure 4.2b). Consequently, CN(G) 6 SN (G) − 5 holds
for SN (G) > 5.
4.3 experimental application 77
0 6 CN(G) (4.14)
and
1
CN(G) 6 SA (G) − 2 , SA (G) > 4 (4.15)
2
hold.
Regarding (4.15):
Let SA (G) > 4. Then, SN (G) > 5 (as the only process model with three nodes is
the SEQ-1 with two arcs and an EPC with four nodes is impossible). According to
(4.13), CN(G) 6 SN (G) − 5 holds. This inequality can be transformed as follows:
(4.16)
CN(G) 6 SN (G) − 5 = (SA (G) − CN(G) + 1) − 5
= SA (G) − CN(G) − 4 | + CN(G)
⇔ 2CN(G) 6 SA (G) − 4 | · 12
1
⇔ CN(G) 6 2 SA (G) − 2
and
hold.
Regarding (4.17): Analogous to the proof of Lemma 4.1, an EPC G with SN (G)
nodes needs at least SN (G) − 1 arcs to be connected (SA (G) > SN (G) − 1). For the
sequential process models SEQ-n, SA (G) = SN (G) − 1 holds (see Table 4.3).
78 analysis of process model metric properties
Regarding (4.18):
CN(G) = SA (G) − SN (G) + 1 ⇔ SA (G) = CN(G) + SN (G) − 1
Let SN (G) > 5.
(4.13)
SA (G) = CN(G) + SN (G) − 1 6 (SN (G) − 5) + SN (G) − 1 = 2SN (G) − 6
Theorem 4.6 For every EPC G, the inequalities
0 6 SC (G) (4.19)
and
2
SC (G) 6 SN (G) − 2 (4.20)
3
hold.
Proof. The two inequalities are proved in separate steps.
Regarding (4.19): An EPC must not have any connectors (e. g., sequential pro-
cess models SEQ-n). The number of connectors cannot be negative. So, SC > 0.
Regarding (4.22):
(SA (G))2
CNCK (G) =
SN (G)
(4.17) (SN (G) − 1)2 (SN (G))2 − 2SN (G) + 1 1
> = = SN (G) − 2 +
SN (G) SN (G) SN (G)
> SN (G) − 2
Regarding (4.24):
In this subsection, the results concerning the process model collection specific
properties are presented.
Descriptive Statistics
The values of the descriptive statistics for the selected process model metrics
applied to the SAP Reference Model EPCs are listed in Table 4.4. Values which
are not meaningful because of the process model metric’s scale type are printed
in italics. Nevertheless, they are not totally skipped as they often lead to fruitful
results9 .
The histograms of the selected process model metrics give more detailed
information than just some compact quantitative numbers. They are depicted in
Figure 4.4, 4.5 and 4.6.
Most process model metrics (including all “size metrics”) have a high frequency
for small values and almost continuously decreasing frequencies for increasing
metric values resulting in a long tail at the right side. The number of nodes metric
(SN ) (see Figure 4.5a) is a typical example for this behavior.
The metrics control flow complexity (CFC) (see Figure 4.6f) and join complexity
(JC) (see Figure 4.6g) have a similar behavior with some few extreme outliers
(see quartiles in Table 4.4). These are caused by contained OR splits with high
out-degree (for metric CFC) and OR joins with high in-degree (for metric JC)
respectively.
The metric coefficient of connectivity (CNC) (see Figure 4.5f), which is defined
|A|
as CNC(G) := |N| , has three different areas of values: (1) CNC < 1 for |A| <
|N| ⇒ |A| = |N| − 1 (as |A| > |N| − 1): The smallest possible value is 23 for process
model SEQ-1 (see paragraph Theoretical Value Ranges in Subsection 4.3.2). The
|A| |N|−1
fraction |N| = |N| converges to 1 for |N| → ∞. The metric values of all sequential
processes SEQ-n form the sequence (an ) = ( 23 , 45 , 67 , . . .) with an = 2n+1
2n
and
are situated in the interval [ 3 , 1). (2) CNC = 1 for |A| = |N| (3) CNC > 1 for
2
|A| > |N|: The upper bound of CNC is 2 (see paragraph Theoretical Value Ranges in
Subsection 4.3.2). Within the SAP Reference Model, 325 process models (63.1%)
have a metric value less than 1 (31 process models with value 23 , 40 process
models with 0.8, . . . ), 75 process models (14.6%) have value 1 and 115 process
models (22.3%) have a value greater than 1. The process model with the second
greatest value 1 31 is an AND-5; the process model with the greatest value 1 15 21
(≈ 1.714) is an AND-17.
9 Stevens states regarding this problem [148, p. 679]: “In the strictest propriety the ordinary statistics
involving means and standard deviations ought not to be used with these scales, for these statistics
imply a knowledge of something more than the relative rank-order of data. On the other hand,
for this ‘illegal’ statisticizing there can be invoked a kind of pragmatic sanction: In numerous
instances it leads to fruitful results. While the outlawing of this procedure would probably serve
no good purpose, it is proper to point out that means and standard deviations computed on an
ordinal scale are in error to the extent that the successive intervals on the scale are unequal in
size.”
Table 4.4: Descriptive statistics of the selected process model metrics for the EPCs of the SAP Reference Model (Part 1 of 2).
200
200
200
150
150
150
frequency
frequency
frequency
100
100
100
50
50
50
0
0
0 5 10 15 20 25 30 0 5 10 15 20 25 30 35 0 10 20 30 40 50
250
60
120
50
100
200
40
80
150
frequency
frequency
frequency
30
60
100
20
40
50
10
20
0
0 20 40 60 80 0 10 20 30 40 50 0 0 2 4 6 8 10
300
300
250
250
250
200
200
200
frequency
frequency
frequency
150
150
150
100
100
100
50
50
50
0
0 2 4 6 8 10 0 5 10 15 0 2 4 6 8 10
400
100
300
80
300
250
60
200
frequency
frequency
frequency
200
150
40
100
100
20
50
0
0 2 4 6 8 10 0 2 4 6 8 10 0 10 20 30 40 50
Figure 4.4: Histograms of the selected process model metrics (Part 1 of 3).
84 analysis of process model metric properties
40
60
40
50
30
30
40
frequency
frequency
frequency
20
30
20
20
10
10
10
0
0
0 20 40 60 80 100 120 140 0 20 40 60 80 100 120 140 0 10 20 30 40
350
50
80
300
40
60
250
30
200
frequency
frequency
frequency
40
150
20
100
20
10
50
0
0
0.0 0.1 0.2 0.3 0.4 0.0 0.2 0.4 0.6 0.8 0.6 0.8 1.0 1.2 1.4 1.6 1.8
Delta D CNC
200
60
300
50
150
250
40
200
frequency
frequency
frequency
100
30
150
20
100
50
10
50
0
CNC_K CN AvgCDeg
60
50
50
150
40
40
frequency
frequency
frequency
30
100
30
20
20
50
10
10
0
0 5 10 15 20 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0
MaxCDeg Pi Xi
(j) dc
C histogram. (k) Π histogram. (l) Ξ histogram.
Figure 4.5: Histograms of the selected process model metrics (Part 2 of 3).
4.3 experimental application 85
300
100
200
250
80
150
200
60
frequency
frequency
frequency
150
100
40
100
50
20
50
0
0
0 1 2 3 4 5 0 10 20 30 40 0.0 0.2 0.4 0.6 0.8 1.0
Lambda MM CH
200
120
100
400
150
80
300
frequency
frequency
frequency
100
60
200
40
50
100
20
0
0.0 0.2 0.4 0.6 0.8 1.0 0 5 10 15 20 25 30 35 0 0 10 100 10^3 10^4 10^5 10^6
CYC TS CFC
35
50
30
100
40
25
80
30
20
frequency
frequency
frequency
60
15
20
40
10
10
20
5
0
0 10 100 10^3 10^4 10^5 0.0 0.1 0.2 0.3 0.4 0.0 0.1 0.2 0.3 0.4 0.5
JC CP CC
Figure 4.6: Histograms of the selected process model metrics (Part 3 of 3).
Looking at the metrics average (dC ) (see Figure 4.5i) and maximum connector
degree (d
cC ) (see Figure 4.5j), one finds 45 process models (8.7%) with value 0.
These are exactly the sequential process models SEQ-n, which do not contain
any connectors at all. As a split connector has exactly one incoming and at least
two outgoing arcs (vice versa for join connectors), there is no process model
with values in the open interval (0, 3). Starting with value 3, the frequencies for
increasing metric values are decreasing. As a process model’s maximum connector
degree is at least as high as its average connector degree, there are larger metric
86 analysis of process model metric properties
values for d cC than for dC . The value distribution for the maximum connector
degree metric is also shifted a little bit upwards compared to the distribution for
average connector degree.
210 of 515 process models (40.8%) have a heterogeneity metric (CH) (see Fig-
ure 4.6c) value of 0 (45 sequential process models SEQ-n and 165 process models
with only one connector type) and nine have value 1 (same number of all three con-
nector types). There are 14 process models with value ≈ 0.455 (ratio 0 : 0.2 : 0.8),
16 process models with value ≈ 0.512 (ratio 0 : 0.25 : 0.75), 56 process models
with value ≈ 0.579 (ratio 0 : 31 : 23 ) and 56 process models with value ≈ 0.631
(ratio 0 : 0.5 : 0.5). Those process models with higher values up to 1 have an even
more equal ratio of connector types.
487 of the 515 process models (94.6%) have a cyclicity metric (CYC) value of
0—meaning that they do not contain any (directed) cycles at all (see Figure 4.6d).
The remaining process models have values almost equally distributed up to
≈ 0.722.
In contrast, only 325 process models (63.1%) have a cyclomatic number metric
(CN) value of 0—so, they do not even contain an undirected cycle (see Fig-
ure 4.5h).
Correlations
Spearman’s rank correlation coefficients (see Section C.2) and Pearson’s product-
moment correlation coefficients for the selected process model metrics applied to
the SAP Reference Model are listed in Table 4.5 and 4.6 respectively. Coefficients
which do not differ significantly from 0 (two-sided test, p = 0.05) are printed in
italics in both tables. Values which are not meaningful because of the process
model metric’s scale type are skipped. Table cells are colored according to their
absolute values—the higher the absolute value the darker the color.
Additional to the correlation coefficients, also the 33 × 33 scatter plot matrix was
looked at. That way, also non-monotonic and non-linear dependencies between
two process model metrics can be identified. Because of length restrictions, only
scatter plots with an “interesting” behavior are depicted.
The “size metrics” SE , SC , SN and SA are highly linear correlated with each
other (all correlation coefficients > 0.93). Two examples are given in Figure 4.7.
The boundaries according to Theorem 4.5 and 4.6 respectively are also drawn
within the figures. They indicate the possibility for the existence of a linear
correlation—what is now confirmed for the SAP Reference Model.
Metric CNCK is also highly linear correlated with the “size metrics” SE , SC ,
SN and SA (all correlation coefficients > 0.94). At first glance, this may be a little
(S (G))2
surprising because of the power of 2 in the definition (CNCK (G) = SA (G) ). For
N
the correlation between the metrics CNCK and SN , Theorem 4.7 indicates the
Table 4.5: Spearman’s rank correlation coefficients of the selected process model metrics for the EPCs of the SAP Reference Model. Coefficients
which do not differ significantly from 0 (two-sided test, α = 0.05) are printed in italics.
SE SE SE SF SS SJ SS SJ SS SJ SC SN SA diam ∆ D CNC CNCK CN dC d
dC Π Ξ Λ MM CH CYC TS CFC JC CP CC
Int E ∧ ∧ X X ∨ ∨
SE 0.363 0.384 0.702 0.373 0.390 0.649 0.231 0.597 0.279 0.470 0.691 0.670 0.648 0.523 −0.689 0.118 0.440 0.629 0.268 0.429 0.610 −0.508 −0.491 0.398 0.653 0.598 0.147 0.395 0.331 0.852 −0.703 −0.589
S
SE 0.431 0.776 0.872 0.505 0.377 0.449 0.514 0.420 0.308 0.671 0.818 0.822 0.922 −0.808 0.547 0.760 0.825 0.651 0.273 0.422 −0.244 -0.013 0.648 0.450 0.463 0.264 0.534 0.552 0.543 −0.630 −0.383
Int
SE 0.738 0.329 0.657 0.252 0.570 0.417 0.588 0.106 0.695 0.670 0.646 0.545 −0.689 0.096 0.434 0.628 0.250 0.496 0.633 −0.544 −0.536 0.367 0.697 0.597 0.149 0.793 0.835 0.336 −0.713 −0.633
E
SE 0.705 0.638 0.513 0.509 0.627 0.530 0.386 0.875 0.973 0.956 0.872 −0.984 0.341 0.741 0.940 0.496 0.574 0.756 −0.580 −0.422 0.576 0.801 0.711 0.205 0.731 0.732 0.742 −0.941 −0.706
SF 0.505 0.402 0.323 0.465 0.306 0.295 0.631 0.807 0.819 0.847 −0.790 0.548 0.777 0.825 0.628 0.343 0.457 −0.207 0.039 0.604 0.403 0.371 0.194 0.513 0.447 0.535 −0.597 −0.330
SS 0.450 0.358 0.456 0.321 0.173 0.743 0.680 0.681 0.633 −0.674 0.378 0.606 0.681 0.500 0.329 0.500 −0.467 −0.459 0.542 0.497 0.525 0.137 0.796 0.565 0.420 −0.574 −0.320
∧
SJ 0.158 0.289 0.210 0.325 0.624 0.556 0.564 0.493 −0.544 0.375 0.526 0.567 0.479 0.292 0.479 −0.453 −0.432 0.529 0.403 0.457 0.245 0.436 0.280 0.568 −0.449 −0.191
∧
SS 0.509 0.264 0.025 0.576 0.512 0.508 0.508 −0.512 0.275 0.433 0.504 0.381 0.151 0.298 −0.388 −0.365 0.455 0.431 0.498 0.232 0.301 0.621 0.313 −0.539 −0.589
X
SJ 0.387 0.128 0.695 0.649 0.649 0.603 −0.645 0.397 0.570 0.645 0.517 0.243 0.439 −0.473 −0.400 0.534 0.541 0.612 0.249 0.423 0.489 0.688 −0.674 −0.682
X
SS 0.267 0.539 0.510 0.515 0.442 −0.505 0.313 0.446 0.515 0.441 0.335 0.468 −0.439 −0.367 0.456 0.525 0.560 0.193 0.688 0.761 0.406 −0.553 −0.599
∨
SJ 0.399 0.396 0.403 0.315 −0.389 0.309 0.379 0.407 0.396 0.226 0.351 −0.322 −0.203 0.382 0.369 0.397 0.036 0.269 0.241 0.646 −0.409 −0.373
∨
SC 0.909 0.913 0.825 −0.900 0.529 0.821 0.912 0.651 0.449 0.675 −0.670 −0.618 0.732 0.757 0.783 0.256 0.768 0.769 0.791 −0.863 −0.708
SN 0.996 0.916 −0.996 0.473 0.840 0.989 0.610 0.543 0.733 −0.570 −0.410 0.668 0.748 0.698 0.218 0.748 0.720 0.765 −0.918 −0.669
SA 0.916 −0.985 0.535 0.878 0.998 0.662 0.554 0.740 −0.594 −0.421 0.702 0.726 0.691 0.233 0.759 0.725 0.772 −0.907 −0.663
diam −0.910 0.522 0.820 0.913 0.629 0.343 0.514 −0.389 −0.248 0.684 0.575 0.601 0.241 0.632 0.639 0.649 −0.769 −0.490
∆ −0.409 −0.797 −0.973 −0.556 −0.532 −0.723 0.545 0.396 −0.630 −0.765 −0.701 −0.206 −0.733 −0.713 −0.754 0.924 0.673
D 0.838 0.577 0.951 0.267 0.361 −0.501 −0.268 0.791 0.162 0.315 0.292 0.440 0.431 0.490 −0.347 −0.271
CNC 0.902 0.861 0.523 0.653 −0.647 −0.437 0.820 0.513 0.571 0.282 0.686 0.658 0.709 −0.723 −0.540
CNCK 0.696 0.562 0.742 −0.609 −0.426 0.722 0.709 0.683 0.243 0.764 0.727 0.774 −0.895 −0.656
CN 0.303 0.460 −0.556 −0.319 0.838 0.317 0.432 0.337 0.560 0.539 0.597 −0.492 −0.392
dC 0.910 −0.712 −0.431 0.278 0.535 0.353 0.076 0.578 0.493 0.471 −0.599 −0.528
d
dC −0.794 −0.549 0.466 0.691 0.545 0.136 0.717 0.635 0.654 −0.777 −0.667
Π 0.783 −0.524 −0.594 −0.561 −0.266 −0.639 −0.639 −0.628 0.672 0.679
Ξ −0.404 −0.524 −0.553 −0.122 −0.533 −0.542 −0.496 0.524 0.587
Λ 0.382 0.514 0.287 0.571 0.594 0.633 −0.543 −0.455
MM 0.771 0.163 0.612 0.655 0.639 −0.822 −0.747
CH 0.211 0.578 0.652 0.663 −0.756 −0.686
CYC 0.172 0.227 0.241 −0.190 −0.190
TS 0.804 0.485 −0.691 −0.509
CFC 0.468 −0.758 −0.744
JC −0.779 −0.707
CP 0.831
4.3 experimental application
87
Table 4.6: Pearson’s product-moment correlation coefficients of the selected process model metrics for the EPCs of the SAP Reference Model.
Values which are not meaningful because of the process model metric’s scale type are skipped. Coefficients which do not differ
significantly from 0 (two-sided test, α = 0.05) are printed in italics.
SE SE SE SF SS SJ SS SJ SS SJ SC SN SA diam ∆ D CNC CNCK CN dC d
dC Π Ξ Λ MM CH CYC TS CFC JC CP CC
Int E ∧ ∧ X X ∨ ∨
SE 0.475 0.475 0.767 0.295 0.468 0.723 0.380 0.575 0.396 0.552 0.680 0.714 0.696 0.520 −0.476 0.349 0.674 0.433 0.308 0.593 −0.373 0.463 0.692 0.016 0.534 -0.016 0.343
S
S EInt 0.510 0.824 0.750 0.635 0.598 0.564 0.685 0.573 0.498 0.783 0.876 0.895 0.866 −0.522 0.595 0.905 0.831 0.129 0.400 −0.122 0.675 0.581 0.120 0.683 0.044 -0.032
SE 0.837 0.217 0.732 0.482 0.760 0.628 0.700 0.213 0.791 0.773 0.747 0.529 −0.469 0.349 0.716 0.422 0.261 0.530 −0.397 0.517 0.746 0.049 0.874 0.191 -0.050
E
SE 0.522 0.767 0.726 0.718 0.778 0.701 0.502 0.930 0.975 0.964 0.791 −0.602 0.534 0.947 0.698 0.281 0.618 −0.362 0.684 0.828 0.079 0.874 0.100 0.084
SF 0.485 0.489 0.274 0.410 0.268 0.326 0.503 0.669 0.678 0.741 −0.471 0.546 0.683 0.599 0.185 0.272 -0.013 0.446 0.288 0.049 0.417 0.007 -0.041
SS 0.624 0.622 0.640 0.462 0.281 0.829 0.804 0.798 0.695 −0.481 0.477 0.784 0.589 0.142 0.363 −0.366 0.627 0.611 0.062 0.775 -0.029 -0.045
∧
SJ 0.403 0.520 0.445 0.473 0.778 0.765 0.766 0.648 −0.426 0.452 0.759 0.609 0.156 0.399 −0.348 0.622 0.532 0.125 0.649 -0.040 0.021
∧
analysis of process model metric properties
SS 0.685 0.548 0.152 0.779 0.716 0.707 0.559 −0.398 0.373 0.693 0.506 0.089 0.278 −0.320 0.564 0.573 0.123 0.602 -0.037 -0.040
X
SJ 0.640 0.285 0.850 0.801 0.805 0.705 −0.474 0.495 0.800 0.659 0.125 0.372 −0.348 0.661 0.676 0.120 0.670 0.002 -0.025
X
SS 0.390 0.748 0.694 0.695 0.566 −0.386 0.412 0.689 0.556 0.143 0.383 −0.318 0.601 0.643 0.074 0.773 0.116 -0.033
∨
SJ 0.510 0.517 0.529 0.453 −0.308 0.352 0.535 0.495 0.155 0.351 −0.208 0.451 0.403 0.037 0.407 0.079 0.087
∨
SC 0.955 0.954 0.805 −0.549 0.565 0.945 0.751 0.174 0.466 −0.426 0.779 0.759 0.124 0.857 0.006 -0.014
SN 0.996 0.864 −0.616 0.600 0.984 0.763 0.253 0.556 −0.345 0.733 0.773 0.096 0.859 0.059 0.033
SA 0.861 −0.598 0.627 0.996 0.818 0.263 0.561 −0.343 0.749 0.759 0.103 0.863 0.057 0.027
diam −0.675 0.648 0.850 0.665 0.148 0.351 −0.237 0.722 0.556 0.121 0.644 -0.027 -0.028
∆ −0.777 −0.573 −0.352 −0.566 −0.600 0.613 −0.495 −0.569 −0.099 −0.462 -0.054 -0.062
D
CNC 0.649 0.676 0.599 0.606 −0.567 0.679 0.454 0.203 0.529 0.033 0.019
CNCK 0.866 0.275 0.564 −0.338 0.756 0.740 0.107 0.862 0.055 0.021
CN 0.271 0.476 −0.259 0.699 0.515 0.131 0.714 0.033 -0.019
dC 0.812 −0.622 0.168 0.357 0.017 0.326 0.245 0.255
d
dC −0.572 0.391 0.639 0.021 0.614 0.342 0.342
Π
Ξ −0.383 −0.463 −0.087 −0.380 -0.057 -0.063
Λ 0.491 0.211 0.653 -0.023 -0.045
MM 0.070 0.743 0.163 0.197
CH
CYC 0.077 -0.014 -0.017
TS 0.217 -0.042
CFC -0.005
JC
CP
88
4.3 experimental application 89
140
● ●
●
40
● ●
●
120
●●
● ●
● ● ●
●●
100
30
● ●
●
● ● ●
80
●
● ● ●
● ●
● ●
S_C
S_A
● ●
● ●● ●
●
20
●
60
● ●● ● ● ●●
● ● ●
●●● ●● ● ●
●● ●● ●
●● ● ● ●
● ● ● ● ●
● ●●
●● ●
●
●
●●● ● ●●●● ● ●
●
40
● ● ●● ● ●● ●●
●●●
● ● ● ●●●
● ●●
● ●●●
●●
10
●●●● ● ●● ● ● ●●
●
● ● ● ●●● ●● ● ●
● ●
●● ●●
● ● ● ●● ●●● ● ● ● ●
●●
●●●
● ●●
●
●●
●●
●● ●●
●●●●
●● ●●●● ●●● ● ● ●●
●●
●● ●●
●● ●●●●●●●●● ● ●
20
●
●●● ●●
●●
●
●●
●● ● ●●●●●●●●●● ●● ●
●●
●●
●● ●●●●
●●●●●●●●●●● ●●
●●
●●●●●
●● ●●●●●●●●●● ●●●●●
●● ●●●
●●●●
●●●
●●●● ●●
●●●●●●●●●● ●●●● ●
●●
●● ●●●●●●●● ● ●
●●
●●
● ●
●●●●●● ●
0
0
S_N S_N
(a) SA /SN plot with theoretical boundaries ac- (b) SC /SN plot with theoretical boundaries ac-
cording to Theorem 4.5. cording to Theorem 4.6.
Figure 4.7: SA /SN and SC /SN plots as examples of strong linear correlations between
“size metrics”.
possibility for such a strong linear correlation. As seen above, the ratio between
S (G)
SA and SN is almost a constant c ( SA (G) ≈ c), so
N
(SA (G))2
CNCK (G) S (G) c · SA (G)
= N ≈ ≈ c2 (4.26)
SN (G) SN (G) SN (G)
holds. The correlation between the metrics CNCK and SN as an example is
depicted in Figure 4.8a.
For metric CNC, things are different. Here, the correlation coefficients for the
pairs of CNC and the “size metrics” are much lower. Also the CNC/SN plot
(see Figure 4.8b) shows no linear or at least monotonic correlation. Yet, there are
points forming several curves. Having in mind that SA (G) = SN (G) + a , a ∈
{−1, 0, 1, . . . , 2SN (G) − 6} (Theorem 4.5), these curves are easily explainable.
There are two proposed process model metrics supposed to measure “density”:
∆ and D. As D is on ordinal scale, only Spearman’s rank correlation coefficients
can be considered during the following analysis. The rank correlation coefficient
between both metrics is −0.409. When one looks at the corresponding ∆/D plot
(see Figure 4.9a), many points with a D value of 0 (325 of 515 process models—
this equates 63.1%) attract attention. The reason is the definition of D, which
assigns D(G) = 0 for EPCs G with SC (G) 6 1. If one removes these points from
the analysis, the rank correlation coefficient is still only 0.598—what is quite small.
So, as a result, one can state that both “density metrics” measure something quite
different.
In contrast, the metrics ∆ and CP are highly correlated (rank correlation coeffi-
cient 0.924, see Figure 4.9b). This is only surprising at first glance as both metrics
measure a ratio between existing arcs and possible arcs—the number of arcs
90 analysis of process model metric properties
● ●
150
1.6
●
●
●
1.4
●
●
●●
● ●
100
●
●
● ● ●
●
CNC_K
● ● ●
CNC
1.2
● ● ● ●● ● ●
●
● ●
● ● ● ●●
●
● ●
●
● ● ● ● ● ● ● ●
●● ●● ● ●
● ● ● ● ● ● ●
● ●● ●● ●●● ● ●●●
● ●●● ●● ●
● ●● ●
● ● ●● ● ● ●● ●● ● ● ●
● ●● ●● ● ●● ● ● ●
● ●●● ● ● ●
● ● ●● ● ● ● ●
●● ● ● ●
1.0
● ● ●
● ● ●●●●●●●
●●●●●●●●●●●●●●●● ●● ● ●●●
●● ●
●
50
●● ●
● ● ●●●●●●●● ● ●●
● ●●
●● ●●
● ●●●●
●
●● ●●●
● ●● ●●●
●
●● ● ●
●●
●●
●
●
●● ●●● ●
●
●● ●
● ● ●● ● ●
●●
●●● ● ●
● ●● ●●●●
● ●●●
● (S_N+3)/S_N
●●●●●
0.8
●● ●
●● ●●●● ●
(S_N+2)/S_N
●●●●●●
●●●●
● ●●
●●
●●● ●● (S_N+1)/S_N
●●●●●
●●
●●●●●●
●●●
S_N/S_N
●●
●
●●
●
(S_N−1)/S_N
0
S_N S_N
(a) CNCK /SN plot with theoretical boundaries (b) CNC/SN plot.
according to Theorem 4.7.
divided by the number of possible arcs between the nodes for metric ∆ and the
weighted number of arcs between tasks and the possible number of arcs between
the tasks in the case of metric CP. This fact supports the assumption that metric
CP with its relatively complicated computation of weights has little additional
benefit compared to metric ∆.
Metric ∆ has plots with the “size metrics” SE , SC , SN and SA very similar to
the x1 /x plot. This is also reflected in quite large absolute values of Spearman’s
rank correlation coefficients (all correlation coefficients < −0.9). For the metrics ∆
and SN (see Figure 4.9c), this was already “predicted” by Theorem 4.8.
Looking at the correlation coefficients of metric cycling (CYC), one gets the
impression that it is not correlated with any other metric. Yet, this indication
is only half the truth. Figure 4.10 depicts the CYC/SN plot. The overwhelming
majority of points has a CYC value of 0 (487 of 515 process models—this equates
94.6%). If one only looks at the remaining points (only 28 points!), one gets
a Spearman’s rank correlation coefficient of −0.885 and a Pearson’s product-
moment correlation coefficient of −0.809. So, the result of this analysis is that
it does not depend on the number of nodes whether an EPC has a cycle (arc
directions are not ignored) or not—yet, if it has at least one, the CYC metric values
decrease for larger EPCs.
At the end of this paragraph, the CN/SN (see Figure 4.11a) and CN/SA (see
Figure 4.11b) plots are depicted in order to show how the points for the SAP
Reference Model lie between the boundaries of Theorem 4.3 and 4.4 respectively.
4.3 experimental application 91
● ●
0.30
0.30
0.25
0.25
0.20
0.20
● ● ● ●● ● ●
● ●
Delta
Delta
● ● ● ● ● ● ● ● ●
0.15
0.15
● ● ● ●
● ● ● ●
●●● ● ●
● ● ●
● ●
● ● ● ● ●● ●● ● ●●
● ● ● ● ●●●● ●● ● ●● ● ●● ● ●
●● ●●
0.10
0.10
● ●● ● ●● ● ● ●●●● ●● ●
● ● ●● ● ●● ● ●
● ●● ●●
● ● ● ●● ● ● ● ● ●● ●●
● ● ●
● ●● ● ●● ●●● ●● ●●●
● ● ● ● ● ●● ●●●● ●● ●●● ●
● ●● ● ● ● ● ●●● ●
●● ● ● ● ● ●●
● ●● ● ● ● ● ●● ● ●●
● ● ● ●● ● ● ●●●●
●
● ● ●●● ●
0.05
0.05
●
● ●●● ●●●●●●
●●● ●●● ●●
●
● ●
●
●
●
● ●●
●
●● ●
●● ● ●●●●●
● ● ●
● ●● ●●
●
●
●●● ●
●
●
●
●
●
●●●●
●
● ●
●
●
● ●●
●
● ● ● ● ● ● ● ● ● ●●
● ●●●
●●
●
●
●
●●
● ●
● ●
●
● ●
● ● ●
●● ●● ● ●● ●●
●
● ●
●
●●●
● ●● ● ● ● ●● ●●●
●
●●● ●●●●● ●●
●●
●
●
● ●● ●● ● ● ●● ●
● ●
●●●●● ●
● ●● ●
● ● ●●
●●●
●● ●●
●●●●
● ●
●●
● ● ●● ● ●● ●
● ●
●●
●●
●●●
●● ●●● ● ● ●
● ● ●●● ● ●
●●
● ●●●●●●
● ● ●●
● ●●●
●
●●●
●
● ● ●● ●●●●●●
●●●●● ●● ●●
● ●
●●
●
● ●
●
0.00
0.00
0.0 0.2 0.4 0.6 0.00 0.05 0.10 0.15 0.20 0.25 0.30
D CP
●
0.30
0.25
0.20
●●
●
Delta
●● ●
0.15
●●
●
●
●●●
●●●
0.10
●●●
●●
● ● ●
●●
●● ●●●
●●●●●
●●●●●
●● ● ●
0.05
●●
●●
●●
●●
●●
●
●●●●
●●
● ●●
●●●
●●● ●
●●
●●
●●●●
●●●●
●●●●●
●●●● ●
●●● ●
● ●●●● ●● ●
●● ●●
● ●●● ●●● ● ●
● ● ●●● ●●
● ●●● ● ● ● ● ●
0.00
0 20 40 60 80 100 120
S_N
Table 4.7: Proportion of variance and cumulative proportion of the 33 components of the
PCA.
0.6
●
●
●
●
● ●
●
●
0.4
●
●
CYC
●
●
●
●
●
●
●
0.2
●
●
● ●
●
●
0.0
●
●●
●●● ●●
●●●●●●
●●
●●●●●
●●●●●
●● ●●●●
●● ●●
●●●●
●●●●●●●●●●●●●●●●●● ●●●●●● ●● ● ● ●●
●● ● ● ● ●● ● ●●● ● ● ● ●
0 20 40 60 80 100 120
S_N
● ●
25
25
20
20
● ● ● ●
15
15
● ●
CN
CN
● ● ● ●
● ● ●●● ● ● ●●●
10
10
●● ● ● ● ● ●● ● ● ● ●
● ● ● ● ● ● ● ● ● ● ● ● ● ●
● ● ● ●
● ● ● ●
● ●
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
5
●●●●●●●
●●●●●●●●●●●●●●●● ●● ● ●●●
●● ● ●●
●●●●●●●●●●●●●●●●●●●●●●● ●●●●
●● ●
●●●
●●●
●●
● ●●
●
●●
●●●●●
●● ●●●●●●●●●●●●●●●●● ● ●●
●● ● ● ●●
● ●●●
●●● ●●
●●
● ●●●●
●●
● ●●●●●●●●●●●●●●●●● ● ●●
●● ● ●
0
S_N S_A
(a) CN/SN plot with theoretical boundaries (b) CN/SA plot with theoretical boundaries
according to Theorem 4.3. according to Theorem 4.4.
nent comprises more than half (54.402%) of the total data variance; the first nine
components (of 33) together more than 90%. So, there is a lot of “redundancy” in
the data of the 33 process model metrics.
Finally, the location of the original 33 process model metrics in the PCA’s
new system of basis vectors (called “loadings” in the PCA literature) is to be
examined. As the first three components comprise more than 71% of the total
data variance, a visualization which is restricted to these first three dimensions of
the 33-dimensional data is meaningful. The results are depicted in Figure 4.12. As
a 3D visualization is hard to interpret as long as it is not interactively rotatable,
three 2D plots are used instead.
94 analysis of process model metric properties
● Xi ● S_E_E
● Pi
CC ●
● MM
● CP
0.2
● ● S_S_XOR
0.2 S_E_Int Delta log.CFC.
● ●● S_S_OR
S_E ● CH
S_N/S_A/
S_S_AND ● ● S_E_S
CNC_K diam TS
● S_J_AND ● S_F
●● S_C ●
● ●
● S_J_XOR
●
● ● ● S_S_XOR
S_E ●S_J_XOR ● CN S_C ● S_S_AND Pi
TS ● ●
● S_S_OR ●
● S_N
0.0
Lambda ●
● S_E_E ● MaxCDeg Delta
● S_A
● S_J_OR
● log.JC. CP ●
●
0.0
● CNC_K
comp2
comp3
● S_E_S S_J_AND
Xi ●
MM ● ● ● AvgCDeg
● CYC
diam ● S_J_OR ● CYC
● Lambda CC ●
log.CFC. ● S_E_Int ●
−0.2
● CH ● D
log.JC. ●
● CNC CN ●
−0.2
CNC ● ● S_F
● MaxCDeg
−0.4
−0.4
● AvgCDeg ● D
comp1 comp1
● S_E_E
● MM
0.2
● S_S_XOR
● log.CFC. ● S_S_OR
CH ●
● S_E_S ● S_E
TS ●
● ● S_C
S_J_XOR
● S_S_AND
● Pi
● S_N ● Delta
0.0
● MaxCDeg
● S_A
● log.JC. ● CP
S_J_AND ● ● CNC_K
comp3
● Xi
● AvgCDeg ● S_J_OR
● CYC ● diam
● Lambda ● CC
● S_E_Int
−0.2
● CN
● CNC ● S_F
−0.4
● D
comp2
Figure 4.12: Selected process model metrics transformed by PCA (first three components).
Looking at these plots (especially Figure 4.12a), one can identify three clusters.
The first one consists of almost all “size metrics” (the metrics SA , SC , SN and
CNCK have almost the same comp1 and comp2 values). The second cluster
comprises the metrics ∆, Π, Ξ, CP and CC and is clearly separated from the first
one. The third cluster contains the remaining metrics. It is not as cohesive as the
other two and is located next to the first cluster with the “size metrics”.
4.4 conclusion
In this chapter, an approach for reducing the experimental effort for the validation
of prediction systems was introduced.
4.4 conclusion 95
Its main idea is to add an additional analysis step before the selection of the
prediction system which shall be validated. In this preceding step, the behavior
as well as important properties of process model metrics which are part of
the potential prediction systems which shall be validated are first analyzed.
Through this, unfavorable properties of process model metrics (e. g., insufficient
dispersion of metric values or strong correlation with other process model metrics)
can be identified before the high effort for an experimental validation of the
corresponding prediction system occurs.
The approach distinguishes between general properties which hold for all
process models because of their definition and process model collection specific
properties which are only true for the examined process model collection.
The approach was tested with 33 EPC process model metrics and 515 pro-
cess models from the SAP Reference Model. During this test, some interesting
properties could be found.
As general properties, mathematical boundaries for the value pairs of some
“size metrics” could be identified. Furthermore, it could be shown, that ∆(G) ≈
1
SN (G) holds.
Even more process model collection specific properties could be discovered for
the SAP Reference Model:
• The metrics CFC and JC have some few extreme outliers. It is very unlikely
that there is a linear dependency between one of these metrics and a process
model quality measure. The existence of a threshold over which a process
model has some undesirable properties (e. g., high error probability) is
much more likely.
• There are many linear or at least monotonic correlations between the ex-
amined process model metrics. This was also confirmed by the result of a
PCA. There are three major clusters for the metrics: The first one consists
of almost all “size metrics”. The second cluster comprises the metrics ∆,
Π, Ξ, CP and CC and is clearly separated from the first one. The third (not
so cohesive) cluster contains the remaining metrics. Consequently, some
metrics do not provide much additional information compared to others.
• The density metrics ∆ and D have quite different behaviors. So, they do not
measure the same concept.
• On the other hand, the metrics ∆ and CP, which both measure some sort of
ratio between existing arcs and possible arcs, are highly correlated. So, it
seems that metric CP, which has a much more complicated computation
rule, has little additional benefit compared to metric ∆.
In Chapter 4, general properties of some EPC process model metrics and process
model collection specific properties of these metrics for the SAP Reference Model
were analyzed.
As most humans are visually thinking beings—preferring pictures to large
tables of numbers—, a visualization of large process model collections based on
process model metric values would be interesting. Yet, the resulting process model
metric data would be very high-dimensional making visualization problematic.
A second interesting question is whether there are clusters of (structurally)
similar process models among a process model collection.
In this chapter, an approach for these two goals is proposed. It comprises
heatmaps, a compact visualization technique for high-dimensional data originally
used in genetics, and scatter plots for dimensionally reduced data using PCA for
visualizing the process model metric data. Additionally, clustering is used for
analyzing
Finally, the proposed approach is applied to the same process model metrics
and process model collection as in Chapter 4 to make the findings comparable.
The remainder of this chapter is organized as follows: In Section 5.2, several
visualization techniques are presented and assessed for their adequateness for
high-dimensional data. Basics of clustering are given in Section 5.3. The visualiza-
tion and clustering approach is introduced in Section 5.4. Afterwards, Section 5.5
shows the results of an experimental application of the approach. The chapter
closes with a conclusion (Section 5.6).
x_5
●
●
●
100
● ●
● ●
●
● ● ●●
● ● ●
x_4
●
● ●
80
● ●
● ●●
● ●
● ●● ●
●
● ●●
● ●
●
●
● ● ●
60
● ●
●
●
●● ● ●
● ●
●
y
● ● ●
● ●
●
● ●
1 2 3 4 5 x_1
40
● ●●
● ●
● ●
●
● ●
● ●● ●● ●
● ●
● ●
20
● ●
●
● ●
●
●● ●
●
●
●
● ●
●
● ●
●
x_3
0
●●
●
0 20 40 60 80 100
x x_2
10
5
f
0
−5
−3 −2 −1 0 1 2 3
Figure 5.1: Examples of several visualization techniques. In (b)–(d), always the same three
vectors ~xred = (1, 2, 3, 4, 5)T , ~xblue = (5, 4, 3, 2, 1)T and ~xgreen = (4, 4, 4, 2, 2)T
are displayed.
Scatter Plots
Scatter plots (see Figure 5.1a for an example) can display two or three dimen-
sions. In the 3-dimensional case, a projection of the third dimension onto the
2-dimensional plane is used. Each vector is represented as a point at the corre-
sponding place in the coordinate system.
Scatter plots are good for visualizing large amounts of vectors. But they are
only applicable for 2D or at most 3D data. Consequently, they are inadequate for
the high-dimensional process model metric data.
Radar Charts
The first ideas of radar charts were published a long time ago by Friedmann in
1862 [46] and by Mayr in 1877 [85, p. 78] respectively.
Radar charts2 (see Figure 5.1b for an example) are drawn in two dimensions
and can display vectors with three or more dimensions and non-negative values.
For each dimension, there exists an axis. The axes start in one single center point
and are uniformly placed around the 360◦ of a circle. For depicting a vector, the
values of the vector components are marked as points on the corresponding axes.
These points are linked by (colored) lines forming a polygon for each vector.
If one wants to display several vectors, one can either draw them in one single
radar chart using different colors for each vector (as in Figure 5.1b) or one can
draw a radar chart for each single vector.
Radar charts soon become confusing when one increases the number of dimen-
sions and/or depicts many vectors. Consequently, they are inadequate for the
high-dimensional process model metric data.
Figure 5.2: Color legend for heatmaps (blue for minimum and red for maximum values).
Andrews Plots
Andrews plots (see Figure 5.1d for an example) were proposed by Andrews in [2].
They can display vectors with an arbitrary number of dimensions. Each vector
~x = (x1 , . . . , xn )T ∈ Rn is mapped to a function
x1
f~x (t) := √ + x2 sin t + x3 cos t + x4 sin 2t + x5 cos 2t + . . . , −π < t < π (5.1)
2
in the plot.
Andrews plots preserve the Euclidean distances between the vectors. That
means that vectors which lie close together in the n-dimensional space are
represented by lines (functions) which also lie close to each other. This property
can be used for identifying clusters or outliers among the vectors.
In Figure 5.1d, for example, the close blue and green lines (functions) indicate
that the corresponding vectors lie close to each other.
Nevertheless, Andrews plots soon become confusing when one increases the
number of dimensions and/or depicts many vectors. Consequently, they are
inadequate for the high-dimensional process model metric data.
5.2.2 Heatmaps
10
●
●
● ●
8
●
6
y
4
2
0
0 2 4 6 8 10
first few components comprise a high proportion of total variance, one can omit
the remaining components (dimension reduction) without loosing much of the
original information of the data set.
If the resulting dimensionally reduced data has at least three dimensions,
scatter plots can be used for visualization.
Nevertheless, one has to keep in mind that this is a lossy visualization technique
and that the original data was transformed into a new coordinate system.
5.3 clustering
5.3.1 Basics
X = C1 ∪ C2 ∪ · · · ∪ Ck , Ci ∩ Cj = ∅, i 6= j . (5.2)
In a good clustering, vectors which fall in the same cluster are quite similar—
while those which fall in different clusters are quite dissimilar. Figure 5.3 gives an
example of a clustering with three clusters. The three different symbols represent
the membership of the vectors to one of these clusters.
There are several different clustering methods. Two often used methods in
practice are hierarchical and partitive clustering. These are explained in more
detail in the following two subsections.
102 visualization and clustering of process model collections
Function AgglomerativeHierarchicalClustering(X)
Input: set X of data vectors
Output: clustering tree (dendrogram) D
1: C←∅
2: for i = 1 to |X| do
3: {initialize: assign each vector to its own cluster}
4: Ci ← {~xi }
5: C ← C ∪ {Ci }
6: end for
7: numberClusters ← |X|
8: repeat
9: {compute distances between all clusters}
10: for all Ci ∈ C do
11: for all Cj ∈ C do
12: if Ci 6= Cj then
13: compute distance d(Ci , Cj ) between clusters Ci and Cj
14: end if
15: end for
16: end for
17: {merge the two clusters Ci and Cj that are closest to each other}
18: Ci,j ← Ci ∪ Cj
19: C ← C\{Ci , Cj } ∪ {Ci,j }
20: numberClusters ← numberClusters −1
21: {store information about two sub-clusters}
22: Ci,j .child1 ← Ci
23: Ci,j .child2 ← Cj
24: D ← Ci,j
25: until numberClusters = 1
26: return D
At the beginning, each vector is assigned to its own cluster. Then, the algorithm
works iteratively. In each step, those two clusters which are closest to each other
5.3 clustering 103
are merged into a new cluster which is located at the next higher level in the
dendrogram.
For the inter-cluster distance d(Ci , Cj ) in line 13 of Algorithm 5.1, several
measures exist. Among these are
• single linkage
• complete linkage
• average linkage
1 X
da (Ci , Cj ) := d(~xi , ~xj ) . (5.5)
|Ci ||Cj | ~x ∈C
i i
xj ∈Cj
~
In each of these measures, d(~xi , ~xj ) is a distance measure between the two
vectors ~xi and ~xj . This could be, for example, the Euclidean distance k~xk2 defined
as
v
uX
u n
k~xk2 := t |xi |2 , ~x ∈ Rn . (5.6)
i=1
X
k X
E(C) = xj − c~i k22
k~ . (5.7)
i=1 x~j ∈Ci
Function KMeans(X, k)
Input: set X of data vectors, number of clusters k
Output: clustering C with k clusters
1: C←∅
2: for i = 1 to k do
3: Ci ← ∅
4: C ← C ∪ {Ci }
5: randomly initialize cluster center (centroid) ~ci
6: end for
7: repeat
8: {compute partitioning for data}
9: for i = 1 to k do
10: Ci ← ∅
11: end for
12: for j = 1 to |X| do
13: add ~xj to that Ci with shortest Euclidean distance between ~xj and ~ci
14: end for
15: {update cluster centers}
16: for i = 1 to k do
P
17: ~ci := |C1 | ~xj ∈Ci ~xj
i
18: end for
19: until partitioning stays unchanged or the algorithm has converged
20: return C
In [105], Milligan and Cooper present and assess 30 procedures for determining
the number of clusters of a data set. One possible solution to this problem is the
Davies-Bouldin index [36] defined as
1X
k
Sc (Ci ) + Sc (Cj )
DB(C) := max . (5.8)
k j∈{1,...,k} dce (Ci , Cj )
i=1 i6=j
Thereby, Sc is defined as
1 X
Sc (Ci ) := k~xj − ~ci k2 (5.9)
|Ci |
~xj ∈Ci
and acts as a dispersion measure quantifying the average centroid distance of the
cluster’s vectors. The measure dce is defined as
Based on the results of the previous two sections, an approach for visualization
and clustering of the high-dimensional process model metric data of process
model collections is proposed in this section.
It is divided into three steps which are subsequently explained.
In the first step, a heatmap is used for depicting the process model metric values
of the process models.
The process model metric values of a process model are displayed in one
row. The different process model metrics form the columns of the matrix. Ex-
ternal attributes as duration, costs, number of errors or understandability (see
Subsection 3.4.2) can be added as additional columns of the heatmap if desired.
In order to get a better insight into the correlations between the different process
model metrics, the columns of the heatmap can be hierarchically clustered.
The second step requires the application of a PCA on the process model metric
data. If the resulting first three components comprise a sufficiently large pro-
portion of total variance, the dimensionally reduced 3-dimensional data can be
visualized using either a 3D scatter plot or three 2D scatter plots.
5.4.3 Clustering
In the third and last step, clusters of structurally similar process models within
the process model collection are searched for. For that purpose, a partitive
clustering algorithm is applied to the process model metric data. The results can
be visualized within a heatmap again.
In this section, the abstract approach presented in the previous Section 5.4 is
applied to a set of selected process model metrics and a process model collection.
AvgCDeg
Pi
Xi
CP
S_S_XOR
CYC
S_E
S_A
S_J_OR
CC
Delta
S_J_XOR
S_E_S
JC
S_C
S_N
CNC_K
MM
CH
MaxCDeg
S_S_OR
S_E_E
CFC
S_F
S_E_Int
diam
TS
Lambda
CNC
D
CN
S_J_AND
S_S_AND
Figure 5.4: Heatmap displaying the values of the 33 selected process model metrics for the
515 selected process models. The rows (i. e., process models) are ordered by
the number of nodes metric (SN ). The columns (i. e., process model metrics) are
hierarchically clustered using (1 − Spearman’s rank correlation coefficient) as
distance measure.
The values of the 33 selected process model metrics for the 515 selected process
models are depicted in the heatmap of Figure 5.4.
5.5 experimental application 107
The values of each process model metric are normalized into the interval
[0, 1] as their domains are too different (see Table 4.4). The metrics control flow
complexity (CFC) and join complexity (JC) are logarithmically normalized as both
have some outliers with extremely high values compared to the large rest of the
values (see Table 4.4 as well as Figure 4.6f and 4.6g).
The rows (i. e., process models) are ordered by the number of nodes metric
(SN ). The columns (i. e., process metrics) are hierarchically clustered using (1−
Spearman’s rank correlation coefficient) (see Section C.2) as distance between two
columns (i. e., process model metrics) within the complete linkage inter-cluster
distance measure of (5.4).
The data is clearly displayed in the heatmap on one page. So, the main goal
of the visualization is fulfilled. Furthermore, several observations can be made,
which are consistent with the findings of Chapter 4:
• There is a negative correlation between most metrics (e. g., the “size metrics”)
and the metrics separability (Π), sequentiality (Ξ), cross-connectivity (CC),
density (∆) and weighted coupling (CP). The negative correlation is especially
strong between SC , SE , SN and SA on the one hand and ∆ and CP on the
other.
• Most metrics have many small and only some large values. Metric heterogene-
ity (CH) shows about one third to one half very small values—the remaining
values are relatively large. For the metrics separability (Π) and coefficient of
connectivity (CNC), most process models have values in the middle of the
domain.
As shown in Table 4.7 of Subsection 4.3.3, the first three components of the PCA
of the selected process model collection comprise more than 71% of the total data
variance. So, a visualization of the scores which is restricted to these first three
dimensions of the 33-dimensional data is meaningful. The results are depicted in
the left column of Figure 5.5. As a 3D visualization is hard to interpret as long as
it is not interactively rotatable, three 2D plots are used instead.
Most process models are located at the origin of the PCA coordinate system.
Especially in Figure 5.5a, one can also note three branches which show to the
top left (in direction of process model G1 ), to the bottom (G2 ) and to the top
right (G3 ). These findings correlate with the loadings of the 33 selected process
model metrics (see right column of Figure 5.5). Process model G1 is the EPC
with the most nodes (SN (G1 ) = 130) and arcs (SA (G1 ) = 138) within the selected
process model collection. Consequently, it is located in the direction of the “size
metrics”. G2 as an AND-17 has the largest average connector degree (dC ) value. So,
108 visualization and clustering of process model collections
● G_1 ● Xi
● ● Pi
● CC ●
● CP
●
5
●● ●
●
● G_3 S_E_Int
0.2
Delta
● ●
● ● S_N/S_A/
● ● S_S_AND
CNC_K diam
● ● S_J_AND ● S_F
●● ● ●● S_C ●
●
●● ● ● ● ● S_S_XOR
S_E ●S_J_XOR ● CN
● TS ●
●
● ● ●
● ● S_S_OR
● ● Lambda ●
●● ●● ● S_E_E
● ● ●● ● ● ● ●
● ● ●
● ● ●● ● S_J_OR
● ●● ●● ● ●●● ● ●●
●● ●
●
0.0
● ● ● ● ● ●●●
● ● ●● ●●●
●
●●●
comp2
comp2
● ● ● ● ●● ●● ● ●●● ● ● S_E_S
● ● ● ●●●● ●●●
● ●● ● ● ● ●●●●● ●●●● ● ●
0
● ● ●● ●●● ● ●● ●●
● ● ●● ● MM ●
● ●●
● ●● ● ●●●
● ●● ● ● ● ●● ●●
●●
●
●
●●
●●● ●
●
● CYC
●
●● ● ●● ●● ● ●●● ●● ●●●●●●
●●
●●●
● ● ● ● ●● ●● ●
●●
●●●●●●
● ● ●● ●●●● ●● ●●●●
● ● ●● ●
● ●●● ●●●●
● ●●● ●
●● ●●● ●●
● ●● ●●●● ●●● ●
●●●●
● ●● ●● ●● ● ● log.CFC. ●
●
●●● ●●● ●● ●● ●●●●
● ● ● ●●● ● ●●
●
● ● ●
●● ● ● CH ● D
● ● ●● ● ● log.JC. ●
● ●
●● ●
●
●
●
● CNC
−0.2
● ●
●
● ● ●
● ● MaxCDeg
−5
●●
●
●
−0.4
● G_2 ● AvgCDeg
comp1 comp1
● G_1 ● S_E_E
5
● ● MM
●
0.2
●
●● ● S_S_XOR
● ● ●
● ● ● log.CFC.
●● S_S_OR
● ●
● ●● ● ● ● CH
● ● S_E
● ● ●●●●
● ●● ●●● ●●● ●●
●
● ●●●
● ●●●● ● ● ● S_E_S
● ● ● ●●●●●●● TS
●● ●● ●● ● ● ● ●● ●
●●
●●●● ● ●●
G_3
● ● ● ●● ● ● ● ● ●●● ●● ●●
● ●
●
● ●●
●●●● ● ●
● ● ●●
●● ●● ● ● ● ●●● ●●●● ●●
●
●● ● ● ● S_J_XOR
● ● ● ● ● ● ●●●●
● ● ● ● ● ● ● ●●●●● ●● ● ●● ●●●● ●
● ●● ● ●
● ●●● ●●● ●●●●● ●●●● ●
●●●● ● S_C ● S_S_AND Pi
● ● ●● ● ●●● ●
●● ●●
●
0
●● ●● ● ● ● ● ●●●●● ●
● ●● ● ●
● ● ●
● ●● ●●● ●● ●
● ● ● ● ●●●● ●● ● ●
●● ● ● ● ● ● ● S_N
● ●● ● ●
0.0
● ●●● ● ● MaxCDeg
● ●● ● ● ● ● ● ●●● ● Delta
● ● ● ●● ●● ●
●●●
●●● ● S_A
● ● ●● ●● ●
● ● ● ●●
● ●● ● ● ● ● ●● ● log.JC. CP ●
●●● ● ● ● CNC_K ●
● ●
●● ●●
comp3
comp3
S_J_AND
● ● Xi ●
● ●● ● AvgCDeg
● ● ●
● ● ●
● diam ● S_J_OR ● CYC
●
● ● Lambda CC ●
S_E_Int ●
●
−5
−0.2
● ●
●
●
CN ●
●
CNC ● ● S_F
−0.4
−10
● G_2 ● D
comp1 comp1
Figure 5.5: Scores and loadings plots of the PCA (first three components). The left column
shows the scores plots of the selected process models. The right column shows
the loadings plots of the selected process model metrics (as in Figure 4.12)
(Part 1 of 2).
G_1 ● ● S_E_E
5
● ● MM
●
0.2
●
● ● ● S_S_XOR
●● ●
● ● ● ●
● ● log.CFC. ● S_S_OR
● ●
●●
● CH ●
● ● ●
● ●●●
●●●●
●● ●
●●●●●
●●
●
● ●●●
●● ● ● ●
● S_E_S ● S_E
●● ●●● ● ●●●
● ●●●●● ●●
● ● ●
● ● ● ● ●●
●
● ●
● ●●●●
● ●
● ●● ● ●●● ● ● ● ●● TS ●
● ●
●
●● ●● ●●
●●
●
●●●
● ●
●●●●●
●
●●●● ●
●●● ● ● ●●
●● ● ● ● S_C
●● ● ● ● ●
●●
●
●
●●
●
●●
●●
●●
●●●●●
●●
●●●
●● ●
●●
●●
●●●●●
● ● G_3 S_J_XOR
● ● ●●● ●
●
●● ● ●●
●●●●
●●
●●●●●
●
● ●●● ● ●
● S_S_AND
0
● ● ●●●
● ●●● ● ● ●
●
●
● ● ● Pi
●●● ● ●● ●● ● ●
● ● ●●● ●●● ●● ● ● ●● ● Delta
●● ● ● ●●● ●● ● ● ●● ● S_N
0.0
● ● MaxCDeg
●●●● ●●●
● ● ●
●
●
●●
●
●●●● ●● ● ●● ● ●● ● S_A
●● ●
●● ● ● ●●● ●●
●● ● ● ● ●● ● ● log.JC. ● CP
● ●● ● ● S_J_AND ● ● CNC_K
●● ●●
comp3
comp3
● ● ● ●
● Xi
● ● ● ● ● AvgCDeg ● S_J_OR
●●● ●
● ● CYC ● diam
●
● ● Lambda
● CC
● S_E_Int
●
−5
−0.2
● ●
●
●
● CN
●
● CNC ● S_F
−0.4
−10
● G_2 ● D
comp2 comp2
Figure 5.5: Scores and loadings plots of the PCA (first three components). The left column
shows the scores plots of the selected process models. The right column shows
the loadings plots of the selected process model metrics (as in Figure 4.12)
(Part 2 of 2).
Before clustering, the input data (normalized metric values from the non-
clustered heatmap) was scaled to mean 0 and variance 1 for each dimension. The
selection of the optimal number of clusters and the optimal clustering with this
cluster number for the input data was done using the Davies-Bouldin index. The
depicted clustering has a Davies-Bouldin index of 1.001289 based on the scaled
values.
The three clusters are not that “spectacular”. They simply segregate the process
models into three sets with middle (top), large (center) and small (bottom) metric
values for the first five metrics. If one looks at the results of the PCA visualization
in Subsection 5.5.3, this observation is not that surprising. In the left column of
Figure 5.5, also no clear cluster structure of the scores is identifiable.
5.6 conclusion
AvgCDeg
Pi
Xi
CC
Delta
CP
S_S_XOR
CYC
S_J_XOR
S_E_S
JC
S_C
S_E
S_N
S_A
CNC_K
MM
CH
MaxCDeg
S_S_OR
S_E_E
CFC
S_J_OR
S_F
S_E_Int
diam
Lambda
TS
CNC
D
CN
S_J_AND
S_S_AND
Figure 5.6: Clustered heatmap displaying the values of the 33 selected process model
metrics for the 515 selected process models. The rows (i. e., process models)
are separated into three clusters (see bar with gray scale values at the left).
The columns (i. e., process model metrics) are hierarchically clustered using
(1 − Spearman’s rank correlation coefficient) as distance measure.
The approach was successfully applied to the same EPC process model metrics
and process models as in Chapter 4.
5.6 conclusion 111
s ha
ha s
costs
PROCESS MODEL
duration
number of errors
internal external
possible reasons complexity possible implications changeability
attributes attributes
flexibility
pr
oc el
e od res understandability
m ss m m
et s s asu
ric od dependency e e
s e oc m
l pr lity
a
process model process model qu
metrics quality measures
values predictability values
Figure 6.1: Chapter visually located within the measurement approach of Subsec-
tion 3.4.2.
Content Validity
Content validity is concerned with whether a measure covers the range of mean-
ings included in the underlying concept (see Subsection 3.4.3 and A.3.2).
Sarshar et al. ask questions on the states before and after special events (EPC)
or transitions (Petri net) are reached [139, pp. 32–34, 39–41]. Mendling et al. name
four aspects of structural process model understandability: understanding of
order, concurrency, exclusiveness and repetition [101, p. 52]. In [102, p. 146],
Mendling and Strembeck ask questions on choices, concurrency, loops and dead-
locks. But these are not used to compute the MSCORE measure for process
models. For [126], the asked questions are not given. It is only stated [126, p. 26]
that the whole measurement approach is similar to that of [101].
Looking at these publications, some questions arise: Do other important aspects
of structural process model understandability exist? How different is the under-
standing based on the different aspects? How can “overall structural process
model understandability” be computed?
Reliability
Reliability requires that measure values obtained by different observers of the
same process model have to be consistent (see Subsection 3.4.3 and A.3.2).
In [138, 139], Sarshar et al. ask only ten questions per process model version
(EPC or Petri net)—even though these process models have more than 80 nodes. In
[101, 102], only eight and six questions per process model are asked, respectively.
And in [101], these questions are even distributed to four different aspects. Reijers
and Mendling ask 12 questions on process models with more than 100 tasks in
[126]. In all mentioned publications, it does not become clear how the nodes
(tasks, functions, events or transitions depending on the used process modeling
language) involved in the questions are selected.
6.3 framework for evaluating modeling technique understanding 117
content
model viewer
characteristic
So, the question arises whether this selection is representative for the process
model. It is possible that complicated parts of the process model have been
omitted or only especially complex parts have been selected. This selection could
have a big influence on the measured values.
In this section, an own approach for measuring structural process model under-
standability is introduced. It tries to overcome the existing measures’ potential
problems with reliability and validity which were identified in Subsection 6.2.2.
At the same time, it is as similar to the existing measurement approaches of
Subsection 6.2.1 as possible so that hypotheses about the existing approaches’
potential problems can be formulated and examined in subsequent experiments
(see Section 6.5).
The approach is based on Gemino and Wand’s framework for evaluating mod-
eling technique understanding which was presented in the previous Section 6.3.
The understandability is measured using comprehension tasks—consistent with
the existing approaches. The major advantages of this approach are:
• The approach systematically creates the asked questions (“comprehension
tasks” in Gemino and Wand’s framework) for all possible process models
6.5 approach for measuring structural process model [...] 119
activity period
time
execution of
task t becomes execution of
task t
executable task t starts
terminates
Definition 6.1 (Activity period) An activity period of task t is the period between a
point in time when t becomes executable and the next point in time when the actual
execution of t terminates (see Figure 6.3).
Now, relations for the four aspects of structural process model understandabil-
ity can be defined.
120 measuring structural process model understandability
Definition 6.2 (Concurrency) For the questions on task concurrency, the relations
c@ , c∃ , c∀ ⊆ T × T with the following meanings are used.
(t1 , t2 ) ∈ c@ ⇔ There is no process instance for which the activity periods of tasks t1
and t2 overlap.
(t1 , t2 ) ∈ c∃ ⇔ There is a process instance for which the activity periods of tasks t1
and t2 overlap at least once (Several executions of t1 and t2 per process instance are
possible!).—But there also exists a process instance for which this does not hold.
(t1 , t2 ) ∈ c∀ ⇔ For each process instance, the activity periods of tasks t1 and t2
overlap at least once.
Definition 6.3 (Exclusiveness) For the questions on task exclusiveness, the relations
e@ , e∃ , e∀ ⊆ T × T with the following meanings are used.
(t1 , t2 ) ∈ e@ ⇔ There is no process instance, for which tasks t1 and t2 are both
executed.
(t1 , t2 ) ∈ e∃ ⇔ There is a process instance, for which tasks t1 and t2 are both
executed.—But there also exists a process instance for which this does not hold.
(t1 , t2 ) ∈ e∀ ⇔ For each process instance, the tasks t1 and t2 are both executed.
Definition 6.4 (Order) For the questions on task order, the relations o@ , o∃ , o∀ ⊆ T × T
with the following meanings are used.
(t1 , t2 ) ∈ o@ ⇔ There is no process instance for which an activity period of task t1
ends before an activity period of task t2 starts.
(t1 , t2 ) ∈ o∃ ⇔ There is a process instance for which an activity period of task t1 ends
before an activity period of task t2 starts.—But there also exists a process instance for
which this does not hold.
(t1 , t2 ) ∈ o∀ ⇔ For each process instance, an activity period of task t1 ends before an
activity period of task t2 starts.
Definition 6.5 (Repetition) For the questions on task repetition, the relations r=1 , r? ,
r∗ , r+ ⊆ T with the following meanings are used.
t ∈ r=1 ⇔ For each process instance, task t is executed exactly once.
t ∈ r? ⇔ For each process instance, task t is executed not once or exactly once.
Both cases really occur.
t ∈ r∗ ⇔ For each process instance, task t is executed not once, exactly once or
more than once. There exists a process instance for which t is executed not once and
another one for which t is executed more than once.
t ∈ r+ ⇔ For each process instance, task t is executed at least once. There exists a
process instance for which t is executed more than once.
At first glance, the definitions of the relations might look a little complicated.
But they are constructed in such a way that the properties of Theorem 6.1 hold,
which is beneficial for the measurement process.
Theorem 6.1 (Properties of relations) The relations have the following properties:
1. The relations c@ , c∃ , c∀ and e@ , e∃ , e∀ are symmetric.
2. For all possible task combinations, exactly one relation per aspect is true.
6.5 approach for measuring structural process model [...] 121
Because of property 2 of Theorem 6.1, the different relations for an aspect can
be grouped to questions on the process model: The question qr (t), for example,
asks which of the relations r=1 , r? , r∗ , r+ holds for task t. Because of property 1 of
Theorem 6.1, qc (t1 , t2 ) = qc (t2 , t1 ) and qe (t1 , t2 ) = qe (t2 , t1 ) hold.
Theorem 6.2 (Maximum number of questions) The maximum number |Qa,max (p)|
of possible different questions for aspect a ∈ {c, e, o, r} on a process model p with n tasks
is
n(n − 1)
|Qc,max (p)| = |Qe,max (p)| = (6.1)
2
|Qo,max (p)| = n(n − 1) (6.2)
|Qr,max (p)| = n . (6.3)
Regarding (6.1): The questions on the aspects concurrency and exclusiveness each
depend on two tasks. As there are n tasks per process model p, n(n − 1) different
pairs of tasks exist. According to property 1 of Theorem 6.1, the relations for both
aspects are symmetric. So, for example, the questions with task order t2 and t1
can be omitted if the questions with the reverse task order t1 and t2 are asked.
n(n−1)
Consequently, |Qc,max (p)| = |Qe,max (p)| = 2 .
Regarding (6.2): The questions on aspect order each depend on two tasks. As
there are n tasks per process model p, n(n − 1) different pairs of tasks exist.
Consequently, |Qo,max (p)| = n(n − 1).
Regarding (6.3): The questions on aspect repetition each depend on one task. As
there are n tasks per process model p, |Qr,max (p)| = n.
As one can see, the maximum number of questions for concurrency, exclusiveness
and order grows quadratically with the number of tasks, while the maximum
number of questions for repetition grows only linearly.
Hypothesis 6.1 The personal structural process model understandability measure values
Ua (p, si ) of a process model p are normally distributed.
Additionally, confidence intervals for the true expected values of the random
variables for the different aspects of structural process model understandability
can be computed. The width of these intervals will decrease for higher numbers
of subjects—meanwhile, the certainty of the true expected value will increase.
It can be expected that the different aspects concurrency, exclusiveness, order and
repetition of structural process model understandability are of varying difficulty.
Consequently, it is important to measure all aspects to get “overall understand-
ability”.
The measures proposed in this subsection require that all possible questions
on an aspect are asked to each subject. As Theorem 6.2 shows, this number
of questions grows quadratically for the aspects concurrency, exclusiveness and
order. Even for practically relevant process models with, for example, ten tasks,
each subject would have to answer 45 questions on concurrency and exclusiveness
respectively as well as 90 questions on order. In the following subsections 6.4.3
and 6.4.4, two possible solutions for this problem are suggested.
A first possibility to reduce the effort for measuring structural process model
understandability is to select only a subset of all possible questions on the
different aspects for being answered by each subject. This approach was also used
in [101, 102].
# correct answers to Qa
Ua (p, s, Qa ) := , a ∈ {c, e, o, r} (6.6)
|Qa |
Here again, the different values of personal partial structural process under-
standability can be seen as outcomes of a random variable. The expected value of
this variable can be estimated according to Definition 6.9.
Definition 6.10 (Coverage rate) The coverage rate of a set of questions Qa ⊆ Qa,max (p)
on aspect a of process model p is defined as
|Qa |
ra (Qa , p) := , a ∈ {c, e, o, r} . (6.8)
|Qa,max (p)|
There are several possibilities to select a certain number of questions (i. e., equal
coverage rate).
124 measuring structural process model understandability
Theorem 6.3 The number of different sets of questions Qa ⊆ Qa,max (p) with |Qa | = m
questions is
One can assume that the different questions are not equally difficult to be
answered. This would have some important implications.
Hypothesis 6.3 The different questions of Qa,max (p) are not equally difficult. This has
two consequences:
1. For the same coverage rate, one gets different values for estimated partial structural
process model understandability depending on the selected questions Qa .
2. The smaller the coverage rate, the bigger the standard deviation of the different
values of estimated partial structural process model understandability for that
coverage rate.
As a consequence, the coverage rate should not be selected too small. Fur-
thermore, the questions for the set Qa should be chosen randomly in order to
minimize the risk of intentionally or unintentionally selecting especially easy
or difficult questions when done by a human. The two recommendations shall
assure that the estimated partial structural process model understandability does
not differ much from the true value of structural process model understandability.
Roughly speaking this means that a subject with good results for one subset of
questions will also be good for the second subset. This is used in inverse direction
in order to “construct” new virtual subject’s answers out of the answers given by
several real subjects.
The set of all possible questions for one aspect is divided into different subsets
which are each answered by different groups of subjects. Afterwards, in each
group the subjects are ordered by their personal partial structural process under-
standability values. Now, new virtual subjects are “created” by combining the
6.5 experimental evaluation 125
answers of one subject from each group. For this step, the best subjects from each
group are combined to the best new virtual subject, the second best subjects to
the second best new virtual subject—and so on.
Using the answers of these so “constructed” virtual subjects, (virtual) personal
structural process model understandability and (virtual) estimated structural
process model understandability can be computed as defined in Definition 6.6
and 6.7 respectively.
Both subsections are equally structured: First, the experiment design is ex-
plained (see Subsection B.3.2 for an explanation of the used terminology). Then,
the experiment’s results are given and analyzed. Finally, a validity evaluation
(see paragraph Validity Evaluation in Subsection B.3.3) is carried out.
6.5.1 Experiment 1
Experiment Design
For a first experiment, Hypothesis 6.1 (normally distributed personal structural
process model understandability values), 6.2 (different estimated structural pro-
cess model understandability values for different aspects) and 6.3 (influences of
coverage rate on estimated partial structural process model understandability)
were selected for examination.
Consequently, a rather small process model had to be chosen in order to be
able to let one single person answer all possible questions (for one aspect). So,
Start
XOR
XOR AND
B C D E
XOR AND
XOR
End
object Finally, the process model depicted in Figure 6.4 was selected. It was
presented to the subjects in the same top-to-bottom-style EPC-like notation as in
[101, 102].
As the process model has only five tasks, all |Qc,max (p)| = |Qe,max (p)| = 10,
|Qo,max (p)| = 20 and |Qr,max (p)| = 5 possible questions for the four aspects (cf.
Theorem 6.2) could be asked to single subjects.
qc (C, D)
qc (A, C)
qc (B, D)
qc (D, E)
qc (A, B)
qc (A, E)
qc (B, C)
qc (C, E)
qc (B, E)
subject Uc (p, s)
solution c@ c@ c@ c@ c@ c@ c@ c@ c@ c∃
s2 c@ c@ c@ c@ c@ c@ c@ c@ c@ c∀ 0.9
s4 c@ c@ c@ c@ c@ c@ c@ c@ c@ c∃ 1.0
s6 c@ c@ c@ c@ c@ c@ c@ c@ c@ c∃ 1.0
s34 c@ c@ c@ c@ c@ c@ c@ c@ c@ c∃ 1.0
s42 c@ c@ c@ c@ c@ c@ c@ c@ c@ c∀ 0.9
s50 c@ c∃ c@ c@ c∀ c∃ c∃ c∃ c∃ c@ 0.3
s52 c@ c@ c@ c@ c∃ - c@ c@ c@ c∃ 0.8
s56 c@ c@ c@ c@ c@ c@ c@ c@ c@ c∃ 1.0
s60 c@ c@ c@ c@ c@ c@ c@ c@ c@ c∃ 1.0
correct 100% 89% 100% 100% 78% 78% 89% 89% 89% 67%
Results
The answers to the questionnaire are given in Table 6.2 (aspect concurrency),
Table 6.3 (aspect exclusiveness), Table 6.4 (aspect order) and Table 6.5 (aspect
repetition).
qe (A, D)
qe (C, D)
qe (A, C)
qe (B, D)
qe (D, E)
qe (A, B)
qe (A, E)
qe (B, C)
qe (C, E)
qe (B, E)
subject Ue (p, s)
solution e∃ e∃ e∃ e∃ e∃ e@ e@ e@ e@ e∃
s2 e∃ e∃ e∃ e∃ e∃ e@ e@ e@ e@ e∃ 1.0
s4 e∃ e∃ e∃ e∃ e∃ e@ e@ e@ e@ e∃ 1.0
s6 e∃ e∃ e∃ e∃ e∃ e@ e@ e@ e@ e∃ 1.0
s34 e∃ e∃ e∃ e∃ e∃ e@ e@ e@ e@ e∃ 1.0
s42 e∃ e∃ e∃ e∃ e∃ e@ e@ e@ e@ e∀ 0.9
s50 e∀ e∀ e∃ e∃ e∃ e@ e@ e@ e@ e∃ 0.8
s52 e@ e@ e@ e@ e@ e@ e@ e@ e∀ e∃ 0.4
s56 e∃ e∃ e∃ e∃ e∃ e@ e@ e@ e@ e∃ 1.0
s60 e∃ e∃ e∃ e∃ e∃ e@ e@ e@ e@ e∃ 1.0
correct 78% 78% 89% 89% 89% 100% 100% 100% 89% 89%
1.0
1.0
● ●● ● ●●
● ●● ●●
● ●●●
●
●● ● ●
●
0.8
0.8
● ● ● ● ●
●
● ●●
0.6
0.6
U_a
U_a
●
0.4
0.4
● ● ●
● ●
●
0.2
0.2
●
0.0
0.0
c e o r c e o r
aspect aspect
(a) Personal structural process model under- (b) Estimated structural process model under-
standability values for the four aspects. standability values and 95% confidence in-
tervals for the four aspects.
four aspects concurrency (c), exclusiveness (e), order (o) and repetition (r) are depicted
in Figure 6.5a.
In order to test the hypothesis that the personal structural process model un-
derstandability values are normally distributed for each aspect (Hypothesis 6.1),
Table 6.4: Answers given for aspect order.
qo (A, B)
qo (A, C)
qo (A, D)
qo (A, E)
qo (B, A)
qo (B, C)
qo (B, D)
qo (B, E)
qo (C, A)
qo (C, B)
qo (C, D)
qo (C, E)
qo (D, A)
qo (D, B)
qo (D, C)
qo (D, E)
qo (E, A)
qo (E, B)
qo (E, C)
qo (E, D)
Uo (p, s)
subject
solution o∃ o∃ o∃ o∃ o@ o∃ o@ o@ o@ o∃ o@ o@ o@ o@ o@ o@ o@ o@ o@ o@
s1 o∀ o∀ o∀ o∀ o@ o∀ o@ o@ o@ o@ o@ o@ o@ o@ o@ o@ o@ o@ o@ o@ 0.70
s3 o@ o@ o@ o@ o@ o∃ o∀ o∀ o∃ o@ o∀ o∀ o∃ o∀ o∀ o∃ o∀ o∀ o∀ o∃ 0.10
s5 o∀ o∀ o∀ o∀ o@ o∃ o∃ o∃ o@ o∃ o∃ o∃ o@ o∃ o∃ o@ o@ o∃ o∃ o@ 0.40
s11 o∀ o∀ o∀ o∀ o@ o∃ o@ o∃ o@ o@ o∃ o∃ o@ o∃ o∃ o@ o@ o∃ o∃ o@ 0.40
s35 o∀ o∀ o∀ o∀ o@ o∀ o∃ o∃ o@ o@ o∃ o∃ o@ o∃ o∃ o@ o@ o∃ o∃ o@ 0.30
s51 o∀ o∀ o∀ o∀ o@ o∀ o∃ o∃ o@ o@ o∃ o∃ o@ o∃ o∃ o@ o∀ o∃ o∃ o@ 0.25
s53 o∀ o∀ o∀ o∀ o@ o∀ o@ o@ o@ o∀ o@ o@ o@ o@ o@ o@ o@ o@ o@ o@ 0.70
s55 o∀ o∀ o∀ o∀ o@ o∃ o@ o@ o@ o@ o@ o@ o@ o@ o@ o@ o@ o@ o@ o@ 0.75
s57 o∀ o∀ o∀ o∀ o@ o∀ o@ o@ o@ o@ o@ o@ o@ o@ o@ o@ o@ o@ o@ o@ 0.70
correct 0% 0% 0% 0% 100% 44% 56% 44% 89% 11% 44% 44% 89% 44% 44% 89% 78% 44% 44% 89%
6.5 experimental evaluation
129
130 measuring structural process model understandability
a Shapiro-Wilk test [142] was conducted for each of the four data sets. For concur-
rency, exclusiveness and repetition, the null-hypothesis that the data is normally
distributed had to be rejected (p 0.05). Only for order, this null-hypothesis
could not be rejected on the α = 0.05 level.
Possible reasons for not finding a normal distribution for concurrency, exclusive-
ness and repetition are:
• The process model is too “easy”. So, most values are near 1.0. As the value
range ends there, there cannot exist any bigger values “symmetric” to the
values lower than 1.0.
• The process model is too “small”. Only five and ten questions were asked
respectively. Consequently, personal structural process model understand-
ability values have a “step size” of 0.2 and 0.1 respectively.
• The number of subjects is too low. Only data from nine participants per
aspect could be collected.
Table 6.6: Estimated structural process model understandability values, standard devia-
tions and 95% confidence intervals for the four aspects.
[42], which does not require normally distributed data, was applied. The lower
and upper confidence interval bounds are also listed in Table 6.6.
Furthermore, the estimated structural process model understandability val-
ues and the 95% confidence intervals for the four aspects are also depicted in
Figure 6.5b.
For testing the hypothesis that the structural process model understandability
values for the four aspects are different (Hypothesis 6.2), Wilcoxon rank-sum tests
for independent values (aspects asked for in different questionnaire groups) [117,
pp. 590–597] and Wilcoxon signed-rank tests for paired values (aspects asked for
in one single questionnaire group) [117, pp. 599–603] were used. Both tests do not
require normally distributed data. Only for the combinations order-concurrency,
order-exclusiveness and order-repetition, the null-hypothesis (data belongs to the
same distribution) could be rejected on the α = 0.05 level.
Here again, a possible reason that the values for concurrency, exclusiveness and
repetition are so equal could be that the process model is too “small” and “easy”
so that no really difficult parts which are of varying difficulty for the different
aspects are included.
1.0
1.0
● ● ● ● ● ●
● ● ● ●
● ●
● ● ● ● ● ● ● ●
● ● ● ● ● ● ●
● ● ● ● ● ● ●
● ● ● ● ● ● ● ●
● ● ●
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
● ● ● ● ● ● ● ●
● ● ● ● ● ●
● ● ● ● ● ●
● ● ● ●
● ● ● ● ● ●
● ● ● ●
0.8
0.8
● ●
● ● ● ● ● ●
●
●
●
0.6
0.6
U_e
U_c
0.4
0.4
0.2
0.2
0.0
0.0
0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0
1.0
● ● ● ● ●
●
● ●
● ● ●
● ● ● ● ● ● ●
● ●
● ● ●
● ● ● ● ●
● ●
● ●
0.8
0.8
● ● ● ● ●
● ● ● ● ● ● ● ●
● ● ● ●
● ● ● ● ● ●
● ● ●
● ● ● ● ● ● ●
● ● ● ● ● ●
● ● ● ● ● ● ●
● ● ● ● ● ●
●
●
● ● ● ● ● ● ● ● ● ● ● ●
● ● ● ● ● ● ● ●
● ● ● ● ● ● ● ●
● ● ● ● ● ● ● ● ● ●
● ● ● ● ● ● ● ● ●
● ● ● ● ● ● ● ● ●
0.6
0.6
● ● ● ● ● ● ●
● ● ●
●
● ● ● ● ● ● ● ● ● ● ●
●
● ● ● ● ● ● ● ● ● ●
● ●
●
● ● ● ● ● ● ● ● ● ● ●
● ● ● ● ● ● ●
● ● ● ● ● ● ● ● ● ● ● ● ●
● ●
● ●
● ●
● ●
● ● ● ● ● ● ●
● ● ● ● ●
● ● ● ● ● ● ● ● ● ● ●
● ●
● ●
● ● ● ● ● ● ● ● ● ●
● ● ● ●
● ●
U_o
● ● ● ●
U_r
● ● ● ● ● ● ● ● ●
● ● ●
● ● ●
● ● ● ● ● ● ● ● ●
● ● ●
● ● ●
● ● ●
● ●
● ● ● ● ● ● ● ● ● ● ●
● ● ●
● ●
●
● ● ● ● ● ● ● ● ● ●
● ●
● ● ●
● ●
● ● ● ● ● ● ●
● ● ● ● ● ●
● ●
● ● ● ● ● ● ● ●
● ●
● ●
● ●
● ●
● ●
● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
● ● ● ● ● ● ● ●
● ●
● ●
● ●
● ●
●
● ● ● ● ● ● ●
● ● ● ● ●
● ●
● ● ● ● ● ● ● ● ● ●
● ●
0.4
0.4
● ● ● ● ● ● ● ●
● ● ● ●
● ● ●
●
● ● ● ● ● ● ● ● ● ● ●
● ● ●
●
● ● ● ● ● ● ●
● ● ●
● ●
● ● ● ● ● ● ● ● ● ● ●
●
● ● ● ● ● ● ● ● ● ●
● ●
●
● ● ● ● ● ● ● ● ●
● ● ● ● ● ●
● ● ● ● ● ● ● ● ● ● ●
● ●
● ●
● ●
● ● ● ● ● ● ● ● ●
● ● ● ● ● ● ● ●
● ● ● ● ● ● ● ●
● ● ● ● ● ● ●
● ● ● ● ● ● ● ●
● ● ● ● ● ● ●
● ● ● ● ● ● ●
● ● ● ● ●
● ●
● ● ● ● ● ● ● ●
●
0.2
0.2
● ● ●
● ● ● ● ●
● ●
● ● ●
● ● ●
● ●
●
● ● ● ●
● ●
●
● ● ●
0.0
0.0
● ● ● ●
0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0
Figure 6.6: Estimated partial structural process model understandability values of the
four aspects depending on coverage rate.
structural process model understandability values exist. The smaller the coverage
rate, the higher the standard deviation and the number of values outside the
confidence interval.
Table 6.7: Data on estimated partial structural process model understandability values of
the four aspects (Part 1 of 2).
Validity Evaluation
Finally, the necessary validity evaluation (see paragraph Validity Evaluation in
Subsection B.3.3) of experiment 1 has to be carried out.
134 measuring structural process model understandability
Table 6.7: Data on estimated partial structural process model understandability values of
the four aspects (Part 2 of 2).
internal validity Internal validity refers to the fact that the effects observed
in the experiment are not caused by a factor which one has no control of or has
not measured (see Definition B.9).
Looking at the threats to internal validity mentioned in Subsection B.3.3, one
can make the following statements:
6.5 experimental evaluation 135
• History: During the experiment which lasted less than one hour, no events
occurred which were able to strongly influence the subjects.
• Maturation: The experiment was so short that factors as, for example,
fatigue, boredom or hunger had no big influence.
• Mortality: As all subjects finished the experiment, this threat played no role.
• Selection: As the subjects were randomly assigned to one of the two ques-
tionnaire groups, possible personal differences should have been balanced.
external validity External validity refers to the extent to which the re-
sults of an experiment can be generalized out of the scope of the study (see
Definition B.10).
Looking at the threats to external validity mentioned in Subsection B.3.3, one
can make the following statements:
• Ecological validity: The process model used in experiment 1 was only a very
small “toy problem”. Consequently, it is questionable whether the results
are generalizable to more realistic process models. This is one reason for
the conduction of the following second and larger experiment.
6.5.2 Experiment 2
Experiment Design
The second experiment was conducted as a cooperation with Jan Mendling
(Humboldt-Universität zu Berlin) and Hajo A. Reijers (Technische Universiteit Eind-
hoven). Its goal was to (re)test the hypotheses with a larger and more realistic
process model as well as with much more subjects.
136 measuring structural process model understandability
Start
XOR
A B
XOR
AND
XOR XOR
AND AND
F J
C D E H I L
G K
AND AND
XOR
XOR
AND
End
object For experiment 2, the process model depicted in Figure 6.7 was used.
It was presented to the subjects in the same notation as in experiment 1.
As the process model has 12 tasks, the number of possible questions for the four
aspects are |Qc,max (p)| = |Qe,max (p)| = 66, |Qo,max (p)| = 132 and |Qr,max (p)| = 12
(cf. Theorem 6.2).
Results
The answers to the questionnaire are given in Tables 6.9–6.17.
single questions First, the rate of correctly answered questions on the four
aspects were analyzed. These values are given in Table 6.18 and as histograms in
Figure 6.8.
As one can see, the aspect order has a quite different behavior compared to the
other three aspects. While those have narrow peaks near the rate of 1.0, the values
for aspect order are more spread over the whole interval with the peak near 0.6.
So, most questions on the aspect order seem to be more difficult to answer by the
subjects than those on the other aspects.
Next, it was analyzed whether a connection between the rates of correct answers
to questions about concurrency and exclusiveness for the same pair of tasks exist.
As both aspects deal with the execution of task pairs during a process instance
execution, such a connection is imaginable. The single value pairs are depicted in
Figure 6.9a. As Spearman’s rank correlation coefficient (see Section C.2) is only
0.465, there is no strong connection.
Afterwards, the same analysis was done for the aspect order and order in reverse
ordering. The single value pairs are depicted in Figure 6.9b. Here, Spearman’s
rank correlation coefficient is −0.209. So, knowing the rate of correct answers to
question qo (t1 , t2 ), no prediction for qo (t2 , t1 ) can be given.
qo (A,B)
qo (A,F)
qo (A,J)
qo (B,C)
qo (B,G)
qo (B,K)
qo (C,D)
qo (C,H)
qo (C,L)
qo (D,E)
qo (D,I)
qo (E,A)
qo (E,F)
qo (E,J)
qo (F,B)
qo (F,G)
qo (F,K)
qo (G,C)
qo (G,H)
qo (G,L)
qo (H,D)
qo (H,I)
qo (I,A)
qo (I,E)
qo (I,J)
qo (J,B)
qo (J,F)
qo (J,K)
qo (K,C)
qo (K,G)
qo (K,L)
qo (L,D)
qo (L,H)
subject Uo (p, s, Qo )
solution o@ o∃ o∃ o∃ o∃ o∃ o@ o@ o∃ o@ o@ o@ o@ o@ o @ o∃ o ∃ o@ o@ o∃ o@ o@ o@ o@ o@ o@ o@ o∀ o@ o∃ o∃ o@ o@
s001 o@ o∀ o∀ o∃ o∃ o∃ o@ o@ o∃ o@ o@ o@ o@ o∃ o @ o∀ o ∃ o@ o@ o∃ o@ o@ o@ o@ o∃ o@ o@ o∀ o@ o@ o∃ o@ o∃ 0.79
s002 o@ o∀ o∀ o∀ o∀ o@ o∀ o∀ o∀ o@ o∃ o@ o@ o∃ o @ o∀ o ∃ o@ o@ o∃ o∃ o@ o@ o∃ o∃ o@ o∃ o∀ o@ o∃ o∃ o∃ o∃ 0.48
s003 o∃ o∃ o∃ o∃ o∀ o∀ o@ o∃ o∃ o@ o∃ o@ o∃ o@ o ∃ o∀ o ∃ o@ o∃ o∃ o@ o@ o@ o@ o∃ o@ o∃ o∀ o@ o∃ o∃ o@ o@ 0.67
s004 o∃ o∃ o∃ o∃ o∃ o∃ o@ o@ o∃ o@ o∃ o∃ o∃ o@ o ∃ o∀ o ∃ o∃ o@ o∃ o∃ o@ o@ o@ o∃ o∃ o@ o∀ o∃ o∃ o@ o∃ o@ 0.61
s005 o@ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o@ o∃ o∃ o∃ o@ o ∃ o∀ o ∃ o@ o@ o∃ o∃ o@ o∃ o@ o∃ o@ o∃ o∀ o@ o∃ o∀ o@ o∃ 0.64
s006 o∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∀ o@ o∃ o∃ o∃ o@ o ∃ o∀ o ∀ o@ o∃ o∃ o∃ o@ o@ o@ o∃ o∃ o@ o∀ o@ o∃ o@ o@ o@ 0.58
s007 o∃ o∀ o∀ o∀ o∀ o∀ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o @ o∀ o ∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o@ o∃ o∀ o∃ o∃ o@ o∃ o∃ 0.27
s008 o∀ o∀ o∀ o∀ o∀ o∀ o@ o∃ o∃ o@ o@ o∃ o@ o@ o ∃ o∀ o ∃ o@ o∀ o∃ o@ o@ o∃ o@ o@ o∃ o@ o∀ o@ o∃ o@ o@ o∃ 0.55
s009 o@ o∀ o∀ o∃ o∃ o∀ o@ o∃ o∃ o@ o@ o@ o@ o@ o @ o∀ o ∀ o@ o@ o∃ o@ o@ o@ o@ o@ o@ o@ o∀ o@ o@ o∃ o@ o@ 0.79
s010 o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o∃ o∃ o ∃ o∀ o ∃ o@ o∃ o∃ o∃ o@ o∃ o∃ o∃ o∃ o∃ o∀ o@ o∃ o@ o∃ o∃ 0.42
s011 o∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∀ o@ o∃ o∃ o∃ o@ o ∃ o∀ o ∃ o@ o∃ o∃ o@ o@ o∃ o∃ o@ o∃ o∃ o∀ o@ o∃ o@ o@ o∃ 0.55
s012 o∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o@ o∃ o∃ o∃ o@ o ∃ o∀ o ∃ o@ o∃ o∃ o∃ o@ o∃ o@ o∃ o∃ o∃ o∀ o@ o∃ o∀ o∃ o∃ 0.52
s013 o∃ o∃ o∀ o∀ o∀ o∃ o@ o∃ o∃ o@ o∃ o∃ o∃ o∃ o @ o∃ o ∃ o@ o∃ o∃ o@ o@ o@ o@ o∃ o@ o∃ o∃ o@ o∃ o∃ o@ o∃ 0.58
s014 o@ o∃ o∀ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o @ o∀ o ∃ o∃ o@ o∃ o∃ o∃ o@ o∃ o∃ o@ o∃ o∀ o∃ o∃ o∃ o∃ o@ 0.52
s015 o@ o∀ o∀ o∀ o∀ o∀ o@ o@ o∃ o@ o@ o@ o∃ o∃ o @ o∀ o ∃ o@ o@ o∃ o∃ o@ o@ o@ o@ o@ o@ o∀ o∃ o∃ o∀ o@ o@ 0.67
s016 o@ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o@ o∃ o∃ o@ o@ o ∃ o∀ o ∃ o@ o@ o∃ o@ o@ o∃ o@ o@ o∃ o∃ o∀ o@ o∃ o∀ o@ o∃ 0.70
s017 o∃ o∀ o∀ o∀ o∀ o∀ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o @ o∀ o ∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o∃ o∃ o∀ o∃ o∃ o∀ o∃ o∃ 0.24
s018 o@ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o@ o@ o@ o@ o@ o @ o∀ o ∃ o@ o@ o∃ o@ o@ o@ o@ o@ o@ o∃ o∀ o@ o∃ o∀ o@ o∃ 0.85
correct 44% 61% 50% 67% 61% 61% 72% 17% 83% 83% 28% 50% 33% 56% 50% 6% 89% 78% 50% 100% 44% 83% 67% 67% 33% 56% 33% 94% 72% 89% 33% 61% 33%
6.5 experimental evaluation
139
Table 6.10: Answers given in group 2.
qo (H,A)
qo (K,H)
qo (D,A)
qo (A,G)
qo (G,D)
qo (A,K)
qo (K,D)
qo (A,C)
qo (B,H)
qo (H,E)
qo (L,A)
qo (B,D)
qo (E,G)
qo (E,K)
qo (D,F)
qo (C,E)
qo (F,H)
qo (B,L)
qo (E,B)
qo (L,E)
qo (F,C)
qo (H,J)
qo (D,J)
qo (G,I)
qo (J,G)
qo (F,L)
qo (I,K)
qo (C,I)
qo (J,C)
qo (I,B)
qo (L,I)
qo (J,L)
qo (I,F)
measuring structural process model understandability
subject Uo (p, s, Qo )
solution o∃ o∃ o∃ o∃ o∃ o∃ o@ o@ o@ o@ o@ o@ o∃ o∃ o@ o@ o∃ o@ o@ o@ o@ o@ o@ o@ o∃ o@ o∃ o∃ o@ o@ o@ o@ o@
s019 o∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o@ o∃ o∃ o∃ o@ o∃ o@ o∃ o@ o∃ o∀ o@ o∃ o∃ o@ o∃ 0.58
s020 o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∀ o@ o∃ o∃ o∃ o@ o∃ o∃ o∃ o@ o∃ o∀ o@ o∃ o∃ o∃ o∃ 0.45
s021 o∀ o∀ o∀ o∀ o∀ o∀ o@ o∃ o@ o@ o@ o@ o∃ o∃ o@ o@ o∃ o@ o@ o@ o@ o@ o@ o@ o∃ o@ o∃ o∀ o@ o@ o@ o@ o@ 0.76
s022 o∃ o∃ o∃ o∃ o∃ o∃ o@ o@ o@ o@ o@ o∃ o∃ o∃ o@ o@ o∃ o@ o@ o@ o@ o@ o@ o@ o∃ o@ o∃ o∀ o@ o@ o@ o@ o@ 0.94
s023 o∃ o∀ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∀ o∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o@ o∃ o∃ o∀ o@ o∃ o∃ o@ o∃ o∃ o∀ 0.36
s024 o∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o∃ o@ o∃ o∀ o∀ o@ o@ o∃ o@ o∃ o@ o@ o@ o∃ o∃ o∃ o@ o∀ o∃ o@ o∃ o@ o@ o∃ 0.64
s025 o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o@ o∃ o∃ o∃ o∃ o@ o∃ o∃ o∃ o∀ o∃ o∃ o∃ o∃ o∃ 0.42
s026 o∃ o∃ o∃ o∃ o∃ o∃ o@ o@ o@ o∃ o@ o@ o∃ o∃ o@ o@ o∃ o@ o@ o@ o@ o@ o@ o@ o∃ o@ o∃ o∀ o@ o∃ o@ o@ o@ 0.91
s027 o∀ o∀ o∀ o∀ o∀ o∀ o@ o@ o@ o∃ o@ o@ o∃ o∃ o∃ o@ o∃ o∃ o@ o@ o∃ o∃ o@ o@ o∃ o@ o∃ o∃ o∃ o∃ o∀ o∃ o∃ 0.52
s028 o∀ o@ o∀ o@ o∃ o∀ o@ o@ o∃ o@ o@ o@ o@ o∃ o∃ o@ o@ o@ o@ o@ o@ o@ o∃ o@ o@ o∃ o@ o∀ o@ o@ o@ o@ o@ 0.58
s029 o∀ o∀ o∀ o∀ o∀ o∀ o@ o∃ o@ o∃ o@ o∀ o∃ o∃ o@ o∀ o∃ o∃ o∀ o∀ o∃ o∃ o∀ o∀ o@ o∃ o∃ o∀ o∃ o∃ o∀ o∃ o∃ 0.24
s030 o∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o@ o@ o∃ o∃ o∃ o∃ o@ o∃ o∃ o∀ o∃ o@ o∃ o@ o@ o@ o∃ o@ o∃ 0.58
s031 o∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o∃ o@ o∃ o∃ o∃ o@ o∀ o∃ o@ o∀ o∃ o@ o@ o∃ o∃ o∃ o@ o∃ o∀ o∃ o∃ o∃ o@ o∃ 0.58
s032 o∀ o∃ o∃ o∀ o∀ o∀ o@ o∃ o∃ o∃ o@ o∃ o∃ o∃ o@ o∃ o∃ o@ o∃ o∃ o@ o@ o∃ o∃ o∃ o@ o∃ o∀ o@ o∃ o@ o@ o∃ 0.52
s033 o@ o∀ o@ o∃ o@ o∀ o@ o@ o@ o@ o@ o@ o∃ o∃ o∃ o@ o∃ o@ o@ o∃ o@ o@ o@ o@ o∃ o@ o∃ o@ o@ o∃ o@ o@ o∃ 0.70
s034 o∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o∃ o@ o∃ o∃ o@ o@ o∃ o∃ o@ o∃ o∃ o@ o@ o∃ o∃ o@ o@ o∃ o∀ o@ o∃ o∃ o@ o∃ 0.55
s035 o∃ o∃ o∃ o∃ o∃ o∃ o@ o@ o@ o@ o@ o@ o∃ o∀ o∃ o@ o∃ o@ o@ o@ o@ o@ o@ o@ o∃ o∀ o∃ o∀ o@ o@ o@ o@ o@ 0.88
s036 o∀ o∃ o∀ o∀ o∀ o∃ o∃ o∃ o@ o∃ o∃ o@ o∃ o∃ o∃ o@ o∃ o∃ o@ o@ o∃ o∃ o@ o@ o∃ o∃ o∃ o∀ o∃ o∃ o@ o∃ o∃ 0.45
s037 o∀ o∀ o∀ o∀ o∀ o∀ o@ o@ o@ o@ o@ o@ o∃ o∃ o@ o∃ o∃ o@ o@ o@ o@ o@ o@ o@ o∃ o@ o∃ o∀ o@ o@ o@ o@ o@ 0.76
s038 o∀ o∀ o∀ o∀ o∀ o∀ o@ o@ o@ o@ o@ o@ o∃ o∃ o@ o@ o∃ o@ o@ o@ o@ o@ o@ o@ o∃ o@ o∃ o∀ o@ o@ o@ o∃ o@ 0.76
correct 55% 60% 60% 60% 60% 60% 80% 40% 50% 35% 70% 45% 85% 85% 65% 60% 90% 70% 55% 55% 65% 70% 50% 60% 85% 70% 85% 15% 70% 40% 55% 65% 35%
140
Table 6.11: Answers given in group 3.
qo (A,D)
qo (A,H)
qo (A,L)
qo (B,E)
qo (B,I)
qo (C,A)
qo (C,F)
qo (C,J)
qo (D,B)
qo (D,G)
qo (D,K)
qo (E,C)
qo (E,H)
qo (E,L)
qo (F,D)
qo (F,I)
qo (G,A)
qo (G,E)
qo (G,J)
qo (H,B)
qo (H,F)
qo (H,K)
qo (I,C)
qo (I,G)
qo (I,L)
qo (J,D)
qo (J,H)
qo (K,A)
qo (K,E)
qo (K,I)
qo (L,B)
qo (L,F)
qo (L,J)
subject Uo (p, s, Qo )
solution o∃ o∃ o∃ o∃ o∃ o@ o@ o@ o@ o∃ o∃ o@ o@ o∃ o@ o@ o@ o@ o@ o@ o@ o∃ o@ o@ o∃ o@ o@ o@ o@ o@ o@ o@ o∃
s039 o∃ o∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∀ o∃ o@ o∃ o∀ o@ o∃ o∃ o@ o@ o∃ o∃ o@ o@ o∃ o∃ o@ o∃ o∃ o@ o∃ o∃ o∃ o@ 0.45
s040 o∃ o∃ o∃ o∃ o∃ o@ o@ o∃ o@ o∃ o∃ o@ o@ o∃ o@ o@ o@ o@ o@ o@ o@ o∃ o@ o@ o∃ o@ o@ o@ o@ o@ o@ o@ o∀ 0.94
s041 o∃ o∀ o∀ o∀ o∀ o@ o∃ o∃ o@ o∃ o∃ o∃ o∃ o∃ o∃ o@ o@ o∃ o∃ o@ o@ o∃ o∃ o@ o∃ o∃ o∃ o@ o∃ o∃ o@ o∃ o@ 0.45
s042 o∃ o∃ o∃ o∃ o∃ o@ o∃ o@ o@ o∃ o∀ o@ o∃ o∃ o∃ o@ o@ o@ o@ o@ o∃ o∃ o@ o∃ o∃ o∀ o∃ o@ o@ o∃ o@ o∃ o∃ 0.70
s043 o∀ o∃ o∃ o∀ o∃ o∃ o@ o∃ o@ o∃ o@ o∃ o∃ o∃ o∃ o@ o∃ o∃ o∃ o@ o@ o∃ o@ o@ o∃ o∃ o∃ o@ o∃ o∃ o@ o∃ o∀ 0.48
s044 o∃ o∃ o∃ o∃ o∃ o@ o@ o@ o@ o∃ o∃ o@ o@ o∃ o@ o∃ o@ o@ o@ o@ o@ o∃ o@ o@ o∃ o@ o@ o@ o@ o@ o@ o@ o∃ 0.97
s045 o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o@ o∃ o∃ o∃ o∃ o∃ o∃ o@ o@ o∃ o∃ o@ o@ o∃ o∃ o@ o∃ o∃ o∃ o@ o∃ o∃ o@ o∃ o∃ 0.61
s046 o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o∃ o∃ o@ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o@ 0.39
s047 o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o∃ o@ 0.39
s048 o@ o@ o@ o∀ o∀ o@ o∃ o@ o@ o∃ o∃ o@ o∃ o∀ o@ o@ o@ o@ o@ o@ o@ o∃ o∃ o@ o∃ o@ o∃ o@ o∃ o∃ o@ o∃ o@ 0.58
s049 o∃ o∃ o∃ o@ o@ o∃ o@ o@ o∃ o∃ o∀ o@ o∃ o∃ o∃ o@ o∃ o∃ o∃ o@ o@ o∃ o@ o@ o∃ o@ o∃ o∃ o∃ o∃ o∃ o∃ o@ 0.48
s050 o∃ o∃ o@ o∃ o∃ o@ o@ o@ o@ o∃ o∃ o∃ o∃ o∃ o@ o@ o@ o∃ o∃ o@ o@ o∃ o∃ o@ o∃ o@ o@ o@ o∃ o∃ o@ o∃ o∃ 0.73
s051 o∃ o∃ o∃ o∃ o∃ o∃ o@ o@ o∃ o@ o@ o@ o@ o@ o@ o@ o∃ o@ o@ o∃ o@ o@ o@ o@ o@ o@ o@ o∃ o@ o@ o∃ o@ o@ 0.64
s052 o∃ o∃ o∃ o∃ o∃ o@ o@ o@ o@ o∃ o∃ o∃ o∃ o∀ o∃ o@ o@ o∃ o∃ o∃ o@ o∃ o∃ o@ o∀ o@ o∃ o@ o∃ o∃ o@ o@ o@ 0.61
s053 o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∀ o@ o@ o∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o∃ o@ o∃ o∃ o@ o∃ o∃ o@ o∃ o∃ o∃ o∃ 0.42
s054 o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o@ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o@ o∃ o∃ o∃ o@ o∃ o∃ o∃ o@ o∃ o∃ o@ o∃ o@ 0.52
s055 o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∀ o∀ o@ o∃ o∀ o@ o∃ o∃ o@ o@ o∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o@ o∃ o∃ o∃ o∀ 0.39
s056 o∃ o∃ o∃ o∃ o∀ o@ o∃ o∃ o@ o∃ o∃ o∃ o∃ o∃ o∃ o@ o@ o∃ o∃ o@ o@ o∃ o∃ o@ o∃ o∃ o∃ o@ o∃ o∃ o@ o∃ o∃ 0.58
s057 o∀ o∀ o∀ o∀ o∀ o@ o@ o∃ o@ o∀ o∀ o@ o@ o∀ o@ o@ o@ o@ o∃ o@ o@ o∀ o@ o∀ o∀ o∃ o@ o@ o@ o@ o@ o@ o@ 0.55
s058 o∃ o∀ o∃ o∀ o∀ o@ o@ o∀ o@ o@ o@ o@ o∃ o@ o∃ o@ o@ o∃ o@ o∃ o@ o∃ o@ o@ o∃ o@ o∃ o@ o@ o∃ o@ o∃ o@ 0.55
s059 o∃ o∃ o∃ o∃ o∃ o@ o@ o∃ o@ o∃ o∀ o@ o@ o∀ o@ o@ o@ o@ o∃ o@ o@ o∃ o@ o@ o∃ o@ o∃ o@ o∀ o@ o@ o@ o∃ 0.82
correct 86% 81% 81% 71% 71% 62% 48% 38% 67% 71% 57% 62% 24% 62% 43% 71% 62% 48% 43% 62% 71% 86% 52% 67% 86% 57% 24% 67% 48% 24% 71% 29% 33%
6.5 experimental evaluation
141
Table 6.12: Answers given in group 4.
qo (D,H)
qo (H,G)
qo (H,C)
qo (G,K)
qo (D,C)
qo (C,G)
qo (H,L)
qo (C,K)
qo (A,E)
qo (B,A)
qo (D,L)
qo (E,D)
qo (G,B)
qo (L,G)
qo (K,B)
qo (L,K)
qo (C,B)
qo (G,F)
qo (L,C)
qo (K,F)
qo (F,A)
qo (I,H)
qo (B,F)
qo (A,I)
qo (I,D)
qo (J,A)
qo (F,E)
qo (K,J)
qo (E,I)
qo (B,J)
qo (J,E)
qo (F,J)
qo (J,I)
measuring structural process model understandability
subject Uo (p, s, Qo )
solution o∃ o∃ o@ o∃ o∃ o@ o∃ o∃ o@ o@ o∃ o@ o@ o@ o@ o@ o@ o@ o∃ o@ o@ o∃ o@ o@ o@ o@ o@ o@ o@ o∃ o@ o∃ o∃
s060 o∃ o∃ o@ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o@ o∃ o@ o@ o@ o@ o@ o∃ o@ o@ o∃ o∃ o@ o@ o@ o∃ o@ o∃ o@ o@ o∃ o@ 0.76
s061 o∀ o∀ o@ o∀ o∀ o@ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o@ o@ o∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o∃ o∃ o@ o∃ o∃ o@ 0.36
s062 o∃ o∃ o@ o∃ o∃ o∃ o@ o@ o@ o@ o@ o∀ o@ o∃ o@ o@ o∃ o@ o@ o@ o@ o@ o@ o@ o∃ o@ o@ o∃ o@ o∀ o@ o@ o∃ 0.61
s063 o∀ o∃ o∃ o∀ o∀ o@ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o@ o@ o∃ o∃ o@ o∃ o∃ o@ o@ o∃ o∃ o∃ o@ o∃ o∃ o@ o@ 0.45
s064 o∃ o∃ o@ o∃ o∃ o@ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o∃ o@ o∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o∃ o∃ o@ o∃ o∃ o∃ 0.48
s065 o∃ o∃ o∃ o∃ o∃ o∃ o∀ o∀ o@ o∃ o∀ o∃ o∃ o∃ o@ o@ o∃ o@ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o∃ o@ o@ - o@ 0.36
s066 o∀ o∀ o@ o∀ o∀ o@ o∃ o∃ o@ o∃ o∃ o@ o∃ o@ o@ o@ o@ o@ o∃ o@ o∃ o∃ o@ o@ o@ o∀ o∃ o@ o@ o@ o@ o∃ o@ 0.67
s067 o∃ o∀ o∃ o∀ o@ o@ o∃ o∃ o@ o∃ o∃ o@ o∃ o∃ o∃ o∃ o@ o∃ o∃ o∃ o∃ o∃ o@ o@ o@ o@ o∃ o@ o∃ o@ o@ o∃ o@ 0.52
s068 o∃ o∃ o@ o∃ o∃ o∃ o∀ o∃ o@ o∃ o∀ o@ o∃ o∃ o@ o@ o∃ o@ o∃ o@ o∃ o∃ o@ o@ o∃ o@ o∃ o∃ o∃ o@ o@ o∃ o@ 0.58
s069 o∃ o∃ o@ o∃ o∃ o@ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o@ o@ o∃ o∃ o@ o∃ o∃ o∃ o@ o∃ o∃ o@ o∃ o@ o∃ o∃ o∃ 0.58
s070 o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o@ o∃ o∃ o@ o@ o∃ o@ o∃ o@ o∃ o∃ o@ o@ o∃ o∃ o∃ o∃ o∃ o@ o@ o∃ o@ 0.58
s071 o∀ o∀ o@ o∀ o∀ o@ o∃ o∃ o@ o∃ o∃ o@ o∃ o@ o@ o@ o@ o@ o∃ o@ o∀ o∃ o∃ o@ o@ o@ o∃ o@ o@ o@ o@ o∃ o∃ 0.70
s072 o∀ o∀ o@ o∀ o∀ o@ o∃ o∃ o@ o∃ o∃ o@ o∃ o@ o@ o∃ o@ o@ o∃ o@ o@ o∃ o@ o@ o@ o@ o∃ o@ o∃ o∃ o@ o∃ o@ 0.70
s073 o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o∀ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o@ 0.33
s074 o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o@ o∃ o∃ o@ o∃ o∃ o@ o∃ o@ o∃ o∃ o@ o@ o∃ o∃ o∃ o∃ o∃ o@ o@ o∃ o@ 0.55
s075 o∃ o∀ o∃ o∀ o∀ o@ o∀ o∀ o@ o∃ o∀ o@ o∃ o@ o@ o@ o@ o@ o@ o@ o∃ o∃ o@ o@ o@ o@ o∃ o@ o∃ o@ o@ o∃ o@ 0.55
s076 o∃ o∃ o@ o∃ o∃ o@ o∃ o∃ o@ o@ o∃ o@ o@ o@ o@ o∃ o@ o@ o∃ o@ o@ o∃ o@ o@ o@ o@ o@ o@ o@ o∃ o@ o∃ o∃ 0.97
s077 o∃ o∃ o∃ o∃ o∃ o∃ o∃ o∃ o@ o∃ o∃ o@ o∃ o∃ o@ o@ o∃ o@ o∃ o@ o@ o∃ o@ o@ o∃ o@ o∃ o∃ o∃ o@ o@ o∃ o∃ 0.67
s078 o∀ o@ o@ o∃ o∃ o@ o∀ o∀ o@ o∃ o∃ o@ o∃ o@ o@ o@ o@ o@ o∃ o@ o@ o∃ o@ o@ o@ o@ o∃ o@ o∃ o∃ o@ o@ o@ 0.70
s079 o∃ o∃ o∃ o∃ o∃ o∃ o∃ o@ o@ o∃ o∀ o@ o∃ o∃ o@ o@ o∃ o∃ o@ o@ o∃ o∃ o@ o@ o∃ o@ o∃ o∃ o∃ o@ o∃ o∃ o@ 0.45
correct 70% 65% 55% 65% 65% 55% 75% 75% 75% 10% 75% 65% 10% 55% 70% 55% 55% 90% 85% 65% 40% 95% 60% 75% 60% 60% 10% 45% 25% 20% 70% 80% 30%
142
Table 6.13: Answers given in group 5.
qc (A,B)
qc (A,F)
qc (A,J)
qc (B,D)
qc (B,H)
qc (B,L)
qc (C,G)
qc (C,K)
qc (D,G)
qc (D,K)
qc (E,H)
qc (E,L)
qc (F,J)
qc (G,I)
qc (H,I)
qc (I,J)
qc (J,L)
Uc (p, s, Qc )
qe (A,D)
qe (A,H)
qe (A,L)
qe (B,F)
qe (B,J)
qe (C,E)
qe (C,I)
qe (D,E)
qe (D,I)
qe (E,F)
qe (E,J)
qe (F,H)
qe (F,L)
qe (G,K)
qe (H,K)
qe (I,L)
Ue (p, s, Qe )
subject
solution c@ c@ c@ c@ c@ c@ c∃ c∃ c∃ c∃ c∃ c∃ c∃ c@ c∃ c∃ c@ e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∀ e∃ e∃ e∀ e@ e∃ e∃ e∃ e∃
correct 83% 89% 83% 89% 83% 89% 83% 56% 83% 56% 94% 89% 94% 94% 39% 100% 83% 72% 94% 94% 94% 78% 100% 100% 100% 100% 100% 83% 94% 94% 100% 100% 100%
6.5 experimental evaluation
143
Table 6.14: Answers given in group 6.
Ue (p, s, Qe )
Uc (p, s, Qc )
qc (D,H)
qe (G,H)
qc (A,G)
qc (C,H)
qc (A,K)
qc (A,C)
qc (C,D)
qe (H,L)
qe (A,E)
qc (D,L)
qe (B,G)
qe (E,G)
qe (G,L)
qe (B,K)
qe (E,K)
qe (D,F)
qc (K,L)
qe (B,C)
qc (C,L)
qe (C,F)
qc (F,G)
qc (B,E)
qc (F,K)
qc (H,J)
qe (A,I)
qe (D,J)
qc (G,J)
qe (J,K)
qc (I,K)
qe (C,J)
qc (B,I)
qc (E,I)
qe (F,I)
subject
measuring structural process model understandability
solution c@ c@ c@ c@ c@ c∀ c∃ c∃ c∃ c∃ c∃ c@ c∃ c∃ c∃ c∃ c@ e∃ e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∀ e∃ e∀ e@ e@ e∃ e∃ e∀
s098 c@ c@ c@ c@ c@ c∀ c∃ c∃ c∃ c∃ c∃ c@ c∃ c∃ c∃ c∃ c@ 1.00 e∃ e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∀ e∃ e∀ e@ e@ e∃ e∃ e∀ 1.00
s099 c∃ c∃ c∃ c∃ c∃ c∀ c∃ c∃ c∃ c∃ c∀ c@ c∃ c∃ c∀ c∃ c∀ 0.53 e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∀ e∃ e@ e@ e∃ e∃ e∃ e∀ 0.81
s100 c@ c@ c∃ c∃ c@ c@ c∃ c∃ c∃ c∃ c∃ c@ c∃ c∃ c∃ c∃ c@ 0.82 e∃ e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∃ e∃ e∀ e@ e∃ e∃ e∃ e∀ 0.88
s101 c∃ c∃ c∃ c∃ c∃ c∀ c∃ c∃ c∃ c∃ c∃ c∀ c∃ c∃ c∃ c∃ c∃ 0.59 e∃ e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∀ e∃ e∀ e@ e@ e∃ e∃ e∀ 1.00
s102 c@ c@ c@ c@ c@ c∀ c∃ c∃ c∃ c∃ c∃ c@ c∃ c∃ c∃ c∃ c@ 1.00 e∃ e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∀ e∃ e∀ e@ e@ e∃ e∃ e∀ 1.00
s103 c@ c@ c@ c@ c@ c∀ c∃ c∃ c∃ c∃ c∃ c@ c∃ c∃ c∃ c∃ c@ 1.00 e∃ e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∀ e∃ e∀ e@ e@ e∃ e∃ e∀ 1.00
s104 c@ c@ c@ c@ c@ c∀ c∀ c∃ c∀ c∃ c∀ c@ c∃ c∃ c∀ c∃ c@ 0.76 e∃ e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∀ e∃ e∀ e@ e@ e∃ e∃ e∀ 1.00
s105 c@ c@ c@ c@ c@ c∀ c∃ c∃ c∃ c∃ c∃ c@ c∃ c∃ c∃ c∃ c@ 1.00 e∃ e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∀ e∃ e∀ e@ e@ e∃ e∃ e∀ 1.00
s106 c∀ c∃ c∃ c∀ c∃ c∀ c∃ c∃ c∃ c∃ c∃ c∃ c∃ c∃ c∃ c∃ c∃ 0.59 e∃ e∃ e∀ e∃ e∃ e∃ e∀ e∃ e∀ e∃ e∀ e∃ e@ e∃ e∃ e∀ 0.88
s107 c@ c@ c@ c@ c@ c∃ c∃ c∃ c∃ c∃ c∃ c@ c∃ c∃ c∃ c∃ c@ 0.94 e∃ e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∀ e∃ e∀ e@ e@ e∃ e∃ e∀ 1.00
s108 c@ c@ c∃ c@ c@ c∀ c∃ c∃ c∃ c∃ c∃ c@ c∃ c@ c∃ c∃ c∃ 0.82 e∃ e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∀ e∃ e∀ e@ e@ e∃ e∃ e∀ 1.00
s109 c@ c@ c∃ c∀ c∃ c∃ c∃ c∃ c∃ c@ c∃ c∃ c∃ c∃ c@ c∃ c@ 0.59 e∃ e∀ e∃ e∃ e∀ e∃ e∀ e∀ e∃ e∃ e∀ e∀ e@ e@ e∃ e∀ 0.62
s110 c@ c@ c@ c@ c@ c∀ c∃ c@ c∃ c@ c∃ c@ c@ c@ c∃ c@ c@ 0.71 e∃ e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∃ e∃ e∀ e@ e@ e∃ e∃ e∀ 0.94
s111 c∃ c∃ c∃ c@ c∃ c∀ c∃ c∃ c∃ c∃ c∃ c@ c∃ c∃ c∃ c∃ c@ 0.76 e∃ e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∀ e∃ e∀ e@ e@ e∃ e∃ e∀ 1.00
s112 c@ c@ c@ c@ c@ c∀ c∃ c∃ c∃ c∃ c∃ c@ c∃ c∃ c∃ c∃ c@ 1.00 e∃ e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∀ e∃ e∀ e@ e@ e∃ e∃ e∀ 1.00
s113 c@ c@ c@ c@ c@ c∀ c∃ c∃ c∃ c∃ c∃ c@ c∃ c∃ c∃ c∃ c@ 1.00 e∃ e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∀ e∃ e∀ e@ e@ e∃ e∃ e∀ 1.00
s114 c@ c@ c@ c@ c@ c∀ c∃ c∃ c∃ c∃ c∃ c@ c∃ c∃ c∃ c∃ c@ 1.00 e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e@ e@ e∃ e∃ e∃ 0.75
s115 c@ c@ c@ c∃ c@ c∃ c∃ c∃ c∃ c∃ c@ c@ c∃ c∃ c∃ c∃ c@ 0.82 e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∀ e@ e@ e∃ e∃ e∀ 0.94
s116 c@ c@ c@ c@ c@ c∀ c∃ c@ c∃ c@ c∃ c@ c@ c@ c∃ c@ c@ 0.71 e∃ e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∀ e∃ e∀ e@ e@ e∃ e∃ e∀ 1.00
s117 c∃ c@ c@ c@ c∃ c∀ c∃ c@ c∃ c∃ c@ c@ c@ c∃ c@ c@ c@ 0.59 e∃ e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∀ e∃ e∀ e@ e@ e∃ e@ e∀ 0.94
s118 c@ c@ c@ c@ c@ c∀ c∃ c∀ c∃ c@ c∃ c@ c∃ c∃ c∃ c∃ c@ 0.88 e∃ e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∀ e∃ e∀ e@ e@ e∃ e∃ e∃ 0.94
s119 c@ c@ c@ c@ c@ c∀ c∃ c∃ c∀ c∃ c∀ c∃ c∃ c∃ c∀ c∃ c@ 0.76 e∃ e∃ e∃ e∃ e∀ e∃ e∀ e∃ e∀ e∃ e∀ e@ e@ e∃ e∃ e∀ 0.94
s120 c@ c@ c@ c@ c@ c∀ c∃ c∃ c∃ c∃ c∃ c@ c∃ c∃ c∃ c∃ c@ 1.00 e∃ e∃ e@ e∃ e∃ e∃ e∀ e∃ e∀ e∃ e∀ e@ e@ e∃ e∃ e∀ 0.94
s121 c@ c@ c@ c@ c@ c∀ c∃ c∃ c∃ c∃ c∃ c@ c∃ c∃ c∃ c∃ c@ 1.00 e∃ e∃ e∃ e∃ e∀ e∃ e∀ e∃ e∀ e∃ e∀ e@ e@ e∃ e∃ e∀ 0.94
s122 c@ c@ c@ c@ c@ c∀ c∃ c∃ c∃ c∃ c∃ c@ c∃ c∃ c∃ c∃ c@ 1.00 e∃ e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∀ e∃ e∀ e@ e@ e∃ e∃ e∀ 1.00
correct 80% 84% 72% 76% 76% 84% 96% 84% 92% 84% 80% 84% 88% 88% 80% 88% 84% 100% 96% 92% 100% 88% 100% 88% 96% 84% 100% 92% 92% 92% 96% 96% 92%
144
Table 6.15: Answers given in group 7.
qc (A,D)
qc (A,H)
qc (A,L)
qc (B,F)
qc (B,J)
qc (C,E)
qc (C,I)
qc (D,E)
qc (D,I)
qc (E,F)
qc (E,J)
qc (F,H)
qc (F,L)
qc (G,K)
qc (H,K)
qc (I,L)
Uc (p, s, Qc )
qe (A,B)
qe (A,F)
qe (A,J)
qe (B,D)
qe (B,H)
qe (B,L)
qe (C,G)
qe (C,K)
qe (D,G)
qe (D,K)
qe (E,H)
qe (E,L)
qe (F,J)
qe (G,I)
qe (H,I)
qe (I,J)
qe (J,L)
Ue (p, s, Qe )
subject
solution c@ c@ c@ c@ c@ c∀ c∃ c∀ c∃ c∃ c∀ c@ c∃ c∃ c∃ c∃ e@ e∃ e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∀ e∃ e∃ e∃ e@ e∃ e∃ e∃
correct 86% 81% 71% 86% 86% 95% 90% 100% 95% 95% 90% 86% 81% 95% 86% 86% 100% 100% 90% 100% 90% 100% 95% 90% 100% 86% 100% 100% 100% 95% 67% 100% 95%
6.5 experimental evaluation
145
Table 6.16: Answers given in group 8.
Ue (p, s, Qe )
Uc (p, s, Qc )
measuring structural process model understandability
qe (D,H)
qc (G,H)
qe (A,G)
qe (C,H)
qe (A,K)
qe (A,C)
qe (C,D)
qc (H,L)
qe (D,L)
qc (A,E)
qc (B,G)
qc (E,G)
qc (G,L)
qe (K,L)
qc (B,K)
qc (E,K)
qe (C,L)
qc (D,F)
qc (B,C)
qc (C,F)
qe (F,G)
qe (B,E)
qe (F,K)
qe (H,J)
qc (A,I)
qc (D,J)
qe (G,J)
qe (I,K)
qc (J,K)
qc (C,J)
qe (B,I)
qe (E,I)
qc (F,I)
subject
solution c@ c@ c@ c@ c@ c∃ c∀ c∃ c∀ c∃ c∃ c@ c@ c∃ c∃ c@ e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃
s144 c@ c∃ c@ c@ c@ c∃ c∃ c∃ c∃ c∃ c∃ c@ c@ c∃ c∃ c@ 0.81 e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∃ e∃ e∃ e∃ 0.94
s145 c@ c@ c@ c@ c@ c∃ c∀ c∃ c∀ c∃ c∀ c@ c@ c∃ c∃ c@ 0.94 e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ 1.00
s146 c@ c@ c@ c@ c@ c∃ c∀ c∃ c∀ c∃ c∃ c@ c@ c∃ c∃ c@ 1.00 e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ 1.00
s147 c@ c@ c∀ c@ c@ c∃ c∃ c∃ c∃ c∃ c∃ c@ c@ c∃ c∃ c@ 0.81 e@ e@ e@ e@ e@ e∀ e∃ e∃ e∃ e∃ e∃ e@ e∃ e∃ e∃ e∃ e@ 0.59
s148 c@ c@ c@ c@ c@ c∃ c∀ c∃ c∀ c∃ c∃ c@ c@ c∃ c∃ c@ 1.00 e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ 1.00
s149 c@ c@ c@ c@ c@ c∃ c∀ c∃ c∀ c∃ c∃ c@ c@ c∃ c∃ c@ 1.00 e∀ e∃ e∀ e∃ e∃ e∀ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ 0.88
s150 c@ c@ c@ c@ c@ c∃ c∀ c∃ c∀ c∃ c∃ c@ c@ c∃ c∃ c@ 1.00 e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ 1.00
s151 c@ c@ c@ c@ c@ c∀ c∀ c∀ c∀ c∀ c∃ c@ c∀ c∃ c∃ c@ 0.75 e∃ e∃ e∃ e∀ e∃ e∀ e∃ e∃ e∃ e∃ e∃ e∀ e∀ e∃ e∃ e∃ e∃ 0.82
s152 c@ c@ c@ c@ c@ c∃ c∀ c∃ c∀ c∃ c∃ c@ c@ c∃ c∃ c@ 1.00 e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ 1.00
s153 c@ c@ c@ c@ c@ c∃ c∀ c∃ c∀ c∃ c∃ c@ c@ c∃ c@ c@ 0.94 e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ 1.00
s154 c@ c@ c@ c@ c@ c∃ c∀ c∃ c∀ c∃ c∃ c@ c@ c∃ c∃ c@ 1.00 e∃ e∃ e∀ e∃ e∃ e∀ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ 0.94
s155 c@ c@ c@ c@ c@ c∃ c∀ c∃ c∀ c∃ c∃ c@ c@ c∃ c∃ c@ 1.00 e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ 1.00
s156 c@ c@ c@ c@ c@ c∀ c∀ c∀ c∀ c∀ c∃ c@ c@ c∃ c∃ c@ 0.81 e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ 1.00
s157 c@ c@ c@ c@ c@ c∃ c∀ c∃ c∃ c∃ c∀ c@ c@ c∃ c∃ c@ 0.88 e∃ e∃ e∃ e@ e∃ e∀ e∃ e∃ e∃ e∃ e∃ e@ e∃ e∃ e∃ e∃ e@ 0.82
s158 c@ c@ c∃ c@ c@ c∃ c∀ c∃ c∀ c∃ c∃ c@ c∃ c∃ c∃ c@ 0.88 e@ e∃ e@ e∃ e∃ e∀ e∃ e∃ e∃ e∃ e∃ e@ e∃ e∃ e∃ e∃ e@ 0.76
s159 c∃ c∃ c@ c@ c∃ c∃ c∃ c∃ c∃ c@ c∃ c@ c@ c∃ c∃ c∀ 0.56 - e∃ e∀ e@ e∀ e∀ e∃ e∃ e@ e@ e∀ e∃ e@ e∃ e∃ e∀ e@ 0.41
s160 c@ c@ c@ c@ c@ c∃ c∀ c∃ c∀ c∃ c∀ c@ c@ c∃ c∃ c@ 0.94 e∃ e∃ - e∃ e∃ e∀ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ 0.94
s161 c@ c@ c@ c@ c@ c∃ c∀ c∃ c∀ c∃ c∀ c@ c@ c∃ c∃ c@ 0.94 e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ 1.00
s162 c@ c@ c@ c@ c@ c∃ c∀ c∃ c∀ c@ c@ c@ c@ c@ c@ c@ 0.75 e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ 1.00
s163 c@ c@ c@ c@ c@ c∃ c∀ c∃ c∀ c∃ c∀ c@ c@ c∃ c∃ c@ 0.94 e∃ e∃ e∃ e∃ e∃ e∀ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ e∃ 1.00
correct 95% 90% 90% 100% 95% 90% 85% 90% 80% 80% 70% 100% 90% 95% 90% 95% 80% 95% 70% 80% 90% 100% 100% 100% 95% 95% 95% 75% 90% 100% 100% 95% 80%
146
6.5 experimental evaluation 147
subject qr (A) qr (B) qr (C) qr (D) qr (E) qr (F) qr (G) qr (H) qr (I) qr (J) qr (K) qr (L) Ur (p, s)
solution r? r? r=1 r=1 r=1 r? r? r? r? r+ r+ r∗
s164 r? r? r=1 r=1 r=1 r? r? r? r? r+ r+ r∗ 1.00
s165 r? r? r=1 r=1 r=1 r? r? r? r? r+ r+ r∗ 1.00
s166 r? r? r=1 r=1 r=1 r? r? r? r? r+ r+ r∗ 1.00
s167 r? r? r=1 r=1 r=1 r? r? r? r? r+ r+ r∗ 1.00
s168 r∗ r? r=1 r=1 r=1 r? r∗ r? r? r+ r+ r∗ 0.83
s169 r? r? r=1 r=1 r=1 r? r? r? r? r+ r+ r∗ 1.00
s170 r? r? r=1 r=1 r=1 r? r? r? r? r+ r∗ r∗ 0.92
s171 r? r? r=1 r=1 r=1 r? r? r? r? r+ r+ r∗ 1.00
s172 r? r? r=1 r=1 r=1 r? r? r? r? r+ r+ r∗ 1.00
s173 r? r? r=1 r=1 r=1 r? r? r? r? r+ r+ r∗ 1.00
s174 r? r? r+ r∗ r∗ r=1 r=1 r? r=1 r+ r? r∗ 0.42
s175 r? r? r=1 r=1 r=1 r? r? r? r? r+ r+ r∗ 1.00
s176 r? r? r=1 r=1 r=1 r? r? r? r? r+ r+ r∗ 1.00
s177 r? r? r=1 r=1 r=1 r? r? r? r? r+ r+ r∗ 1.00
s178 r? r? r=1 r=1 r=1 r? r? r? r? r+ r+ r∗ 1.00
correct 93% 100% 93% 93% 93% 93% 87% 100% 93% 100% 87% 100%
Additionally, also the 95% confidence intervals for the estimated (partial) struc-
tural process model understandability values of the 13 data sets were computed.
For o1, o2 and o4, the method for estimating confidence intervals for means of
normal distributions (see experiment 1) was used. For the other ten data sets, the
bootstrap approach which does not require normally distributed data was applied.
The lower and upper confidence interval bounds are also listed in Table 6.19.
The estimated (partial) structural process model understandability values and
the 95% confidence intervals for the 13 data sets are also depicted graphically in
Figure 6.12.
Finally, it was analyzed whether the distributions of the personal partial struc-
tural process understandability values of the four data sets of each of the aspects
concurrency, exclusiveness and order are the same. For that purpose, a Kruskal-
Wallis rank sum test was conducted for each of these three aspects. The null-
hypothesis (same distribution) could not be rejected on the α = 0.05 level for
all of them. So, the difficulty of the four subsets of questions for each of these
aspects seems to be quite equivalent.
Table 6.18: Correct given answers per question for the four aspects (Part 1 of 2).
two different groups of subjects. In the next step, Spearman’s rank correlation
coefficient between the personal partial structural process model understandabil-
ity values from the two halves was computed. This was repeated 5,000 times for
each aspect.
The corresponding empirical cumulative distribution functions are depicted in
Figure 6.13. The medians were 0.714 (concurrency), 0.818 (exclusiness) and 0.933
(order). So, the approach seems to be legitimate.
The resulting virtual subjects are listed in Table 6.20.
In the remainder of this section, the resulting virtual personal structural process
model understandability values are denoted as U∗a (p, s), the virtual estimated
structural process model understandability values as U b ∗ (p, S) and the virtual
a
estimated partial structural process model understandability as U b ∗ (p, S, Qa )
a
(a ∈ {c, e, o, r}).
Table 6.18: Correct given answers per question for the four aspects (Part 2 of 2).
(virtual) subjects for the four aspects concurrency, exclusiveness, order and repetition
are depicted in Figure 6.14a.
In order to test the hypothesis that the personal structural process model un-
derstandability values are normally distributed for each aspect (Hypothesis 6.1),
a Shapiro-Wilk test for each of the four data sets was done. For concurrency, exclu-
siveness and repetition, the null-hypothesis that the data is normally distributed
had to be rejected (concurrency: p = 0.023; all others: p 0.05). Only for order,
this null-hypothesis could not be rejected on the α = 0.05 level.
Table 6.19: Estimated (partial) structural process model understandability values, standard deviations and 95% confidence intervals for the four
aspects and 13 data sets.
o1 o2 o3 o4 c5 c6 c7 c8 e5 e6 e7 e8 r9
b a (p, S, Qa ) or U
U b r (p, S) 0.579 0.609 0.583 0.579 0.816 0.835 0.881 0.898 0.942 0.941 0.946 0.905 0.945
standard deviation 0.163 0.186 0.169 0.154 0.147 0.167 0.139 0.118 0.098 0.093 0.081 0.159 0.153
lower conf. interval bound 0.498 0.521 0.513 0.506 0.747 0.769 0.818 0.844 0.893 0.900 0.907 0.830 0.861
upper conf. interval bound 0.661 0.697 0.653 0.651 0.878 0.898 0.932 0.948 0.983 0.972 0.978 0.965 1.000
150
6.5 experimental evaluation 151
30
20
25
15
20
frequency
frequency
15
10
10
5
5
0
0
0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0
concurrency exclusiveness
5
4
10
frequency
frequency
3
2
5
1
0
0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0
order repetition
Figure 6.8: Histograms for rates of correct answers for the four aspects.
For testing the hypothesis that the structural process model understandability
values for the four aspects are different (Hypthesis 6.2), Wilcoxon rank-sum tests
for independent values were conducted. This test does not require normally dis-
tributed data. Only for the combination exclusiveness-repetition, the null-hypothesis
(data belongs to same distribution) could not be rejected on the α = 0.05 level
(p = 0.110).
1.0
1.0
● ●● ● ●●● ●●● ● ●
● ●
● ●● ● ●● ●
● ● ● ●
● ● ● ●
● ● ● ● ● ●
● ●
● ●
● ● ● ●
●● ● ● ●
●
● ● ● ●
0.8 ● ●
0.8
● ● ● ● ●
● ●
● ● ● ● ●
● ● ●
● ● ● ●
● ● ● ●● ● ●
● ● ● ● ● ● ●
● ● ● ● ● ●● ●
● ● ● ● ●
● ● ● ●
0.6
0.6
● ● ●● ● ● ● ●
exclusiveness
orderRevers
● ●
● ●
● ● ● ● ● ● ● ●
●
● ● ●● ●
● ●
● ●
● ●
● ●
0.4
0.4
● ● ●
●
● ●
● ● ● ●
●
●
●
●
● ● ●
0.2
0.2
●
●
●
● ● ●
●
0.0
0.0
0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0
concurrency order
(a) Rates of correct answers concur- (b) Rates of correct answers order/order in re-
rency/exclusiveness. verse order.
Figure 6.9: Scatter plots with rates of correct answers for experiment 2.
1.0
● ●
● ●
●
●
●
●
0.8
●
●
● ●● ●
●
● ● ● ●●
●
●● ●●
● ● ●
● ●● ●
0.6
● ● ●
●●● ●● ● ● ●
●● ● ● ● ● ●
● ● ● ● ● ●
U_a
● ●
● ●
● ● ●● ●●
● ● ●
0.4
● ●
● ● ●
●
●
● ●
0.2
0.0
o1 o2 o3 o4 r9
aspect/group
Figure 6.10: Personal (partial) structural process model understandability values for the
aspects order and repetition.
The values depending on the coverage rate are depicted in Figure 6.15. The
dashed horizontal lines are the lower and upper 95% confidence interval bounds
for the (virtual) estimated structural process model understandability values of
the four aspects.
6.5 experimental evaluation 153
1.0
● ● ●
●●
● ●●
●●●● ●
●
●●● ●
●● ●●
●●●
●●
● ●● ● ● ●●●●
●
●
●●● ● ● ● ● ●
●
●●● ●● ● ● ● ●●●
0.8
●● ● ●●
●
● ●
●
●●
● ●
0.6
● ●
● ●●
●
●
● ●
U_a
0.4
0.2
0.0
c5 c6 c7 c8 e5 e6 e7 e8
aspect/group
Figure 6.11: Personal partial structural process model understandability values for the
aspects concurrency and exclusiveness.
1.0
● ● ● ●
● ●
●
● ●
0.8
●
0.6
● ● ●
U_a
0.4
0.2
0.0
o1 o2 o3 o4 c5 c6 c7 c8 e5 e6 e7 e8 r9
aspect/group
Figure 6.12: Estimated (partial) structural process model understandability values and
95% confidence intervals for the 13 data sets.
1 The highest number of possible subsets exists for the aspect order and coverage rate 0.5. Here,
132 38 different subsets exist.
66 ≈ 3.8 × 10
154 measuring structural process model understandability
1.0
1.0
● ●
0.8
0.8
0.6
0.6
F(correlation)
F(correlation)
●
● ●
0.4
0.4
●
● ●
0.2
0.2
●
0.0
0.0
●
0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0
●
0.8
●
●
●
●
●
0.6
F(correlation)
●
●
●
●
●
0.4
●
●
0.2
●●
●
●
●
●●
●●
●
●●●
●●
● ● ●●
●●
●● ● ● ●● ●
●●
0.0
● ● ●
● ●
●
Figure 6.13: Empirical cumulative distribution functions for Spearman’s rank correlation
coefficients between two halves of questions of experiment 1.
process model understandability values and the rate of values lower and higher
than the confidence interval bounds of the four aspects are listed for different
coverage rates.
Also for these tables, a probabilistic algorithm had to be used: For each analyzed
coverage rate, 100,000 subsets of questions were randomly selected. Exact values
could only be computed for very small and very large coverage rates as well as
for the aspect repetition.
The different values for the mean in Table 6.23a, 6.23c and 6.23d compared to
Table 6.21 are caused by rounding errors.
Table 6.22 and Figure 6.15 support the hypothesis—the aspect repetition having
the weakest effect: For the same coverage rate, many different (virtual) estimated
6.5 experimental evaluation 155
virtual subject
partial structural process model understandability values exist. The reason for this
is the different difficulty of the single questions as already shown in Table 6.18
and Figure 6.8. The smaller the coverage rate, the higher the standard deviation
and the number of values outside the confidence interval.
For the process model used in experiment 2, a coverage rate of 0.25 produces
less than 1% lower or upper outliers for all four aspects.
Validity Evaluation
Also for the second experiment, the necessary validity evaluation (see paragraph
Validity Evaluation in Subsection B.3.3) is carried out.
156 measuring structural process model understandability
virtual subject
• Maturation: It was not measured how much time the subjects needed for
answering the online questionnaire. As the number of asked questions did
not differ that much between experiment 1 (20 or 25 questions per subject)
and experiment 2 (33 questions per subject), it is believed that factors as,
for example, fatigue, boredom or hunger had also no big influence for
experiment 2.
6.5 experimental evaluation 157
virtual subject
Table 6.21: (Virtual) estimated structural process model understandability values, stan-
dard deviations and 95% confidence intervals for the four aspects.
1.0
1.0
●● ●
●
●●●● ● ●
●●
●●
●● ●
●●
●
● ●●● ● ●
● ● ● ●● ● ●
● ●
● ●
● ● ● ●
● ●
0.8
0.8
● ●
● ●
● ●
●
●
● ●
● ●
0.6
0.6
● ● ● ●
●●
●
●
U_a
U_a
●
●
●
●
0.4
0.4
●
●
●
0.2
0.2
0.0
0.0
c e o r c e o r
aspect aspect
(a) (Virtual) personal structural process model (b) (Virtual) estimated structural process
understandability values. model understandability values and 95%
confidence intervals.
1.0
1.0
●●● ● ●●● ●●● ●●●●●●●●● ●
●
● ●
●
●● ●● ● ● ● ●
● ●
● ● ●
● ●
●●●●●●●●
●
●●●●●●●● ●●●●●●
●●
●
●●●●
●
●●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●
● ●
● ●
●● ● ●● ●●● ● ●●●●●
● ●
● ● ● ● ● ●
● ●
● ●
● ●
● ●
● ●
● ●
● ● ●●
●●●●●●●● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ●
●●
●● ●
●●● ●●● ● ● ● ●
● ●
● ●
●●● ●●● ●● ●●
● ●●
●●
●
●
● ●
● ●
●
●
●
●
●
●
● ●
●
● ●
● ●
●
● ●
● ●
●
●
●
● ●
●
●
●
●
● ●
● ●
●
●●
● ●●
● ●
●
●●
●
●
●●
●●
●
●●●
●●●
●●●●
●● ● ● ● ● ●
● ●● ● ● ● ●
● ● ● ● ● ● ● ●
● ● ● ● ● ●●●●●● ●●●
●
●● ●● ●●●●●●●●●●●
● ●
●
●
●
●
●
●
● ● ● ●
● ●● ●
●●● ●●●●●●
● ●
● ●
●
● ●
●
●
● ●
●
●
●
●
●
●
● ●
●
●
●
●
● ●
● ●
●
●
●
●
●
●
●
● ●
●
●
●
● ●●● ●●●●● ●
● ●
● ●
● ●
● ●
● ● ● ● ●●● ● ● ●
● ●
● ●
● ●
● ●
● ●
● ●
● ● ●
●
● ●
● ● ●
●●● ●●● ●●●●
● ●●
● ●●
●
●
●
●
●
●
● ●
● ●
●
●
●
●
●
●
●
● ●
●
● ●
●
●
●
● ●
● ●
● ● ●●● ●●● ●●●
●●●
●●●
●●●
●
●
●
●
●
●
●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●
●
● ● ● ● ● ●
● ●
● ●
● ● ●
● ● ● ●
● ● ● ● ● ● ●
●● ● ●
●●●●● ● ●●
●
●●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
● ●●
●
●
●●
● ●●
● ●●
● ●
●●●● ●● ● ●
●●●●●● ●
●● ●
● ●
●
●
●
●
●
●
●
●
● ●
●
●
●
●
●
●
●● ● ●
● ● ●
● ● ●
●
● ● ● ●
● ●
●
● ●
●
● ●
● ●
●
● ●
● ●
● ●
● ●
● ●
● ●
● ● ●
● ●
● ●
● ●
●●●●
● ●●
●●
● ●● ● ●
● ● ●●
● ●
●
● ●
● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●●●●●● ● ● ● ● ●
●●●●● ●●●●●
●●●
● ●
●
●
●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●●
● ●
●●●●●●● ● ● ●
●●●● ●●●●●
●●●
● ●
●
●
●●● ●●● ●●●
●●● ●
●● ●
● ● ●
●
● ● ●
● ●
● ● ●
● ●
● ●
● ● ● ● ● ●●● ●●● ●●●
●●● ●
●● ●
● ●
●●● ●
● ●
● ●
● ● ●
● ●
● ● ●
● ●
● ●
● ●
● ●●● ●
●
●● ●● ● ● ● ● ● ● ● ● ● ● ● ●● ●● ●● ● ●
0.8
0.8
●
●●●●● ● ●
● ●
● ●
● ● ●
● ●
● ●
● ●
● ● ●
●●●●● ● ●
● ●
●
● ●● ●●●●● ●
● ●
● ●
● ●
● ●
● ● ● ●● ●●●●
●●●● ● ● ●
●
● ● ● ●
● ●●●● ●●●
●●● ●●● ●
●●●
●●● ●
● ● ● ●●● ●●● ● ●
● ●●●
●●
●
●●
● ● ●
● ●
●
●
●
●
● ● ●● ●●
● ●●
●● ●●● ●
●●●
●●
●●●●
● ●
●
●
●
●
●
●
●● ●●●
●●●● ● ●●
● ● ●●
●●● ●●● ●●●●●●●●● ●●● ●
● ●● ●
●●●●● ● ●
●● ●
● ● ●●
●● ●● ●● ●● ●●
●●● ●●●
●●●
●●
●
●
●
●●● ●●● ● ● ●
●● ● ●
●●●●●
●● ● ●
●
●
●●● ●
●● ●●● ●
0.6
0.6
● ●●
●● ●●
●●
●●● ●
● ●
●
U_e
U_c
●
●
0.4
0.4
●
0.2
0.2
0.0
0.0
0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0
1.0
● ● ● ● ●
● ● ● ●
●● ● ● ● ● ● ●
● ● ● ● ● ● ●
● ●
●●●●●● ● ● ● ● ● ● ●
● ●
● ●
● ● ● ● ●
●●● ● ● ● ● ● ● ● ● ● ●
●●●●●
●● ● ● ● ●
●
●
●
●
● ●
●●● ● ● ●
●●●●●●
●●
● ●
●●●●
●●
●●●●●
●●
●
● ●
●●
●●●
●
●●
●●●●●●
●●●
●● ●
●●●● ● ●
0.8
0.8
●● ●
●
●
●●
●●●● ●
●
● ●
●
●●
●●●●●●
●●
●● ●
●
●
●
●
●
● ●
●●● ●●
●
●●●
●●●
●
●
●
●
●
●
●
●
●
●
● ●
●
● ●
● ●
● ●
●●
●●●●●●●
●
●
●
●
●
●
●
●
●
●
●
● ●
●●●
●●● ●
● ● ●
● ●
● ●
●
●
●●
●●●●● ●
●
● ● ●
●
● ●
● ●
● ●
● ●
● ●
● ●
● ●
● ●
● ●
● ●
●
●●●● ●
● ●
● ●
● ●
● ●
● ●
● ● ●
●
●●●●●●● ●
●
● ● ●
● ●
● ●
● ●
● ●
● ●
● ●
●●● ●
● ●
● ●
● ● ●
● ●
● ●
● ●
●
● ●
● ● ●
●● ●●
●●●●● ●
●
●
● ●
● ●
●
● ●
●
●
●
●
● ●
● ●
●
● ●
● ●
●
●
●
● ●
● ●
●
●●● ●
● ●
● ●
●
● ●
● ●
● ●
● ●
● ●
●
● ●
● ●
● ●
● ●
● ●
● ●
●
●●●●
● ●
●
● ●
● ●
● ●
● ●
● ●
● ●
●
● ●
● ●
● ●
● ●
● ●
● ●
● ●
● ●
●
●
●●
●●●●●● ●
● ● ●
● ●
● ●
● ●
● ● ●
● ●
● ●
● ●
● ●
● ●
● ●
● ● ●
●
0.6
0.6
●●● ●
●
● ●
● ●
● ●
● ●
● ●
● ●
● ●
● ●
● ●
● ●
●
● ●
● ●
● ●
● ●
● ● ●
●
●●● ●
● ●
● ●
● ●
● ●
● ●
● ●
● ●
● ●
● ●
● ● ●
● ●
● ●
●
● ●
● ●
● ●
● ●
●●●
●●●
●●●●●● ●
● ●
● ●
● ●
● ●
● ●
● ●
● ●
● ●
● ●
● ●
● ●
●
● ●
● ●
● ●
● ●
● ●
● ●
●●
●
●●●●●
●●
●● ●
● ● ●
● ● ●
● ●
● ●
● ●
● ●
● ●
● ● ●
● ●
● ●
● ●
● ●
● ●
● ●
●●
●●● ●●
●●●
●●●●●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●
● ●
●
●
●
●
●
●
●
●
● ●
● ●
●
●
●
●
● ●
●
●
●
●
● ●
●
●
● ●●●
●●●●
●
●●
●● ●
●
● ●
● ●
● ●
● ●
● ●
●
● ●
● ●
●
● ●
● ●
●
● ●
●
● ●
●
● ●
● ●
● ●
● ●
●
●● ● ● ●
● ● ●
● ● ● ● ● ● ●
● ● ●
● ● ●
●●● ●●
●
●●●
●●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●
●
●
● ● ● ● ● ● ● ●
U_o
● ● ● ●
U_r
● ● ● ● ● ● ● ● ●
●●
●●●●●●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●●●
●● ●
● ● ●
● ● ●
● ●
● ● ●
● ●
●
●●
●●●●● ●
●
● ●
● ●
●
● ●
●
● ●
●
● ●
● ●
● ●
●
● ● ●
● ●
● ● ● ● ●
●●●●
● ●
●
● ●
● ●
● ●
● ●
●
●
●
●
●●●●●●●
●● ●
● ● ●
● ●
● ●
●● ●
● ● ●
●
● ● ●
●●●●●●
●●● ●
●
●
●
●
● ●
● ●
● ●
● ● ● ●
● ●
0.4
0.4
●●●● ● ●
●●
●●●●●●●
●
●
●
●
●
●
●
● ●
●●●
● ●
●
● ●
●
●● ● ● ●
●●●●●●
●●
●●●●
●
●
●
●
●
●
●
●
●
● ●
●
● ●
●
●●
●●●●●●● ●
●
●
●●●● ●
●
●●●●●●
●
●●●
●
●●●●
●●
●●●●●●●
●
●●●
●●●
●
●●
●●●●●
●●●●
●
●
●●
●●●●●●
●
0.2
0.2
●
● ●●●
●●●●●●
●●●●
●
●●●●●●
●
●●
●●
●●●●●
●●●●
●●●●●●
●
●●●
●
●
0.0
0.0
0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0
Figure 6.15: (Virtual) estimated partial structural process model understandability values
of the four aspects depending on coverage rate.
6.6 conclusion
Table 6.22: Data on (virtual) estimated partial structural process model understandability
values for the four aspects (Part 1 of 4).
Table 6.22: Data on (virtual) estimated partial structural process model understandability
values for the four aspects (Part 2 of 4).
process model with five tasks, 18 subjects; experiment 2: process model with 12
tasks, 178 subjects) were conducted in order to examine these hypotheses. The
results of these experiments are quite consistent.
They support the hypothesis that different aspects of structural process model
understandability are of varying difficulty (only exclusiveness and repetition are
162 measuring structural process model understandability
Table 6.22: Data on (virtual) estimated partial structural process model understandability
values for the four aspects (Part 3 of 4).
quite similar in case of the second and larger process model). Thus, all different
aspects have to be measured in order to get a feeling of the “overall structural
process model understandability”.
6.6 conclusion 163
Table 6.22: Data on (virtual) estimated partial structural process model understandability
values for the four aspects (Part 4 of 4).
Furthermore, the hypothesis that asking only a small part of the set of possible
questions for one aspect can cause values to differ substantially from the real
value was strongly confirmed. Consequently, the coverage rate of asked questions
should not be too small. With respect to the larger process model of experiment 2,
a coverage rate of 0.25 resulted in less than 1% outliers (higher or lower than 95%
confidence interval) for all four aspects. Finally, the asked questions should be
selected randomly in order to minimize the risk of choosing particularly easy or
difficult questions.
In both experiments, only the aspect order was normally distributed. This aspect
also had the lowest values of all examined aspects—what is not directly intuitive.
Arguably, concurrency and exclusiveness are more complicated matters than order.
This fact should be further examined in future work.
Another future issue is the selection of suitable coverage rates which minimize
the measuring effort and the differences from the real structural process model
understandability value. It should be investigated whether the ideal coverage rate
is indicated relative or absolute to the process model size and whether it depends
on other (structural) process model properties.
Furthermore, it should also be examined whether other aspects of structural
process model understandability exist.
EFFECTS OF PROCESS MODEL GRANULARITY
7
7.1 introduction
ha
has s
costs
PROCESS MODEL
duration
number of errors
internal external
possible reasons complexity possible implications changeability
attributes attributes
flexibility
pr
oc el
e od res understandability
m ss m m
et s s asu
ric od dependency e e
s e oc m
l pr lity
a
process model process model qu
metrics quality measures
values predictability values
Figure 7.1: Chapter visually located within the measurement approach of Subsec-
tion 3.4.2.
tion 7.3. The conducted experiment and its results are shown in Section 7.4. The
chapter closes with a conclusion (Section 7.5).
In this section, the process model granularity heuristic, which was introduced
by Vanderfeesten et al., is presented. In Subsection 7.2.1, the underlying process
model granularity metric is first explained. The actual heuristic is presented in
Subsection 7.2.2.
In this chapter, the definitions from [168, pp. 426–429] are used with some
modifications:
• In [168], all definitions are based on so-called operations structures (see
Definition 1 in [168, p. 426]). This process modeling language is almost
equivalent to the PDMs of Definition 2.11 presented in Subsection 2.3.2. The
operations of a operations structure correspond to the production rules of a
PDM. As PDMs are more frequently used, the original definitions of [168]
are adapted to PDMs in this chapter.
• In [168], references to resource classes or roles which are able to execute the
operations and activities are given. As they are not relevant for the analysis
here, they are omitted in this chapter.
Based on a PDM, the contained data elements and production rules are parti-
tioned into different activities.
Definition 7.1 (Activitiy) An activity T ⊆ F based on a PDM (D, C, pre, constr, cst,
flow) is a set of production rules.1
1 Remember: F is the set of production rules of a PDM (D, C, pre, constr, cst, flow) (see Defini-
tion 2.11).
7.2 process model granularity heuristic 167
S
As a shorthand, the notation T̂ := (p,cs)∈T ({p} ∪ cs) for the data elements
processed in an activity T is introduced.
The different activities can be combined to a process model which processes
and computes the data elements of the PDM in a valid sequence. For details on
how to specify the control flow or how to check the correctness and soundness
of the process model, the reader is referred to [127]. For the purpose here, the
following definition is sufficient.
Definition 7.2 (Process model) A process model2 based on a PDM (D, C, pre, constr,
cst, flow) is a set S of activities on this PDM.
Based on these notations, metrics for process model cohesion and coupling can be
defined.
Process model cohesion measures “to what extent operations3 ‘belong’ to each
other within one activity” [168, pp. 420–421]. It consists of two components.
The first one, activity relation cohesion, quantifies how much the production rules
of an activity are related. For that purpose, it measures the average overlap of
production rules. Two production rules overlap if they share input or output data
elements.
Definition 7.3 (Activity relation cohesion) For an activity T based on a PDM (D, C,
pre, constr, cst, flow), the activity relation cohesion λ(T ) is defined as
|{((p1 ,cs1 ),(p2 ,cs2 ))∈T ×T |(({p1 }∪cs1 )∩({p2 }∪cs2 ))6=∅∧p1 6=p2 }|
|T |·(|T |−1) for |T | > 1
λ(T ) := .
0 for |T | 6 1
(7.1)
The total cohesion of an activity is simply the product of its relation and
information cohesion.
Definition 7.5 (Activity cohesion) For an activity T based on a PDM (D, C, pre,
constr, cst, flow), the activity cohesion c(T ) is defined as
Definition 7.6 (Process model cohesion) For a process model with a set S of activi-
ties based on a PDM (D, C, pre, constr, cst, flow), the process model cohesion ch4 is
defined as
P
c(T )
ch := T ∈S . (7.4)
|S|
Process model coupling quantifies how strong the activities of a process model
are connected to each other. Two activities are connected if they share at least one
data element. The coupling metric measures the fraction of connected activity
pairs.
Definition 7.7 (Process model coupling) For a process model with a set S of activi-
ties based on a PDM (D, C, pre, constr, cst, flow), the process model coupling cp5 is
defined as
|{(T1 ,T2 )∈S×S|T1 6=T2 ∧(T̂1 ∩T̂2 )6=∅}|
|S|·(|S|−1) for |S| > 1
cp := . (7.5)
0 for |S| 6 1
Definition 7.8 (Process model coupling/cohesion ratio) For a process model with
a set S of activities based on a PDM (D, C, pre, constr, cst, flow), the process model
coupling/cohesion ratio ρ6 is defined as
cp
ρ := . (7.6)
ch
Vanderfeesten et al. state that the proposed metrics and the heuristic are inspired
by software engineering “where an old design aphorism is to strive for strong
cohesion, and loose coupling” [168, p. 421].
Consequently, the statement of the heuristic is that a process model with a
smaller value of the process model granularity metric (process model coupling/
cohesion ratio) of Definition 7.8 is to be preferred over another one with a larger
value. Yet, it does not describe how different alternative process models can be
found. [168, p. 429]
Vanderfeesten et al. establish the following two hypotheses about the implica-
tions of their heuristic [168, pp. 425–426]:
Hypothesis 7.1 The smaller the value of the process model granularity metric of a
process model, the smaller the probability of run-time mistakes.
Hypothesis 7.2 The smaller the value of the process model granularity metric of a
process model, the larger the understandability and—consequently—the maintainability.
For testing Hypothesis 7.2, experiments in which subjects have to change ex-
isting process models (e. g., finding and correcting errors or realizing special
modification tasks) would have to be conducted. These experiments would be
very time-consuming for the participants and therefore making it hard to find
volunteers among students of a university. Consequently, only Hypothesis 7.1
(error probability) is looked at in this section.
For an appropriate experimentation system, three main requirements were
identified:
In Subsection 7.4.1, the experiment design is explained. The results are pre-
sented in Subsection 7.4.2. In Subsection 7.4.3, an alternative error probability
model which better explains the results of the experiment is proposed. The section
closes with the validity evaluation of the experiment (Subsection 7.4.4).
object For this experiment, the PDM depicted in Figure 7.3, which is presented
as an example in [129, 168], was applied. It was used in the abstract fashion
described in Section 7.3. So, only the structural properties—and consequently
the process metric values—remained unchanged. The used production rules are
shown in Figure 7.4.
Based on the PDM, the three different process model alternatives (Figure 7.5)
already proposed in [129, 168] were used. The respective partition into activities
is shown in Figure 7.6.
value(27) == false
42
value(27) == true
value(31) == false
value(34) == false
39 40 41
value(31) == true value(34) == true
36 35 34 38 37
31 32 33
30 27
23 21 22
28 29 19 20
24 25 26
18
12 13
data elements15 . If they were correctly executed, the first and last instances of the
set were routed directly from activity C to G at the XOR split—the others had to
take the branch with all the other activities.
15 Basic data elements are data elements whose values are not computed by any production rule.
Instead, their values have to be given for each process instance before the execution.
7.4 experimental evaluation 173
Figure 7.4: Production rules used in the experiment. vi (18) stands for the integer variable
representing data element 18, vb (27) for the boolean variable representing
data element 27.
For each process model alternative, several teams were used, which each
processed the same set of process instances which was mentioned above. Each
team got exactly as many subjects as there are activities in its process model
alternative. As the subjects executing activity AE in alternative 3 have much
more work than other subjects, two types of teams were used for this alternative
to analyze the effect of this possible bottleneck: The first got the “normal” six
subjects (number of activities in alternative 3)—the second got seven subjects
(two for the resource role of activity AE).
The web-based experimentation system presented in Section 7.3 was used for
the experiment. The subjects accessed the system via web browser from different
desktop computers (one computer per subject) which were located in a single
computer laboratory. As they did not know their other team members, they could
only “communicate”16 via the experimentation system.
16 As explained in Section 7.3, the subjects could only exchange variable values via the web-based
experimentation system. Communication via text messsages was impossible.
174 effects of process model granularity
F
value(27) == true
value(27) == false D
value(27) == true A2
Start C XOR B AND AND AND A3 A4 AND XOR G End
A1
value(27) == false
D
F
value(27) == true
value(27) == false D
Figure 7.5: Three different process model alternatives used in experiment [168, pp. 423–
424].
Table 7.1: Metric values for the three process model alternatives.
cp ch ρ
alternative 1 0.714 0.183 3.9
alternative 2 0.611 0.105 5.8
alternative 3 0.8 0.114 7.0
factor The process model granularity of the three different process model
alternatives acted as factor in the experiment.
F E AE value(31) == false
value(31) == false
value(34) == false
41 39 39
value(34) == true value(31) == true value(31) == true
34 37 38 36 35 36 35
31 18 32 31 18 32
27 22 18
27 22 30 27 22 30
C
27
A4 28 29
value(31) == false
39
23 21 22
value(31) == true
24 25 26
36 35
19 20
31 30 18 32
G
A3 value(27) == false
31 42
value(27) == true
A1 A2
28 29
30 27 22 39 40 41 27
24 25 25 26
28 29
A
B
30
25
D
40
28 29 18
32 33 27 18 24 25 26 12 13
Figure 7.6: Partitioning of the PDM in smaller activities [168, pp. 423–425].
7.4.2 Results
The number of incorrect process instances (over all teams) for the different process
model alternatives are shown in Table 7.3.
First, it was checked whether there is a significant difference between alterna-
tive 3 with six and seven subjects. For that purpose, Pearson’s chi-square test
[117, pp. 643–648] was used. The null-hypothesis that the numbers belong to the
same distribution could not be rejected on the α = 0.05 level. Consequently, both
cases were mixed together for the further analysis (row “sum alternative 3” in
Table 7.3).
Afterwards, the actual hypothesis was looked at. As one can see in Table 7.3,
alternative 1, which should be the best process model alternative according to
Hypothesis 7.1, has the highest ratio of incorrect process instances closely followed
by alternative 3, which should be the worst. Again, a chi-square test was done to
test the alternatives for significant differences. Only for the pair alternative 1 and
2, the null-hypothesis (no difference) could be rejected (p ≈ 0.030). So, the results
of the experiment do not support Hypothesis 7.1.
Next, an analysis on the activity level was conducted. The results of pairwise
chi-square tests are shown in Table 7.4. Looking at Table 7.3, one sees that the error
probabilities of activities A–AE have exactly the opposite order than predicted
by Hypothesis 7.1—even though, there is only a significant difference between
alternatives 1 and 3. This was a motivation to further search for alternative factors
of influence.
In a next step, the possible influence of the number of data elements and
production rules (see Table 7.5) on the error probability of activities (see last row
of Table 7.3) was analyzed and depicted in Figure 7.7. For that purpose, both
Spearman’s rank correlation coefficient (see Section C.2) and Pearson’s product-
moment correlation coefficient (see Section C.1) were computed. For the number
of data elements, one gets 0.95 (Spearman) and 0.78 (Pearson) respectively—
as well as 0.97 (Spearman) and 0.85 (Pearson) respectively for the number of
production rules. So, roughly speaking, larger activities are more error-prone.
These results are interpreted as follows: Hypothesis 7.1, that process model
granularity (a global process model property) influences the error probability
Table 7.3: Error statistics for the different process model alternatives (alternative 3 with six and seven subjects, respectively).
# incorrect
# incorrect
activities C
# incorrect
activities B
# incorrect
activities F
# incorrect
activities D
# incorrect
activities G
# incorrect
activities A
# incorrect
activities E
# incorrect
activities A1
# incorrect
activities A2
# incorrect
activities A3
# incorrect
activities A4
# incorrect
activities AE
# process instances
process instances
A–AE incorrect
with at least one of
alternative 1 29/60 9/60 1/41 3/41 1/41 0/60 5/41 14/41 - - - - - 18/41
48.3% 15.0% 2.4% 7.3% 2.4% 0.0% 12.2% 34.1% 43.9%
alternative 2 14/50 0/50 0/40 2/40 0/40 0/50 - - 0/40 1/40 2/40 10/40 - 13/40
28.0% 0.0% 0.0% 5.0% 0.0% 0.0% 0.0% 2.5% 5.0% 25.0% 32.5%
alternative 3, 6 subjects 26/60 3/60 3/47 9/47 1/47 7/60 - - - - - - 9/47 9/47
43.3% 5.0% 6.4% 19.1% 2.1% 11.7% 19.1% 19.1%
alternative 3, 7 subjects 24/60 0/60 2/48 9/48 1/48 4/60 - - - - - - 15/48 15/48
40.0% 0.0% 4.2% 18.8% 2.1% 6.7% 31.3% 31.3%
sum alternative 3 50/120 3/120 5/95 18/95 2/95 11/120 - - - - - - 24/95 24/95
41.7% 2.5% 5.3% 18.9% 2.1% 9.2% 25.3% 25.3%
sum 12/230 6/176 23/176 3/176 11/230 5/41 14/41 0/40 1/40 2/40 10/40 24/95
5.2% 3.4% 13.1% 1.7% 4.8% 12.2% 34.1% 0.0% 2.5% 5.0% 25.0% 25.3%
7.4 experimental evaluation
177
178 effects of process model granularity
Table 7.4: Results of chi-square tests for error statistics on activity level (α = 0.05). For
cells marked with “+”, the null-hypothesis (no difference) was rejected.
activities A–AE
activity D
activity G
activity C
activity B
activity F
# alternative 1 vs. 2 + - - - - -
# alternative 1 vs. 3 + - - - + +
# alternative 2 vs. 3 - - + - + -
Table 7.5: Number of data elements and production rules per activity.
activity AE
activity A1
activity A2
activity A3
activity A4
activity A
activity D
activity G
activity C
activity B
activity E
activity F
# data elements 6 4 7 5 5 6 9 3 3 6 7 14
# production rules 2 2 4 1 2 3 5 1 1 2 4 8
during process instance execution, might not be true. Instead, activity size seems
to have a big influence on the error probability of an activity. From a subject’s
point of view, the remaining process model is some kind of “black box”. He/she
only sees its own activity with the contained production rules. This fact motivates
the following alternative error probability model.
If one further assumes, for the sake of simplicity, that all error probabilities pi
of the n activities of a process model are equal with value p, one gets
0.35
0.35
● ●
0.30
0.30
0.25
0.25
● ● ● ●
0.20
0.20
error probability
error probability
0.15
0.15
● ●
● ●
0.10
0.10
0.05
0.05
●
● ●
●
● ●
● ●
● ●
● ●
0.00
0.00
● ●
4 6 8 10 12 14 1 2 3 4 5 6 7 8
Comparing the error probabilities PerrA and PerrB of two alternative process
models, one gets the following theorem.
Theorem 7.1 Given are two alternative process models A and B with nA and nB
activities, respectively, and the activity error probability pA and pB , respectively.
Then, process model A is more error-prone than process model B (PerrA > PerrB ) if
nB
1. pA > 1 − (1 − pB ) nA or
ln(1−p )
2. nA > nB · ln(1−p B ) .
A
Regarding 1.)
Regarding 2.)
If one applies Theorem 7.1 on the more special case that one process model
alternative has larger and more error-prone but less activities than the other one,
one gets the following corollary.
Corollary 7.1 Given are two alternative process models A and B. Alternative B has
larger and more error-prone (pA < pB ) but less activities (nA > nB ) than alternative A.
Then, process model A is more error-prone than process model B (PerrA > PerrB ) if
nB
1. 1 − (1 − pB ) nA
< pA < pB or
ln(1−p )
2. nA > nB · ln(1−p B ) .
A
Regarding 1.) The proposition follows from case 1 of Theorem 7.1 together with
the precondition pA < pB .
ln(1 − pB )
nA > nB · (7.9)
ln(1 − pA )
holds.
As
Let us now look at the following example for Corollary 7.1: Alternative B
has larger and more error-prone but less activities than alternative A. So, while
alternative B has nB = n activities with error probability pB = 0.075, alternative A
has nA = 2n activities with error probability pA = 0.05. As one can easily check
using case 2 of Corollary 7.1, alternative B is less error-prone for all values of n.
Generally, one finds many parameters for this error probability model so that
the process model with the larger and more error-prone but less activities has a
smaller error probability than the alternative process model.
These findings on the error probability model are consistent with the inter-
pretation of the results of the experiment made above. Hypothesis 7.1 could be
wrong. Instead of process model granularity, the size (and consequently the error
probability) and the number of activities in a process model could be the main
reasons for different error probabilities of alternative process models.
internal validity Internal validity refers to the fact that the effects observed
in the experiment are not caused by a factor which one has no control of or has
not measured (see Definition B.9).
Looking at the threats to internal validity mentioned in Subsection B.3.3, one
can make the following statements:
• History: During the experiment which lasted less than one hour, no events
occurred which were able to strongly influence the subjects.
• Maturation: The experiment was so short that factors as, for example,
fatigue, boredom or hunger had no big influence.
• Mortality: As all subjects finished the experiment, this threat played no role.
external validity External validity refers to the extent to which the re-
sults of an experiment can be generalized out of the scope of the study (see
Definition B.10).
Looking at the threats to external validity mentioned in Subsection B.3.3, one
can make the following statements:
182 effects of process model granularity
• Population validity: The subjects were students with the necessary math-
ematical knowledge for the used abstract functions—yet had no special
training in the area of BPM. As no domain knowledge was needed and even
knowledge about process modeling was unimportant, it is believed here to
be most likely that other subjects—including professionals in BPM—would
show the same effects. For process models with more realistic activities,
things could be quite different.
• Ecological validity: The process models used in the experiment had only
abstract production rules. So, no special domain knowledge was necessary
for the subjects. It is not clear, whether process models with more realistic
activities—requiring domain knowledge—would show similar results.
7.5 conclusion
In this chapter, a short introduction into the process model granularity heuristic
of Vanderfeesten et al. was given. This heuristic is aimed at giving a possible
solution to the process model granularity problem (finding the proper size of
activities in a process model) during the design phase. According to a hypothesis
postulated by Vanderfesten et al., process models with smaller process model
granularity metric values (high cohesion and low coupling) are less error-prone
during process instance execution.
An experimentation system consisting of a small web-based workflow system
for analyzing the hypothesis of Vanderfeesten et al. was presented. It uses abstract
production rules. Consequently, subjects do not need any domain knowledge.
Furthermore, it makes different experiments comparable.
The results of an experiment involving 165 students using the experimentation
system were reported. They do not support the hypothesis of Vanderfeesten et al.
Instead, an alternative error probability model was suggested which is able to
explain the results. According to this model, the error probability of a process
model depends on the size (number of data elements and number of production
rules) of its activities (larger activities have a higher error probability) and the
number of its activities.
The findings of the conducted experiment and the proposed alternative error
probability model should be further examined in future work. Furthermore, it
should also be tested whether process models with more realistic activities—even
if they require some domain knowledge—show the same results related to the
hypothesis of Vanderfeesten et al.
CONCLUSION AND OUTLOOK
8
Subsection 8.1 provides a conclusion of the results of this thesis. Possible future
work in the area of process measurement is presented in Subsection 8.2.
8.1 conclusion
– On the other hand, the metrics ∆ and CP, which both measure some
sort of ratio between existing arcs and possible arcs, are highly corre-
lated. So, it seems that metric CP, which has a much more complicated
computation rule, has little additional benefit compared to metric ∆.
8.2 outlook
Within the thesis, the following open research questions were identified:
a.1 definitions
holds.
Furthermore, the “measure” shall be “additive” in the following manner: If
one “combines” two objects a1 and a2 (binary operation •), the function µ on A
shall also “preserve” the binary operation • such that ∀a1 , a2 ∈ A
holds.
In order to express these requirements in a formal mathematical way, one needs
the following definitions.
First, the term “relational system” has to be defined [130, p. 51].
For the purpose of measurement, two different relational systems are necessary:
an empirical relational system and a formal relational system.
The empirical relational system [130, p. 51] represents that part of reality which
one wants to measure. In the example above, the empirical relational system is
A = (A, H, •) of type (2; 1) with binary relation H and (binary) operation •.
The formal relational system [180, p. 40] represents the system with which
the “magnitude” of the measured property is expressed. In the example above,
the formal relational system is B = (R, >, +) of type (2; 1). If the set R of real
numbers is part of the formal relational system, the term numerical relational
system is used [130, p. 51].
The function µ which maps an object from an empirical relational system to an
object of a formal relational system (or a numerical value of a numerical relational
system) is called measure [180, p. 40].
A.1 definitions 191
If one wants the measure to “preserve” the properties of the empirical relational
system, the “mapping may not be arbitrary” [180, p. 40]. Finally, this leads to the
definition of a scale [130, pp. 51–52, 54] [180, pp. 40–41].
hold.
Then, the triple (A, B, µ) is a scale.
So, the fundamental problem of measurement is to find a scale for each property
one wants to measure. Especially for non-physical properties, it is hard to find
scales in such a way that it is generally accepted that they “preserve” the original
properties of the empirical relational system.
Consequently, the term “scale” is only rarely used. Instead, one falls back to
the term “measure” (Definition A.3) whose requirements are not as strict as for
“scale”.
Even though the term “measure” would be correct, the term “metric” is (more)
often used in process measurement (e. g., [6, 19, 21–24, 54, 97–100, 129, 165–168]).
Zuse already realized that for the area of software measurement [180, pp. 28–29].
Mathematically spoken, the term “metric” is incorrect in the context of process
measurement as it is used in mathematics to describe a kind of “distance measure”
between vectors of a more-dimensional space (see the following definition [170]).
In order to be consistent with the majority of publications in this area, the term
“metric” is used in this thesis for a measure of an internal (structural) attribute
(see Subsection 3.4.2) of a process model (cf. the metrics in Chapter 4, 5 and 7).
For external attributes (e. g., structural understandability in Chapter 6), the term
“measure” is used.
Definition A.6 (Admissible transformation) Let (A, B, µ) be a scale, A the set un-
derlying A and B the set underlaying B. A function Φ : µ(A) 7→ B is called an
admissible transformation if and only if (A, B, Φ ◦ µ) is also a scale.
Stevens’ five scale types are induced by five corresponding sets of admissible
transformations (see Table A.1). From top to bottom of the table, the requirements
for these transformations become harder. Thereby, the set of admissible transfor-
mations of a scale type is a subset of all sets of admissible transformations of
the scale types which are above in Table A.1. Consequently, scales of a special
scale type also have all scale types “higher” in the table. Furthermore, each scale
type has a set of basic empirical operations which can be used for comparing the
empirical property which is measured.
A nominal scale has the scale type with the weakest requirements. On such a
scale, only the (in)equality of the measured property of two or more objects can be
determined. According to Stevens [148, p. 678], two types of nominal assignment
can be distinguished: (1) numbering each object with a distinct number for
identification (e. g., shirt numbers in team sports) or (2) numbering each object
which is member of a class with the same number (e. g., sex—0 for male and 1
for female). Actually, the first case is a special case of the second one where each
class has exactly one member.
Using an ordinal scale, objects can be ordered according to the size of their
measured property. An example are the rank numbers in sports league tables.
The smaller the number, the better the team played in that season. Yet, there is no
information about the “distances” between the different ranks. So, the question
“Is the ‘distance’ between the teams on rank 1 and 2 as large as between rank 2
and 3?” cannot be answered with this scale type.
An interval scale provides this information about “distances”. Examples are
temperature measured in Celsius or Fahrenheit. Here, differences between scale
Table A.1: Hierarchy of different scale types [148, p. 678] [130, p. 64].
values can be compared. So, one can say, for example, that the temperature
difference between 10 ◦ C and 20 ◦ C is as large as between 20 ◦ C and 30 ◦ C.
A ratio scale has a defined and meaningful zero point. So, also ratios of scale
values can be compared. An example is absolute temperature measured in Kelvin.
Here, the zero point is defined as the lowest physically possible temperature.
For an absolute scale, finally, only the identity function is allowed as admissible
transformation. Counting is an example for this scale type.
These scale types do not only allow to classify scales—they also imply which
statistical operations can legitimately be applied to empirical data [148, p. 677].
In order to discuss this in more detail, the following definition of meaningfulness
[130, p. 59] is necessary.
• median [117, pp. 19–21], quantiles [117, pp. 25–26] and mean [117, pp. 16–17]
as measures of location,
• range [117, p. 22], interquartile range [156, p. 55], median absolute devia-
tion1 , standard deviation [117, p. 22, 35–36] and coefficient of variation [117,
pp. 33–34] as measures of dispersion as well as
While defining generally accepted measures for physical properties (e. g., length,
mass, temperature, etc.) is relatively easy in physics, the same purpose is much
harder for non-physical properties as, for example, intelligence, religiosity, preju-
dice, etc. in psychology and sociology.
In Subsection A.3.1, the general procedure for finding adequate measures is
introduced. Subsection A.3.2 presents necessary properties which these measures
have to fulfill.
This subsection is based on parts of a textbook by Babbie [5, pp. 118–140]. There,
one can also find further details.
The fundamental problem when measuring non-physical properties (e. g., the
magnitude of prejudices of a person) is that these properties themselves do not
exist in reality. Instead, these terms represent mental images in the brains of
humans summarizing certain related observations and experiences which these
persons have made during their lives. The terms are then used to communicate
the corresponding mental constructs.
To illustrate this fact, one can think of the term “prejudice” as an example.
Presumably, everybody has made certain of the following observations and
experiences during his/her life or has—at least—heard about them:
• People saying nasty things about minority groups.
• People believing that women are inferior to men and that they should stay
at home, care for their children and do the housework instead of working
in a job.
Reliability
Babbie gives the following definition for reliability [5, p. 141].
The two subsequent examples shall help to explain the term. For easier under-
standability, the examples are taken from measuring physical properties.
Imagine, for example, that one is interested in a person’s weight. One possible
technique could be to ask two different people to estimate the person’s weight.
Presumably, these estimates would vary a lot. As an alternative, one could use
a bathroom scale for measuring the person’s weight. If the person steps on the
scale twice, it would likely report the same weight in both cases. So, the second
method is much more reliable than the first one. [5, p. 141]
Kan gives another good example [69, pp. 70–71]: If one wants to measure
the body height of a person, one gets more reliable values if one uses a precise
operationalization which specifies the time of the day to take the measurement,
the scale to use, who takes the measurement, whether the measurement should
be taken barefooted and so on.
Frequent threats to reliability are:
Validity
Validity is defined by Babbie [5, p. 143] as follows:
Definition A.15 (Validity) Validity refers to the extent to which a measure adequately
reflects the real meaning of the concept under consideration.
Definition A.16 (Face validity) Face validity is that quality of an indicator that makes
it seem a reasonable measure of some variable.
Even though one might discuss whether there exist better indicators or not, it
seems to make sense without a lot of explanation that the frequency of church
attendance is some indication of a person’s religiosity—the measure is valid “on
its face”.
Definition A.18 (Construct validity) Construct validity is the degree to which a mea-
sure relates to other variables as expected within a system of theoretical relationships.
Imagine one wants to study the influence of marital satisfaction on the proba-
bility of cheating the spouse. For that purpose, a measure for marital satisfaction
is developed whose validity has to be assessed. Furthermore, the hypothesis that
marital satisfaction decreases the probability of cheating the spouse is made. If
the measure relates to this expectation, that constitutes evidence for its construct
validity.
Definition A.19 (Content validity) Content validity is the degree to which a measure
covers the range of meanings included within a concept.
(a) Reliable but not valid. (b) Valid but not reliable. (c) Reliable and valid.
Figure A.1: A graphical analogy of the difference between reliability and validity [5,
p. 145].
Graphical Analogy
Babbie gives a good graphical analogy of the difference between reliability and
validity (see Figure A.1) [5, p. 145]: He proposes to think of measurement as
analogous to repeatedly shooting at the bull’s-eye on a target. Then, reliability
looks like a tight pattern of shots somewhere on the target (see Figure A.1a)—
validity is a set of shots arranged around the bull’s-eye (see A.1b). A failure
of validity reveals itself as shots which are systematically off the mark (see
Figure A.1a)—a failure of reliability as randomly distributed shots around the
target (see Figure A.1b). Only reliable and valid metrics (see Figure A.1c) are
useful.
BASICS OF EMPIRICAL RESEARCH
B
In this thesis, controlled experiments are used twice for examining effects of
measuring structural process model understandability (Chapter 6) and effects of
process model granularity (Chapter 7). This chapter gives the necessary theoretical
background to the general need of empirical research and conducting controlled
experiments in particular. It is mainly based on textbooks by Juristo and Moreno
[68] as well as Wohlin et al. [178].
b.1 motivation
The area of BPM has developed and proposed a wide range of modeling lan-
guages (see Section 2.3), tools, methods and process model metrics. In many
publications, additional hypotheses about the alleged superiority of a special
modeling approach, a tool, etc. are being made. Often, the authors present some
motivation why the postulated hypothesis shall be correct—yet, a stringent
validation is often missing (see Section 3.6).
Juristo and Moreno emphasize the importance of validation in scientific re-
search [68, p. 24]: “For a body of knowledge to be considered scientific, its truth
and validity must be proven. A particular item of knowledge is considered to be
scientifically valid if it has been checked against reality. [. . . ] Scientific research is
the antithesis of opinion. Ideally, researchers do not opine, they explain objective
results. Their studies are not based on subjective factors, like emotions, opinions
or tastes. Scientific investigations are objective studies, based on observations of
or experimentation with the real world and its measurable changes.”
As most aspects of BPM cannot be mathematically modeled, a mathematical
proof of a hypothesis is often not possible1 .
Furthermore, one has to state that BPM is a human-intensive discipline. So,
it “can be considered as a social process in that the artifacts (methods/tools/
paradigms) to be used are affected by the experience, knowledge and capability
of the user”2 [68, p. 26].
These facts show the need for empirical studies which “have traditionally been
used in social sciences and psychology, where we are unable to state any laws of
nature, as in physics” [178, p. 5].
1 There are exceptions: Some properties of process models (e. g., soundness, reachability, existence
or absence of deadlocks) are mathematically provable.
2 Originally, Juristo and Moreno write about software engineering. Yet, their statement is transfer-
able to BPM.
202 basics of empirical research
There are several empirical approaches—each with its special advantages and
disadvantages. Consequently, one has to select the adequate approach with regard
to the question which is to be examined.
In this section, three approaches (surveys, case studies and controlled experi-
ments) are presented in more detail. At the end of the section, these approaches
are compared according to their properties.
b.2.1 Surveys
Surveys are often done in retrospect, after having used a (new) technique for
some time in order to get “a snapshot of the situation to capture the current
status” [178, p. 10].
They consist of a set of questions which a group of persons—representatives
from the population to be studied—is asked to answer. Afterwards, it is tried to
generalize the results to the whole population.
Using surveys, one can collect data about subjective experiences, opinions,
attitudes, etc. Yet, it is not possible to get objective values as, for example, the
number of errors or the execution time of a process.
The questions of a survey can be asked either using (online) questionnaires
or by conducting interviews. The advantages of (online) questionnaires are that
the researcher is not needed while the respondents are answering the questions
and the time for answering can be chosen within a specific period of time.
Interviews (by telephone or face-to-face), on the other side, also have some
advantages: Typically, the response rates are higher for interviews than for (online)
questionnaires. Furthermore, the number of “do not know” and “no answer” are
lower as the interviewer can answer possible questions of the respondents. [178,
pp. 8, 10–12]
Each approach has its advantages and disadvantages. So, the choice of the
empirical approach “depends on the prerequisites for the investigation, the
purpose of it, available resources and how we would like to analyze the collected
data” [178, p. 17].
The presented approaches can be compared using the following three factors
[120, p. 19] [178, pp. 16–17].
• Level of control
The main goals of empirical approaches are to gain insights into depen-
dencies, influencing factors and to examine proposed hypotheses. So, it
is necessary that “interesting” variables (i. e., variables which are poten-
tial influencing factors) are well adjustable to desired values and that the
whole investigation is controllable by the experimenter according to his/her
wishes.
• Level of replication
A very important requirement of empirical approaches is the possibility of
replication—either by the same experimenter or by other researchers. Only
this way, the confidence in the observed results can grow.
A necessary prerequisite for replication is that the investigation can be re-
peated under the same conditions. That means that all important conditions
of the first run are determinable and re-adjustable for further runs.
• Investigation costs
The investigation costs of an empirical approach are comprised of the
required time for preparation, execution and final evaluation of the investi-
gation and the required financial resources. These financial costs are mainly
caused by labor costs of the persons involved in the investigation. The labor
204 basics of empirical research
time and costs of the experimenter (for preparation and evaluation) are nor-
mally negligible compared to those of the participants of the investigation.
So, the main factor for the costs is the size of the investigation (i. e., the
number of participants and the amount of time they have to spend).
• Surveys
Surveys have a high level of control as the asked questions and response
options are definable in advance. Yet, “only” subjective values (e. g., personal
experiences, opinions and attitudes, etc.) are measurable.
A survey can be repeated with exactly the same set of questions and
response options—either with the same group of subjects or another. So,
this approach has a high level of replication.
Finally, the investigation costs are quite low even for large groups of sub-
jects: The effort for preparation is almost independent from the number of
subjects—the final evaluation of the given answers can be automated when
using online questionnaires. The subjects can choose when to answer the
questions within a specific period of time. The investigation is independent
from running projects in a company or expensive experimental setups.
• Case studies
Case studies have a low level of control. They are always connected to—and
depending from—a running project. The development of the project over
time is not predictable in advance. So, it is hardly possible to get exactly
the desired constellation.
Consequently, also the level of replication is low as it is almost impossible
to run the case study under the same circumstances in another project.
The investigation costs are medium (between those of surveys and controlled
experiments).
B.3 controlled experiments 205
• Controlled experiments
For controlled experiments, special experimental designs with exactly the
desired circumstances can be used. So, they have a high level of control.
As these experimental designs allow to run an experiment several times
under equal circumstances (with different groups of subjects), this approach
has also a high level of replication.
A big drawback are the relative high investigation costs for controlled experi-
ments. They are caused by the laborious preparation for such an experiment
and mainly by the expenditure of time of the subjects. Participation in an
experiment may not only be very time-intensive for each participant (e. g.,
for experiments which simulate longer processes)—it also prevents him/her
from productive work at the same time in case of using employees of a
company as subjects.
As high levels of control and replication are essential for the empirical anal-
yses in this thesis, controlled experiments were selected among the presented
approaches. The high investigation costs of a controlled experiment—at least the
monetary costs—could be minimized by using students as subjects.
According to Wohlin et al. [178, p. 31], the starting point of a controlled ex-
periment is the idea of an existing relationship between a cause and an effect.
This relationship is expressible as a theory or—even more formally—it can be
formulated as a hypothesis.
In order to test this assumed relationship, one can conduct a controlled experi-
ment. It has the advantage of relative high control about the exact experimental
conditions.
In the design of the experiment, one uses several different values (so-called
levels) of the presumed influencing factor(s). During its execution, one observes
the outcome (so-called response variable(s)) of these different conditions.
If the controlled experiment is properly set up, conclusions about the assumed
relationship and/or hypothesis can be drawn after the analysis of the experiment.
Figure B.1 illustrates the general configuration of a controlled experiment. The
used technical terms are explained in Subsection B.3.2.
206 basics of empirical research
factor(s)
experiment design
. .
independent . . experiment response variable(s)
variables . .
controlled
variables
Definition B.2 (Object) An object is a thing (e. g., a process model) which is used (e. g.,
executed or analyzed) or created within a controlled experiment [178, p. 34].
There are two types of independent variables: factors and controlled variables.
3 Originally, Juristo and Moreno write about software engineering. Yet, their statement is transfer-
able to BPM.
B.3 controlled experiments 207
Definition B.7 (Response variable) A response variable (also called dependent vari-
able) measures the outcome of a controlled experiment—with it the possible influence of
the independent variables [68, p. 59] [178, p. 33].
According to Wohlin et al., the “starting point for an experiment is insight, and
the idea that an experiment would be a possible way of evaluating whatever
we are interested in. In other words, we have to realize that an experiment is
appropriate for the question we are going to investigate.” [178, p. 35]
An experiment is very time- and resource-intensive. Furthermore, many (com-
mon) mistakes can be made which may make the results useless. So, after the
decision for an experiment, it is necessary to carefully plan its realization. Wohlin
et al. propose an experiment process for how to perform experiments (see Fig-
ure B.2) which is explained in the rest of this subsection.
Definition
The first phase of the experiment process is the definition phase [178, pp. 37,
41–46]. In this phase, the foundation of the experiment is determined. All further
steps are then based on this foundation.
208 basics of empirical research
Wholin et al. propose a goal definition template published in [16, pp. 255–256]
and [81, pp. 243–245] for that purpose. It comprises the aspects
• object of study,
• purpose,
• quality focus,
• perspective and
• context.
object of study The object of the study is the entity which is studied in the
experiment. It can be, for example, an approach, a process definition language, a
metric, a tool or a theory.
quality focus The quality focus is the primary effect of interest in the
experiment. Examples are effectiveness, cost, reliability and maintainability.
perspective The perspective defines the viewpoint from which the exper-
iment’s results are interpreted. Possible perspectives are, for example, project
analyst, developer, user and researcher.
Planning
The definition phase is followed by the planning phase [178, pp. 37–38, 47–74].
While the definition phase defines why the experiment is conducted, the planning
phase determines how it is conducted [178, p. 47]. It can be divided into the steps
• context selection,
• hypothesis formulation,
• selection of variables,
• selection of subjects,
• experiment design,
• instrumentation and
• validity evaluation.
B.3 controlled experiments 209
context selection In the context selection step [178, pp. 48–49], the “situa-
tion” of the experiment is determined.
In order to get the most general and realistic results, it would be best to run
an experiment in connection with a large real world project (“on-line”) with
professionals involved. Because of the high costs, the additional risks for the
project or simply a missing fitting project, one often has to switch to experiments
with students dealing with smaller “toy problems” which are not linked with a
real project (“off-line”).
Summing up, one can characterize the context of an experiment according to
the four dimensions
selection of variables In this step [178, p. 51], the independent and re-
sponse variables are selected. Normally, these variables can be identified by
looking at the hypothesis formulated in the previous step. Thereby, the inde-
pendent variables measure the possible influencing values—while the response
variable(s) measure(s) the outcome. The identified independent variables have
to be further divided into factors (different variable values are used during the
experiment) and controlled variables (variable value is kept constant throughout
the experiment).
selection of subjects The next step [178, pp. 51–52] consists of the selec-
tion of the subjects for the experiment. Here, only an abstract decision is made.
Concrete individuals are not searched until the following operation phase.
The selected subjects can be seen as a sample of the overall population desired
for the experiment. In order to generalize the results, the sample has to be
representative for that population.
210 basics of empirical research
The number of selected subjects (sample size) impacts the quality of the gen-
eralization of the results—the larger the sample, the lower the generalization
error. The desirable sample size depends on the properties of the used statistical
hypothesis test method and the variability of the overall population.
experiment design In this step [178, pp. 52–62], the experiment design is
chosen.
A controlled experiment consists of applying several different levels of the
factor(s) to the subjects for later statistical analysis. Thereby, the experiment
design determines how this series of single tests is organized. A design consists
of the used factor levels, the order of the single tests and the assignment of the
subjects to these tests.
Depending on the number of factors in the experiment, one distinguishes
between designs with one or at least two factors. As only experiments with one
factor are used in this thesis, only information about these designs are given here.
Details about designs with at least two factors can be found, for example, in [178,
pp. 58–62] and [107].
Independent of the selected experiment design, there are two main principles:
randomization and balancing.
Most statistical methods which are used for later analysis of the collected
data require that the observations come from independent random variables.
Randomization is used to meet this requirement. It is applied on the allocation of
objects and subjects to the single tests as well as the test order. Furthermore, it
is used to assure the representativeness of the selected subjects for the overall
population. It works as it averages out other effects.
Balancing means the assignment of the same number of subjects to each factor
level of the experiment. Even if it is not absolutely necessary, it simplifies and
strengthens the statistical analysis of the data.
Experiment designs with one factor and two or more levels are normally
used to compare, for example, several tools, modeling methods, etc. Using a
completely randomized design, the subjects are randomly assigned to exactly one of
the different factor levels. In the case of a randomized complete block design, each
subject participates in every single test (every factor level). Only the order in
which a subject is exposed to the factor levels is randomly determined.
The measurement instruments are used for collecting data during the ex-
periment. Possible examples are (online) questionnaires (as in Chapter 6) and
computer-based experimentation environments (as in Chapter 7).
validity evaluation The last step in the planning phase is validity evalua-
tion [178, pp. 63–73].
It is important to deal with the question of validity already in the planning
phase in order to gain usable results. If possible problems are identified, it is still
possible to adjust the experiment before it is actually executed.
In the literature, two important aspects of validity are distinguished.
Definition B.9 (Internal validity) Internal validity refers to the extent to which one
can accurately infer that the independent variable caused the effect observed on the
response variable and that it is not a result of a factor which one has no control of or has
not measured. [27, p. 217] [178, p. 64]
Wohlin et al. explain [178, p. 65]: “Threats to internal validity concern issues that
may indicate a causal relationship, although there is none. Factors that impact on
the internal validity are how the subjects are selected and divided into different
classes, how the subjects are treated and compensated during the experiment, if
special events occur during the experiment etc. All these factors can make the
experiment show a behavior that is not due to the treatment but to the disturbing
factor.”
Definition B.10 (External validity) External validity refers to the extent to which the
results of an experiment can be generalized out of the scope of the study—namely across
variations in people, settings, treatments, outcomes and times [27, p. 247] [178, p. 65].
Wohlin et al. illustrate [178, p. 65]: “Threats to external validity concern the
ability to generalize experiment results outside the experiment setting. External
validity is affected by the experiment design chosen, but also by the objects in the
experiment and the subjects chosen. There are three main risks: having wrong
participants as subjects, conducting the experiment in the wrong environment
and performing it with a timing that affects the results.”
In the subsequent lists, important threats to internal and external validity are
presented in more detail.
Important threats to internal validity:
• History
The history threat comprises events which occur between the beginning of
an experiment and the measurement of the response variable and could
possibly influence the subjects’ behavior [27, pp. 219–220] [31, p. 51].
• Maturation
The maturation threat refers to changes in the internal conditions (both
biological and psychological processes) of a subject as a function of time
as, for example, age, learning, fatigue, boredom and hunger. These changes
could influence the subject’s behavior. [27, pp. 220–221] [31, p. 52]
212 basics of empirical research
• Instrumentation
This threat is caused by changes in the measurement of the response variable
over time. Automatic and physical measurements normally show no or only
small temporal changes. Yet, in cases in which a human observer measures
and evaluates the response variable, this could really become a problem
similar to that of maturation. [27, p. 221] [31, p. 52]
• Testing
The testing threat refers to the fact that answers given to questions of a test
by a subject—and thereby the subject’s achieved score—can change (the
score often increases) for a second run of the test. This is caused as the
subject already has experience with the questions—even if he or she was
not given the correct answers before the second run. [27, p. 221] [31, p. 52]
• Regression
The regression artifact refers to the fact that extreme scores in a particular
distribution will tend to move—or regress—toward the mean of the distri-
bution as a function of repeated testing. In other words: The scores of the
high groups may become lower and vice versa—without any change of the
experimental settings itself. [27, pp. 222–223] [31, pp. 52–53]
• Mortality
The mortality threat (called “attrition” in [27, pp. 223-224]) refers to possible
problems caused by subjects dropping out of an experiment before it is
finished. This can destroy the representativeness of the remaining subjects
and lead to effects not caused by the different factor levels but by properties
of a subject which increase the possibility of leaving the experiment early.
[27, pp. 223-224] [31, p. 53]
• Selection
The selection threat exists when different selection procedures are used
for assigning the subjects to the different groups of the experiment. This
could lead to inequalities between the groups according to properties as
intelligence, age and previous knowledge which also influence the outcome
besides the actual factor(s). The best method to prevent this threat is to
randomly choose the subjects from the overall population and to randomly
assign them to the different groups of the experiment. [27, p. 224] [31, p. 53]
• Population validity
Population validity (called “interaction of selection and treatment” in [31,
p. 73] and [178, p. 73]) refers to the ability to generalize the results from
the subjects of an experiment (sample of the population) to the overall
population one is interested in. [27, pp. 248–250] [31, p. 73] [178, p. 73]
B.3 controlled experiments 213
Christensen explains the problem more detailed [27, p. 248]: For that sake, he
distinguishes between two kinds of populations. The target population is the
larger population one actually wants to generalize to (e. g., all persons with
BPM knowledge). The experimentally accessible population is that population
available to the researcher (e. g., students with BPM knowledge). The desired
generalization now consists of two steps. In the first step, one generalizes
from the sample of subjects to the experimentally accessible population.
If the sample is randomly selected and large enough, this is uncritical.
The second step requires to generalize from the experimentally accessible
population to the target population. This seldom can be made with any
degree of confidence as the experimentally accessible population is rarely
representative of the target population. In the example given above, one does
not know exactly how representative students with BPM knowledge are to
the population of all persons with BPM knowledge including professionals.
• Ecological validity
Ecological validity (called “interaction of setting and treatment” in [31,
p. 74] and [178, p. 73]) refers to the generalizability of the results of an
experiment across settings and environmental conditions. Examples are a
laboratory situation compared to a real project situation or a toy problem
compared to a real world process. [27, p. 251] [31, p. 74] [178, p. 73]
• Temporal validity
Temporal validity (called “interaction of history and treatment” in [31, p. 74]
and [178, p. 73]) refers to the extent to which the results of an experiment
can be generalized across time. The subjects’ behavior could differ, for
example, between Mondays—returning from weekend—and Fridays—after
a hard work week—or between the end of a labor-intensive project and the
day returning from a recreative holiday. [27, pp. 251–252] [31, p. 74] [178,
p. 73]
Operation
The next phase of the experiment process is operation [178, pp. 38, 75–80]. In this
phase, the different treatments of the subjects are actually conducted. It consists
of the three steps
• preparation,
• execution and
• data validation.
data validation The phase finishes with the data validation step. Here, the
collected data is checked for reasonability. In case of a questionnaire, for example,
it is checked whether the subjects have understood how to answer the questions
(not whether the answers are correct!) and whether the answers are serious4 .
Otherwise, the answers of a subject are rejected.
• hypothesis testing.
hypothesis testing The second step is hypothesis testing [117, pp. 483–
496] [178, pp. 49–50, 92–95]. In this step, the “correctness” of the hypothesis
formulated in the planning phase is examined using the collected data and
statistical methods.
4 If a subject, for example, has chosen the same answer possibility for each question or the time
spent for answering an online questionnaire is extremely short, it is very likely that this subject’s
answers are not serious.
B.3 controlled experiments 215
Table B.2: Possible outcomes of hypothesis tests [117, p. 488] [156, p. 424].
decision
do not reject H0 reject H0
type I error
H0 true correct decision
P(type I error) = α
reality
type II error
H0 false correct decision
P(type II error) = β
• Type I error
A type I error occurs if a true H0 is rejected. In other words, a non-existing
cause-effect relationship is expected to exist. The probability of this type of
error is
• Type II error
A type II error occurs if a false H0 is not rejected. In other words, an existing
cause-effect relationship is expected not to exist. The probability of this type
of error is
Both types of errors cannot be eliminated. Furthermore, they are related in that
way that decreasing the probability of one error type increases the probability of
the other one. Normally, one tries to decrease the probability of type I errors—
minimizing the risk that a cause-effect relationship expected to exist is still not
existing [117, p. 491]. Typical values of α are 0.05 or less.
At this point, one needs a decision rule which tells whether or not to reject the
null hypothesis H0 according to the collected data in the experiment.
For that purpose, one selects a test statistic T or a random variable whose
sampling distribution is known under the assumption that H0 is true. Now, one
can compute the realization t of the test statistic T for the observed experiment
data.
The range of possible values of the test statistic T is partitioned into two
regions—the critical region (or region of rejection) R and the region of non-
rejection R. The critical region consists of those sample realizations of T for which
the null hypothesis H0 is rejected. The boundary between the critical region and
the region of non-rejection is called the critical value tc . In order to compute tc
and subsequently R, one can modify (B.1) into
So, the location and size of the critical region depend on the null hypothesis, the
alternative hypothesis and the level of significance α.
In the above example, the so-called t-test [178, p. 99] can be used. In this case,
the test statistic T is a Student’s t-distributed t [117, pp. 357–361] which is mainly
computed as the difference between the average error probabilities for both
techniques multiplied by some correction factors (including the sample variances
for both techniques). This test statistic is symmetric to the t value 0, its expected
value and maximum are both 0. A possible example of such a test statistic T is
depicted in Figure B.3. The critical region is at the lower tail of the distribution.
The critical value tc can be computed so that
P(type I error) =
tc
t
critical region
In the critical region, the average error probability for the new technique, which
was measured during the experiment, is significantly lower than that for the
existing technique on the significance level α. Consequently, the null hypothesis
(equal error probabilities) is rejected.
Beside the significance level α, there also exists the so-called p-value. While
α has to be chosen before an experiment is actually executed and influences the
position of the critical value tc , the p-value is computed after the experiment
execution in the “opposite” direction: Here, the critical value tc is chosen so that
it is equal to the realization t of the test statistic T . Afterwards, p is computed as
the probability of a type I error using that critical region.
The general hypothesis test procedure is summarized by Panik as [117, pp. 495–
496]:
1. Formulate the null hypothesis H0 (assumed true) and the alternative hy-
pothesis HA (it determines the location of the critical region R).
5. Compute the value t of the test statistic (according to the measured experi-
ment data).
Finally, the results of the hypothesis test have to be interpreted: If the null
hypothesis could be rejected, the assumed influence of the independent variable(s)
on the response variable(s) is most likely. Considering the external validity of
the experiment, general conclusions can be made for settings similar to those
218 basics of empirical research
of the experiment. Furthermore, the results can cause decisions about future
applications of the analyzed tool, method, etc. in (commercial) projects and give
ideas for additional experiments. [178, pp. 38, 112–113]
No specific hypothesis tests are explained here. The used tests are presented
within the thesis at the place where they are used.
• X influences Y.
• Y influences X.
Watkins et al. give an example for the last case [171, p. 142]: In a sample of
elementary school pupils, there is a strong correlation between shoe size and
scores on a test of mathematical abilities. Nevertheless, neither mathematical
skills cause feet to get bigger nor vice versa. In fact, the lurking variable age
influences both foot size and mathematical knowledge.
In the remaining chapter, two often used measures of correlation—Pearson’s
product-moment correlation coefficient for measuring linear dependencies (Sec-
tion C.1) and Spearman’s rank correlation coefficient for measuring monotonic
dependencies (Section C.2)—are introduced.
10
10
10
●
●
● ●
●●
●
●
● ●
● ●
●●
●
● ●
●
● ●
●
● ●
8
8
●
●
●● ●
●
● ●
● ●
● ●
●● ●●
●●
●
●●
●●●
● ●● ●
●
● ●● ●
● ●●● ●
●
●
●●● ●
●●
●● ●
6
6
●
● ●●
● ● ●●
●●●
● ●
●● ●
●
● ●
●●●
● ●
●●
● ●●
●●
●
●● ●● ●
●
●● ● ●●
● ●●● ●●
● ●● ●
●●
y
y
●
y
●● ●
●
●●● ●● ●●●●
● ●●●
●● ●
●
●●●
●
●
● ●● ●●
●
● ●
● ●●
●
●
● ●●●
● ●
●●
4
4
●●● ●
●● ●● ●
●
●
●
●● ●●●
●
●
●●● ●●
●● ●
●● ●●
●
●
●●●
●
● ●● ●
●
● ●●
● ●
●
●● ●
●●
● ●●
2
2
●
●● ●
●●
●● ●
●
●
●
● ●
●
●
●
● ●
●●
●
● ●
●
0
0
0 2 4 6 8 10 0 2 4 6 8 10 0 2 4 6 8 10
x x x
10
10
● ● ● ●
● ●●
● ●
●● ●
●● ● ●
● ●
● ●
●
● ●
●● ● ●● ●
●
● ●
●
● ● ●
8
8
●
●
● ● ●●
● ● ●
● ● ●
● ●
● ● ● ● ●
●
● ●● ● ● ●● ●
● ● ●
● ● ● ●● ●
● ●
● ● ●
● ●
● ● ●●
● ●
● ●●
● ● ● ● ● ●
● ● ● ●
6
6
● ● ●●
● ● ● ● ● ● ●
●● ● ● ● ● ●
● ● ●
● ●● ●● ● ● ● ● ●●
● ● ●● ● ● ● ● ● ● ●● ● ●
● ● ● ● ●● ●
y
y
● ● ● ● ● ● ●
● ● ● ● ● ●
● ● ● ● ● ●
● ● ● ● ●
● ●● ● ● ● ● ● ●
● ● ● ● ● ● ●
● ● ● ● ●
● ● ● ●
4
4
● ● ● ● ●
● ●
● ●● ● ●● ● ●
●● ● ●● ● ●●
●● ●● ● ●
● ●
● ● ●
● ● ● ● ● ●
● ● ● ●
●
● ● ● ● ● ● ●● ●
●
● ●● ●
●
●● ● ● ● ●
●
●
2
2
● ● ● ●
● ●● ●
●●
●
●
● ●
●● ●
● ●
● ● ●
● ● ●
● ●● ● ●
0
0
0 2 4 6 8 10 0 2 4 6 8 10 0 2 4 6 8 10
x x x
10
10
●
● ●● ●●
●
● ●
●●
●●
● ●●
● ●
●● ●●●
● ● ●
●●
● ●
● ●
● ● ●
● ●
●
●● ●
● ●
● ●
● ●
●
8
●
●
● ●
● ●
● ●●
●
● ●
● ●
● ● ● ●
● ● ● ● ● ● ●
● ●
●
●
● ● ● ●
● ● ●● ●
6
● ● ● ● ●
● ● ● ●
● ● ● ● ●
● ● ● ●● ● ● ●
● ● ● ● ●
●
● ● ● ●●
● ●● ● ● ● ●
y
● ● ●● ●
● ●
● ● ● ● ● ● ● ●
● ●
● ● ● ● ● ●
● ● ●
●
● ●
● ●●●
● ●● ● ●
● ● ● ●
4
● ●
● ●
●●
● ● ●
● ● ● ●
● ●
● ●
● ● ● ●
● ●
●
● ●
● ●
● ● ●
●
● ● ● ●
● ●
●
● ●
●●
2
●
●
● ●
●
●
●
● ● ●
● ● ●
● ●
● ●
● ● ●
● ● ●●
●
●● ●
●
●●● ●●
●
●●
● ●
●
● ●
●● ●●● ●
●●●●
●● ●
0
0 2 4 6 8 10 0 2 4 6 8 10 0 2 4 6 8 10
x x x
Figure C.1: Pearson’s correlation/Spearman’s rank correlation for different sets of value
pairs (Part 1 of 2).
Definition C.1 ((Sample) covariance) The (sample) covariance cov of a set of ob-
served value pairs (xi , yi ), i ∈ {1, . . . , n} is defined as
1 X
n
cov := (xi − x)(yi − y) (C.1)
n−1
i=1
For value pairs with a positive linear dependency (the larger the x value, the
larger the y value), the (sample) covariance has a positive value—for pairs with a
C.1 pearson’s product-moment correlation coefficient 221
10
10
10
●● ●
●
●
●
●
●
●
● ● ●
●
● ● ●
● ●● ● ●
●
●
●
8
8
● ●
8
● ● ●
●
● ●
●● ● ●
● ● ● ● ●●●
●● ● ●● ● ● ●
● ●
● ● ● ●● ●
● ● ●● ●
● ● ● ●
● ● ● ●
● ● ●
● ● ● ●
●
6
6
● ● ● ● ● ●● ●
●● ● ● ● ● ● ● ●
● ● ●
● ●
● ● ●
●
● ● ●● ●●
● ● ● ●
● ● ● ●
● ● ● ● ● ●●
●
● ● ●● ● ●
● ●●
● ● ● ● ●
y
y
● ● ● ● ● ●● ● ●
●● ● ●● ●
● ● ● ● ● ● ● ● ●
●● ●
● ●● ● ● ●
● ● ● ● ●● ●
● ● ● ● ●
● ● ●● ●● ● ● ● ●
● ●
● ● ● ●
●●
4
4
● ● ● ●
● ● ● ●
● ● ● ● ●
●● ●
● ● ● ● ●
● ● ●●
● ● ●
● ● ●●
●
● ● ●
● ●●●● ● ●●
●
● ● ● ●
● ● ●
● ● ● ● ●
2
2
● ●
● ● ●
● ●
● ● ●
● ●
● ● ●
● ●
●
● ●● ●
●●●
● ● ●
●
● ●● ●● ●
0
0
0 2 4 6 8 10 0 2 4 6 8 10 0 2 4 6 8 10
x x x
10
10
●
●
●●●
●●
●
● ●
●
●● ●
●
●● ●
●
8
8
●
●●
● ●
● ●
●● ●
●● ●
●●● ●
● ●●● ●
●
●
●●
● ●
●
●●● ●
● ●●
●
●● ●● ●
●●
● ●●●
●
●
●
6
●●
6
●
● ●
●●
● ●
●● ●● ●
●
●● ●
●● ●●
●
●● ● ●
●● ●●
●●
●● ●●● ●
●● ● ●● ● ● ●●●
●●●●●●
y
y
● ● ●
●● ●● ●●●
●● ● ●
●●●
●
●
●
● ●
●
● ●
●
●● ● ●●
●● ●
4
4
●
●● ●
●●
●
● ●
●
●
● ●● ●●
●●● ●
● ●● ●
●
● ●●
●● ●● ●
●
●● ●
●● ●
●
● ●● ●
● ●●
●
●●
●● ● ●
●●
●● ●
●
2
2
●
●
●
●
●
●● ●
●●
●● ●
●
●
●● ●
●
● ●●
●
●
●●
0
0 2 4 6 8 10 0 2 4 6 8 10 0 0 2 4 6 8 10
x x x
Figure C.1: Pearson’s correlation/Spearman’s rank correlation for different sets of value
pairs (Part 2 of 2).
X Y
X Y X Y Z
negative dependency (the larger the x value, the smaller the y value) a negative
one.
Yet, covariance has a critical disadvantage: the covariance values of different
sets of value pairs are not comparable. If one measures, for example, the same
set of value pairs with two different units of measurement (e. g., xi0 := 2xi and
yi0 := 2yi ), one gets two different values for the covariance. As the underlying
dependencies are the same in both cases, this effect is unsatisfying.
This problem can be solved by dividing the covariance by the product of the
standard deviations of the xi and yi —resulting in the definition of Pearson’s
(sample) product-moment correlation coefficient [109].
222 measuring correlations
As this measure has the following desired properties, its values are now
comparable for different sets of observed value pairs.
Theorem C.1 For the Pearson’s (sample) product-moment correlation coefficient r, the
properties
1. −1 6 r 6 1 and
hold.
holds with equality if and only if there is a c ∈ R such that ai = rbi for all
i ∈ {1, . . . , n}. (Cauchy-Bunyakovsky-Schwarz inequality) [38, Theorem 2.1, p. 5]
regarding 2.)
• For positive dependencies (the larger the xi , the larger the yi ), Pearson’s
product-moment correlation coefficient is positive (see Figure C.1a, C.1b,
C.1d and C.1e)—for negative dependencies (the larger the xi , the smaller
the yi ), it is negative (see Figure C.1j, C.1k, C.1m and C.1n).
• The less perfect the linear dependency is, the smaller the absolute value
of Pearson’s product-moment correlation coefficient (see Figure C.1d, C.1e,
C.1j and C.1k).
• If the value pairs are nearly randomly distributed (no linear dependency at
all), Pearson’s product-moment correlation coefficient is approximately 0
(see Figure C.1g).
12
12
●
10
10
●
●
● ●
● ● ●
●
● ● ●
8
8
●
● ●
●
y
y
●
6
6
●
● ●
●
4
4
●
2
2
0
0
0 5 10 15 20 0 5 10 15 20
x x
●
●
12
12
10
10
● ●
●
●
8
● ●
●
●
● ●
●
●
●
y
●
●
6
● ●
●
●
●
4
4
2
2
0
0 5 10 15 20 0 5 10 15 20
x x
Figure C.3: Anscombe’s quartet: Four data sets with Pearson’s product-moment correla-
tion coefficient of 0.816 (Figure C.3a–C.3c) and 0.817 (Figure C.3d), respec-
tively [3].
(Figure C.3d). Nevertheless, it can be easily seen that some of them are not
linearly dependent.
So, when having found a high correlation coefficient, visually verifying the
actual existence of a linear dependency by depicting the values in a scatter plot is
a good idea.
C.2 spearman’s rank correlation coefficient 225
example Using the values of Table C.1, the computation of the ranks and
Spearman’s rank correlation coefficient shall be demonstrated.
For the determination of the ranks, valid ranking orders for the xi and yi have
to be found. As one can easily see, both the xi and yi have ties. Consequently,
all three x values of 2 get the rank 3 (as defined in the second alternative of
equation C.8) and the two y values of 8.7 get rank 6.5. The computation of the
other values’ ranks is “intuitive”.
According to Theorem C.2, the means of the ranks are rankx = ranky = 8+1 2 =
4.5.
Inserting the computed ranks and their means in (C.12) results in
X
n
2
rank(xi ) − rankx
i=1
= 12.25 + 2.25 + 2.25 + 2.25 + 0.25 + 2.25 + 6.25 + 12.25 = 40
as well as
X
n
2
rank(yi ) − ranky
i=1
= 2.25 + 0.25 + 4 + 12.25 + 6.25 + 4 + 0.25 + 12.25 = 41.5
[1] BPM: Not just workflow anymore. Survey, AIIM, 2007. http://www.
xerox.com/downloads/usa/en/x/XGS_article_bpm_report1.pdf (last ac-
cessed on 2011-12-01).
[6] S. Balasubramanian and Mayank Gupta. Structural metrics for goal based
business process design and evaluation. Business Process Management Journal,
11(6):680–694, 2005. doi:10.1108/14637150510630855.
[9] Victor R. Basili, Gianluigi Caldiera, and H. Dieter Rombach. Goal question
metric paradigm. In John J. Marciniak, editor, Encyclopedia of Software
Engineering, volume 1, pages 528–532. Wiley, 1994.
[10] Jörg Becker and Dieter Kahn. The process in focus. In Jörg Becker, Martin
Kugeler, and Michael Rosemann, editors, Process Management: A Guide for
the Design of Business Processes, pages 1–12. Springer, 2003.
[17] Axel Buchner. Basic topics and approaches to the study of complex problem
solving. In Peter A. Frensch and Joachim Funke, editors, Complex Problem
Solving: The European Perspective, pages 27–63. Erlbaum, 1995.
[18] Elwood S. Buffa and Rakesh K. Sarin. Modern Production/Operations Man-
agement. Wiley Series in Production/Operations Management. Wiley, 8th
edition, 1987.
[19] J. Cardoso, J. Mendling, G. Neumann, and H. A. Reijers. A discourse on
complexity of process models. In Johann Eder and Schahram Dustdar,
editors, Business Process Management Workshops: Proceedings of the BPM 2006
International Workshops, volume 4103 of Lecture Notes in Computer Science,
pages 117–128, 2006. doi:10.1007/11837862_13.
[20] Jorge Cardoso. About the data-flow complexity of web processes. In
Proceedings of the 6th International Workshop on Business Process Modeling,
Development, and Support (BPMDS’05), pages 67–74, 2005.
[21] Jorge Cardoso. How to measure the control-flow complexity of web pro-
cesses and workflows. In Layna Fischer, editor, Workflow Handbook 2005,
pages 199–212. Future Strategies Inc., 2005.
[22] Jorge Cardoso. Process control-flow complexity metric: An empirical val-
idation. In Proceedings of the 2006 IEEE International Conference on Services
Computing (SCC 2006), pages 167–173, 2006. doi:10.1109/SCC.2006.82.
[23] Jorge Cardoso. Business process quality metrics: Log-based complexity of
workflow patterns. In Robert Meersman and Zahir Tari, editors, On the
Move to Meaningful Internet Systems 2007: Proceedings of the OTM Confederated
International Conferences CoopIS, DOA, ODBASE, GADA, and IS 2007 (Part
I), volume 4803 of Lecture Notes in Computer Science, pages 427–434, 2007.
doi:10.1007/978-3-540-76848-7_30.
bibliography 231
[25] Lewis Carroll. Through the Looking-glass and What Alice Found There. Macmil-
lan, 1887.
[26] Shyam R. Chidamber and Chris F. Kemerer. A metrics suite for object
oriented design. IEEE Transactions on Software Engineering, 20(6):476–493,
1994. doi:10.1109/32.295895.
[28] BPM ist für viele Neuland. Computerwoche, (9):26, 2010. – In German.
[33] Thomas Curran, Gerhard Keller, and Andrew Ladd. SAP R/3 Business
Blueprint: Understanding the Business Process Reference Model. Prentice Hall
PTR, 1998.
[39] Dietrich Dörner. The Logic of Failure: Recognizing and Avoiding Error in
Complex Situations. Basic Books, 1997.
232 bibliography
[40] Dietrich Dörner and Alex J. Wearing. Complex problem solving: Toward
a (computersimulated) theory. In Peter A. Frensch and Joachim Funke,
editors, Complex Problem Solving: The European Perspective, pages 27–63.
Erlbaum, 1995.
[43] Norman E. Fenton and Shari Lawrence Pfleeger. Software Metrics: A Rigorous
and Practical Approach. International Thomson Computer Press, 2nd edition,
1996.
[45] Peter A. Frensch and Joachim Funke. Definitions, traditions, and a gen-
eral framework for understanding complex problem solving. In Peter A.
Frensch and Joachim Funke, editors, Complex Problem Solving: The European
Perspective, pages 3–25. Erlbaum, 1995.
[49] Félix García, Mario Piattini, Francisco Ruiz, Gerardo Canfora, and
Corrado A. Visaggio. FMESP: Framework for the modeling and evalu-
ation of software processes. Journal of Systems Architecture, 52(11):627–639,
2006. doi:10.1016/j.sysarc.2006.06.007.
[50] Micheal R. Garey and David S. Johnson. Computers and Intractability: A Guide
to the Theory of NP-Completeness. A Series of Books in the Mathematical
Sciences. Freeman, 1979.
bibliography 233
[51] Andrew Gemino and Yair Wand. Evaluating modeling techniques based
on models of learning. Communications of the ACM, 46(10):79–84, 2003.
doi:10.1145/944217.944243.
[53] Volker Gruhn and Ralf Laue. Adopting the cognitive complexity measure
for business process models. In Yiyu Yao, Zhongzhi Shi, Yingxu Wang, and
Witold Kinsner, editors, Proceedings of the 5th IEEE International Conference
on Cognitive Informatics (ICCI 2006), volume 1, pages 236–241, 2006. doi:
10.1109/COGINF.2006.365702.
[54] Volker Gruhn and Ralf Laue. Complexity metrics for business process
models. In Witold Abramowicz and Heinrich C. Mayr, editors, Business
Information Systems: Proceedings of the 9th International Conference on Business
Information Systems (BIS 2006), volume P-85 of Lecture Notes in Informatics
(LNI), pages 1–12, 2006.
[57] Sallie Henry and Dennis Kafura. Software structure metrics based on
information flow. IEEE Transactions on Software Engineering, 7(5):510–518,
1981.
[60] IEEE Computer Society. IEEE Standard 1061 - IEEE Standard for a software
quality metrics methodology, 1998.
[61] Alfred Inselberg. The plane with parallel coordinates. The Visual Computer,
1(2):69–91, 1985. doi:10.1007/BF01898350.
[65] Svante Janson, Tomasz Łuczak, and Andrzej Ruciński. Random Graphs.
Wiley-Interscience Series in Discrete Mathematics and Optimization. Wiley,
2000.
[68] Natalia Juristo and Ana M. Moreno. Basics of Software Engineering Experi-
mentation. Kluwer, 2001.
[69] Stephen H. Kan. Metrics and Models in Software Quality Engineering. Addison-
Wesley, 2nd edition, 2002.
[71] Gerhard Keller and Thomas Teufel. SAP R/3 Process-oriented Implementation:
Iterative Process Prototyping. Addison-Wesley, 1998.
[73] Ryan K. L. Ko, Stephen S. G. Lee, and Eng Wah Lee. Business process
management (BPM) standards: a survey. Business Process Management
Journal, 15(5):744–791, 2009. doi:10.1108/14637150910987937.
[75] Gang Soo Lee and Jung-Mo Yoon. An empirical study on the complexity
metrics of petri nets. Microelectronics and Reliability, 32(3):323–329, 1992.
doi:10.1016/0026-2714(92)90061-O.
[77] Frank Leymann and Dieter Roller. Production Workflow: Concepts and Tech-
niques. Prentice Hall PTR, 2000.
[79] Udo Lindemann, Maik Maurer, and Thomas Braun. Structural Complexity
Management: An Approach for the Field of Product Design. Springer, 2009.
doi:10.1007/978-3-540-87889-6.
[83] Robert M. May. Simple mathematical models with very complicated dy-
namics. Nature, 261(5560):459–467, 1976. doi:10.1038/261459a0.
[87] Joachim Melcher, Jan Mendling, Hajo A. Reijers, and Detlef Seese.
On measuring the understandability of process models (experimen-
tal results). Research report, Universität Karlsruhe (TH), Institut
AIFB; Humboldt-Universität Berlin; Eindhoven University of Technology,
2009. urn:nbn:de:swb:90-119933, http://digbib.ubka.uni-karlsruhe.
de/volltexte/1000011993 (last accessed on 2011-12-01).
236 bibliography
[88] Joachim Melcher, Jan Mendling, Hajo A. Reijers, and Detlef Seese. On
measuring the understandability of process models. In Stefanie Rinderle-
Ma, Shazia Sadiq, and Frank Leymann, editors, Business Process Management
Workshops: Revised Papers of the BPM 2009 International Workshops, volume 43
of Lecture Notes in Business Information Processing, pages 465–476, 2010.
doi:10.1007/978-3-642-12186-9_44.
[89] Joachim Melcher and Detlef Seese. Process measurement: Insights from
software measurement on measuring process complexity, quality and per-
formance. Research report, Universität Karlsruhe (TH), Institut AIFB,
2008. urn:nbn:de:swb:90-92259, http://digbib.ubka.uni-karlsruhe.
de/volltexte/1000009225 (last accessed on 2011-12-01).
[90] Joachim Melcher and Detlef Seese. Towards validating prediction systems
for process understandability: Measuring process understandability. In
Viorel Negru, Tudor Jebelean, Dana Petcu, and Daniela Zaharie, editors,
Proceedings of the 10th International Symposium on Symbolic and Numeric
Algorithms for Scientific Computing (SYNASC 2008), pages 564–571, 2008.
doi:10.1109/SYNASC.2008.24.
[91] Joachim Melcher and Detlef Seese. Towards validating prediction systems
for process understandability: Measuring process understandability (exper-
imental results). Research report, Universität Karlsruhe (TH), Institut AIFB,
2008. urn:nbn:de:swb:90-92605, http://digbib.ubka.uni-karlsruhe.
de/volltexte/1000009260 (last accessed on 2011-12-01).
[92] Joachim Melcher and Detlef Seese. Visualization and clustering of business
process collections based on process metric values. Research report, Uni-
versität Karlsruhe (TH), Institut AIFB, 2008. urn:nbn:de:swb:90-98483,
http://digbib.ubka.uni-karlsruhe.de/volltexte/1000009848 (last ac-
cessed on 2011-12-01).
[93] Joachim Melcher and Detlef Seese. Visualization and clustering of business
process collections based on process metric values. In Viorel Negru, Tudor
Jebelean, Dana Petcu, and Daniela Zaharie, editors, Proceedings of the 10th
International Symposium on Symbolic and Numeric Algorithms for Scientific
Computing (SYNASC 2008), pages 572–575, 2008. doi:10.1109/SYNASC.
2008.37.
[94] Joachim Melcher and Detlef Seese. Empirical analysis of a proposed process
granularity heuristic (experimental details). Research report, Universität
Karlsruhe (TH), Institut AIFB, 2009. urn:nbn:de:swb:90-120163, http://
digbib.ubka.uni-karlsruhe.de/volltexte/1000012016 (last accessed on
2011-12-01).
[97] Jan Mendling. Testing density as a complexity metric for EPCs. Technical Re-
port JM-2006-11-15, Vienna University of Economics and Business Admin-
istration, 2006. http://www.mendling.com/publications/TR06-density.
pdf (last accessed on 2011-12-01).
[98] Jan Mendling. Detection and Prediction of Errors in EPC Business Process
Models. PhD thesis, Vienna University of Economics and Business Adminis-
tration, 2007.
[99] Jan Mendling. Metrics for Process Models: Empirical Foundations of Verifi-
cation, Error Prediction, and Guidelines for Correctness, volume 6 of Lecture
Notes in Business Information Processing. Springer, 2008. doi:10.1007/
978-3-540-89224-3.
[100] Jan Mendling and Gustaf Neumann. Error metrics for business pro-
cess models. In Johann Eder, Stein L. Tomassen, Andreas Opdahl,
and Guttorm Sindre, editors, Proceedings of the CAiSE ’07 Forum at
the 19th International Conference on Advanced Information Systems En-
gineering, volume 247 of CEUR Workshop Proceedings, pages 53–56,
2007. urn:nbn:de:0074-247-5, http://sunsite.informatik.rwth-aachen.
de/Publications/CEUR-WS/Vol-247/FORUM_14.pdf (last accessed on 2011-
12-01).
[101] Jan Mendling, Hajo A. Reijers, and Jorge Cardoso. What makes process
models understandable? In Gustavo Alonso, Peter Dadam, and Michael
Rosemann, editors, Business Process Management: Proceedings of the 5th In-
ternational Conference BPM 2007, volume 4714 of Lecture Notes in Computer
Science, pages 48–63, 2007. doi:10.1007/978-3-540-75183-0_4.
[114] State of the business process management market 2008. White paper, Oracle,
2008.
[115] Joseph A. Orlicky, George W. Plossl, and Oliver W. Wight. Structuring the
bill of material for MRP. Production and Inventory Management, 13(4):19–42,
1972.
[116] Martyn A. Ould. Business Processes: Modelling and Analysis for Re-engineering
and Improvement. Wiley, 1995.
[117] Michael J. Panik. Advanced Statistics from an Elementary Point of View. Elsevier
Academic Press, 2005.
[118] Heinz-Otto Peitgen, Hartmut Jürgens, and Dietmar Saupe. Chaos and
Fractals: New Frontiers of Science. Springer, 2nd edition, 2004. doi:10.1007/
b97624.
bibliography 239
[131] Joseph Lee Rodgers and W. Alan Nicewander. Thirteen ways to look at the
correlation coefficient. The American Statistician, 42(1):59–66, 1988.
[132] Elvira Rolón, Francisco Ruiz, Félix García, and Mario Piattini. Applying
software metrics to evaluate business process models. CLEI Electronic
Journal, 9(1), 2006.
[135] Laura Sanchez, Andrea Delgado, Francisco Ruiz, Félix García, and Mario
Piattini. Measurement and maturity of business processes. In Jorge Cardoso
and Wil van der Aalst, editors, Handbook of Research on Business Process
Modeling, pages 532–556. Information Science Reference, 2009.
[136] Laura Sánchez González, Félix García Rubio, Francisco Ruiz González, and
Mario Piattini Velthuis. Measurement in business processes: a systematic
review. Business Process Management Journal, 16(1):114–134, 2010. doi:
10.1108/14637151011017976.
[138] Kamyar Sarshar, Philipp Dominitzki, and Peter Loos. Comparing the
control-flow of EPC and Petri net from the end-user perspective: Statistical
results of a laboratory experiment. Working Paper 25, Universität Mainz,
Information Systems & Management (ISYM), 2005. urn:nbn:de:0006-0252,
http://d-nb.info/975462385/34 (last accessed on 2011-12-01).
[139] Kamyar Sarshar and Peter Loos. Comparing the control-flow of EPC
and Petri net from the end-user perspective. In W. M. P. van der Aalst,
Boualem Benatallah, Fabio Casati, and Francisco Curbera, editors, Business
Process Management: Proceedings of the 3rd International Conference BPM 2005,
volume 3649 of Lecture Notes in Computer Science, pages 434–439, 2005.
doi:10.1007/11538394_36.
bibliography 241
[140] Sven Schnägelberger. BPM: mehr Hype als Realität? Computerwoche, (14):21,
2008. – In German.
[146] Dieter Spath and Anette Weisbecker, editors. Business Process Management
Tools 2008: Eine evaluierende Marktstudie zu aktuellen Werkzeugen. Fraunhofer
IRB Verlag, 2008. – In German.
[154] Jeffrey Travers and Stanley Milgram. An experimental study of the small
world problem. Sociometry, 32(4):425–443, 1969.
242 bibliography
[166] Irene Vanderfeesten, Jorge Cardoso, and Hajo A. Reijers. A weighted cou-
pling metric for business process models. In Johann Eder, Stein L. Tomassen,
Andreas L. Opdahl, and Guttorm Sindre, editors, Proceedings of the CAiSE’07
Forum at the 19th International Conference on Advanced Information Systems
Engineering, pages 41–44, 2007.
[167] Irene Vanderfeesten, Hajo A. Reijers, Jan Mendling, Wil M. P. van der
Aalst, and Jorge Cardoso. On a quest for good process models: The cross-
connectivity metric. In Zohra Bellahsène and Michel Léonard, editors,
Advanced Information Systems Engineering: Proceedings of the 20th International
Conference CAiSE 2008, volume 5074 of Lecture Notes in Computer Science,
pages 480–494, 2008. doi:10.1007/978-3-540-69534-9_36.
[168] Irene Vanderfeesten, Hajo A. Reijers, and Wil M. P. van der Aalst. Evaluating
workflow process designs using cohesion and coupling metrics. Computers
in Industry, 59(5):420–437, 2008. doi:10.1016/j.compind.2007.12.007.
[169] Juha Vesanto and Esa Alhoniemi. Clustering of the self-organizing map.
IEEE Transactions on Neural Networks, 11(3):586–600, 2000. doi:10.1109/72.
846731.
[173] Chris Weber. Complexity is in the eye of the beholder. Blog entry, 2008. http:
//pragmaticprose.com/complexity-is-in-the-eye-of-the-beholder
(last accessed on 2011-12-01).
[174] Gerald M. Weinberg and Daniela Weinberg. On the Design of Stable Systems.
Wiley Series on Systems Engineering and Analysis. Wiley, 1979.
[177] Leland Wilkinson and Michael Friendly. The history of the cluster heat map.
The American Statistician, 63(2):179–184, 2009. doi:10.1198/tas.2009.0033.
[178] Claes Wohlin, Per Runeson, Martin Höst, Magnus C. Ohlsson, Björn Regnell,
and Anders Wesslén. Experimentation in Software Engineering: An Introduction,
volume 6 of Kluwer International Series in Software Engineering. Kluwer, 2000.
244 bibliography
[180] Horst Zuse. Software Complexity: Measures and Methods, volume 4 of Pro-
gramming Complex Systems. Walter de Gruyter, 1991.