978 3 319 62274 3 PDF
978 3 319 62274 3 PDF
978 3 319 62274 3 PDF
Ming J. Zuo
Lin Ma
Joseph Mathew
Hong-Zhong Huang Editors
Engineering
Asset
Management
2016
Proceedings of the 11th World Congress
on Engineering Asset Management
Lecture Notes in Mechanical Engineering
About this Series
• Engineering Design
• Machinery and Machine Elements
• Mechanical Structures and Stress Analysis
• Automotive Engineering
• Engine Technology
• Aerospace Technology and Astronautics
• Nanotechnology and Microengineering
• Control, Robotics, Mechatronics
• MEMS
• Theoretical and Applied Mechanics
• Dynamical Systems, Control
• Fluid Mechanics
• Engineering Thermodynamics, Heat and Mass Transfer
• Manufacturing
• Precision Engineering, Instrumentation, Measurement
• Materials Engineering
• Tribology and Surface Technology
Engineering Asset
Management 2016
Proceedings of the 11th World Congress
on Engineering Asset Management
Editors
Ming J. Zuo Lin Ma
Department of Mechanical Engineering Brisbane, Queensland
University of Alberta Australia
Edmonton, Alberta
Canada
v
vi Preface
The joint congress hosted an excellent technical program and several opportu-
nities for networking through social events including the congress dinner. The
conference program comprised nine keynote speeches, one workshop, and 20 reg-
ular technical paper oral sessions.
The keynotes were:
1. Professor Narayanaswamy Balakrishnan, McMaster University, Ontario,
Canada, on “Parametric and semiparametric inference for one-shot device
testing.”
2. Adjunct Professor Joseph Mathew, Asset Institute, Brisbane, Australia, on
“Innovations in engineering asset management.”
3. Professor Jin Wang, Liverpool John Moors University, Liverpool, UK, on “Risk-
based decision making in design and operation of large engineering systems
under uncertainties.”
4. Professor Dong Ho Park, Hallym University, Chuncheon, Korea, on “Two-
dimensional system maintenance strategy and its effects.”
5. Professor Gao Jinji, “Process machinery diagnosis and maintenance based on
industry internet and big data analysis.”
Prof Gao receiving the ISEAM Lifetime Achievement Award from the ISEAM Chair, Joe Mathew
We would like to express our gratitude to the sponsors for providing their
generous support. The sponsors were Sichuan Provincial Key Laboratory of Reli-
ability Engineering, University of Electronic Science and Technology of China
(UESTC), National Collaborative Innovation Center of Major Machine
Manufacturing in Liaoning, Dalian University of Technology (DUT), Armed
Force Engineering Institute, Institute of Special Vehicle, National University of
Defense Technology, Northwestern Polytechnical University, The National Taiwan
University of Science and Technology, Sichuan Go Yo Intelligence Technology
Co., Ltd., Reliability committee of Sichuan Provincial Mechanical Engineering
Society, International Society of Engineering Asset Management (ISEAM),
European Safety and Reliability Association (ESRA), European Federation of
National Maintenance Societies (EFNMS), The Korean Reliability Society
(KORAS), Reliability Engineering Association of Japan (REAJ), Polish Safety
and Reliability Association (PSRA), Equipment Support Commission of China
Ordnance Society, Reliability Committee of Chinese Operations Research Society,
IEEE Chengdu Section, National Natural Science Foundation of China (NSFC),
and Center for System Reliability and Safety, UESTC.
We are grateful for the voluntary assistance provided by members of the
International Advisory Committee, the Program Committee, the Organization
Committee, and the Conference Secretariat.
Last but not the least, we would like to appreciate the Congress General Chair
and Co-Chairs: Professor Hong-Zhong Huang, UESTC, Adjunct Prof. Joseph
Mathew, Asset Institute, Australia, Prof. Carlos Guedes Soares, University of
Lisbon, Portugal, and Prof. Dong Ho Park, Hallym University, Korea; the Program
Preface ix
xi
xii Contents
1 Introduction
working areas, such as quality, production, safety, production cost, working envi-
ronment, delivery on time, etc. Therefore, a reliable and efficient maintenance not
only increases the profitability, but it also improves the overall performance of the
company [2]. Maintenance is responsible of reducing the probability of failure and
unplanned stoppages in order to minimize the impact of failure consequences on
company competitive advantages and goals through maintaining the continuity of a
production process, production cost and product quality at a predetermined rate.
Companies’ production systems are exposed to different types of failures. These
failures could be due to technical issues or other elements involved in the produc-
tion process, e.g. human resources, operating conditions, systems’ quality, facilities
reliability, etc., which should not be neglected when dealing with maintenance [2–
4]. Different types of failures have usually different consequences on competitive
advantages. For example a paper/pulp machine can be exposed to many different
failures, e.g. electrical, electronic, hydraulic, pneumatic and mechanical, which
could have different impact on the machine performance and the company com-
petitive advantages, such as production delay, lower quality, higher production
costs, etc. In order to achieve proper utilization of the available resources, failures
with significant impacts should be identified and prioritized in order to effectively
reduce/eliminate their consequences. This study presents a model that aims to
identify and prioritize failures according to their significance and impact on the
company’s competitive advantages in order to be able to plan and conduct cost-
effective maintenance. In order to investigate whether the problem addressed in this
study has been treated before, a survey was conducted and three papers [5–7] were
found relevant to the problem in this paper.
Al-Najjar [6] presented Maintenance Function Deployment (MFD) model,
which aims to pinpoint, analyze, and prioritize causes behind losses in the working
areas belonging to the competitive advantages, such as losses due to bad quality and
delayed deliveries. MFD breaks down these losses in a backwards method to
approach the root-causes behind the losses which are usually spread in the operative
level of different disciplines in a production process. MFD also provides business
based production analysis. This is done through quantifying losses (in production
time and economy), in accordance with the strategic goals of the company, and
identifying causes behind them. Then it breaks down the causes and its costs into
their root causes. The author used an example to test the model, and the results
showed that the model could be used to identify, analyze and quantify losses in
order to make cost effective improving decisions. In Al-Najjar and Jacobsson [7], a
model that demonstrates the interactions between man-machine-maintenance-econ-
omy (MMME) was developed, in order to support cost-effective decisions. The
model systematically gathers, categorizes, estimates and quantifies the losses in the
production time due to failures in order to identify and prioritize the problem areas
in the production process. A software program then was built based on the model
and tested in a case study at the automaker FIAT, Italy. The results of the case study
showed a possibility to identify the problematic areas. Also, as the model compares
the time loss in different time periods, it captures the deviations over time for
different categories. Ahamed Mohideen et al. [5] presented a method that reduces
the breakdown costs and recovering time in the construction plant systems. It starts
A Model for Increasing Effectiveness and Profitability of Maintenance. . . 3
by categorizing and analyzing the breakdown records in order to identify the main
breakdowns and the sub-breakdowns using cause effect analysis, and then ranking
them using Pareto analysis. The model was tested by a case study on four types of
machines in a construction company using four years breakdown records. Major
contributing failures and their causes were identified and a strategical plan is
proposed accordingly.
The survey showed that the impact of failures -in particular- on the companies’
competitive advantages is not payed the proper attention. The model developed in
this paper considers in particular the impact of failures on the competitive advan-
tages that companies strive to achieve, maintain and improve continuously, in order
to prioritize, analyze and support cost effective decisions for failures that have the
most significant impact. This is why the problem addressed in this study is: How to
increase maintenance effectiveness, i.e. maintenance capability of achieving a
desired result, and profitability?
The next section will discuss the impact of failures on the competitive advan-
tages. Then we will present the model development and model test using real data.
It will be followed then by results discussion and conclusion.
Failures in production process could occur due to causes arise from different
reasons, for example machines, operators, maintenance staff, management and
working environment. This is why other elements than the machine should be
considered in the model development. The spectrum and contents of maintenance
in this article is referring to a broader definition used by Total Quality Maintenance
(TQMain); TQMain aims to maintain the quality of the elements (working areas)
constituting the production process (and not only the machines) cost effectively
[3, 4].
In manufacturing companies, recorded breakdown data is very important to
enhance the production process. Special management systems and techniques are
used to classify failures and losses into predetermined categories [7–10]. The
categorization helps engineers and management to determine the breakdown that
occurs most often in each category in order to improve the related procedures. This
may help eliminating the breakdown of multiple machines in the related category
[10]. Generally, failures could be defined according to different factors, such as
failure impact on: production cost, product quality, operation safety or reason of
correction action [10]. However, these failure classifications are not sufficient to
determine the failure impact on the competitive advantages (CA) of companies.
Therefore the model developed aims to determine the impact of failures on a
company CA using the failure database available in the company. The common
strategic goals of companies in a wide range of industries is to deliver on time a
high product quality at a competitive price with low violation of environment,
accepted by the society and keeping the manufacturing machine in a reliable
condition [6]. The CA that will be considered in this paper are: high product quality,
on time delivery, competitive price, low violation of environment, society
4 B. Al-Najjar and H. Algabroun
acceptance and maintain good assets condition, i.e. asset value, see Fig. 1. Com-
panies should always utilize their resources efficiently and effectively in order to
reach the CA with a reasonable profit margin. Failing in achieving this will threaten
the company existence. In general, failures in the operative level impact negatively
on achieving these goals [3]. In order to assess the significance of failures’ conse-
quences according to their impact on the CA of a company, we choose first to
classify failures (or sort out failures registered in the company databases) according
to their working areas. This step is necessary to:
1. Analyse failures, i.e. to identify root cause, damage initiation and development,
imminent failures and failure modes.
2. Specify how failure’ consequences converted from technical to economic.
3. Quantify the impact of each failure with respect to predetermined competitive
advantages or strategic goal defined by the company, see Fig. 1.
1. Product quality failures: Failing in maintaining the condition of the cutting tools
and machines could negatively impact the quality of the final product. This may
cause rejection of processed items or final products, and therefore will cause
extra costs due to the need of the re-production of the rejected items.
2. Delivery on time failures: Long downtime of producing machines due to failures
will shorten the production time, which at the end, affects the production plan
and delivery schedules. Therefore, late deliveries and additional production
expenses, e.g. due to penalties, should be expected.
3. Competitive price failures: Failures, e.g. in a bearing of a rotating machine, in
serially connected production process machines will lead to downtime and
production stoppage of the whole production line. Such failures increase the
non-added values time, which in turn increase the production cost. The latter will
be reflected at the end in the final product price. This will affect a company
achieving the production cost at a particular level in order to be able to offer their
customers a competitive price with a reasonable profit margin.
4. Society acceptance failures: Improper maintenance of, e.g. a diesel engine could
in some cases increase the pollution ejected to the environment, which will lead
to extra costs to adapt working conditions of the engine to the national or
international regulations.
5. Machine condition (Asset value) failures: Improper operation and maintenance
of machines will lead to increased deterioration and reduced machines reliabil-
ity, and consequently, continuously reduced value of the company’s assets.
3 Model Development
1. The impact of
each failure on the
CAs CA. n
High quality Competitive On time
price delivery
2.The economic
losses per each CA Due to bad Due to failures Due to Etc.
quality and stoppages delivery
delay
3. Prioritization of
failures
Prioritization of failures (using Pareto chart)
4. Needed
maintenance Assessment of the costs needed for the actions required to prevent failure reoccurrence
investment to
prevent
reoccurrence of
failures
time, income and profit margin in the whole production line. Also, this will cost
additional expenses to remove the oil to avoid harming soil. It may also cause
accidents and fires which expose staff and properties safety to high risk and
catastrophic consequences. These failures may lead to delay in the delivery of
production according to the delivery time schedule which in turn results to pay
penalties, losses of customer if it is repeated and consequently market shares.
Therefore, prioritizing failures due to their significance can be done through
summing the impacts of every failure on the CAs and applying Pareto diagram.
The highest prioritized failure is that which cause more damage and leads to highest
economic losses. Significant failures can be prioritized then and suggest the most
profitable maintenance actions through comparing the economic losses with the main-
tenance expenses. In this way, it will be possible to conduct cost-effective maintenance
through prioritizing the most profitable maintenance actions, see also Fig. 2.
In order to find the impact of each failure on each CA, reasonable estimation of
each failure impact on the losses per each CA were made utilizing the authors’
experience in consultation with the case company, see Table 2. Then a summation
of each failure impact on each CA was performed (see Table 2). There were only
two categories of losses connected to two CA i.e. “Bad quality” and “Losses in
profit margin”.
A Pareto diagram is then applied to determine the failures of the most economic
significance, i.e. those failures whose losses are profitable to avoid. Figure 3 shows
A Model for Increasing Effectiveness and Profitability of Maintenance. . . 9
the Pareto diagram application. Pareto analysis shows that only failures of “Raw
material quality”, “Bearing” and “Gear” cause about 78% of the total loss on the
CAs. Whereas adding the failure of “Cooling system” will increase the loss by 8%
only, therefore it was decided to prioritize only the first three mentioned failures.
The prioritized failures are then analysed to improve the related process or
technique in order to prevent the failures’ reoccurrence. Then, the needed invest-
ment is assessed. However, the needed investment should be compared with the
losses in order to make a cost effective decision, whether to neglect, postpone or
solve the problem. A target value should be estimated, before being able to make a
cost effective decision. The target value refers to the expected losses after applying
a particular solution—as a percentage of the total failure losses. Also, determining a
target value will be useful in order to enhance the continuous improvement and
activities follow-up.
When failures are prioritized according to their significance, the investments
needed are identified and target values are set for each CA. A cost-effective
decision can then be achieved, see Table 3.
In the next section, a discussion of the results and how CA-Failures support cost
effective decisions will be presented.
5 Discussions
After analysing the failure data using the model CA-Failures, it is shown that the
losses are mainly due to two categories i.e. “Bad quality“ and “Losses in profit
margin” and the losses were 28.7% and 71.3% of the total failure respectively (see
Table 1). Failures of “Gear”, “Raw materials quality” and “Bearing” cause most of
the failures losses, i.e. 78% in total (see Fig. 3 and Table 3). To prevent these losses,
10 B. Al-Najjar and H. Algabroun
Table 3 Analysis of the prioritized failures (Prioritized failures are “Gear”, “Raw material
quality” and “Bearing”)
Suggested solutions
for process
Loss per improvement or Investment Target value i.e. the
prioritized failure prevention needed for each expected losses after
failure as a (Suggested in solution (as a the investment (as a
percentage of consultation with by percentage of the percentage of the
Failures the total losses the company) total losses) total losses)
Gear 27 Apply CBM using 15 2
vibration
Maintenance staff 5
training
Raw 26 To find and establish 17 Unknown
material an analysis technique
quality for quality control of
the raw materials
Bearing 25 Apply CBM using 13 1
vibration
Training of operator to 2
increase awareness
about the bearing
condition
Maintenance staff 5
training
Shaft 3 – – –
Cooling 5 – – –
system
User 3 – – –
error
Cutting 7.7 – – –
tools
Pump 3.3 – – –
Total 100
investments are needed; by comparing the needed investments and the target value
a cost-effective decision could be made. In the given case study, failures of “Gear”
need two actions to prevent the failure reoccurrence, i.e. applying CBM using
vibration monitoring techniques and training of maintenance staff. The total invest-
ment needed for that is estimated to be equivalent to 20% of the total losses
(i.e.100%). After applying the suggested solution, the “Gear” failure losses are
expected to be reduced from 27% to only 2% (i.e. 25% of the total losses will be
saved). By this investment, the company will gain about 5% in the first year, see
Table 3.
In order to prevent “Bearing” failures, we reutilize CBM using vibration mon-
itoring technique and trained maintenance staff and operators. This requires addi-
tional 20% of the total loss to be invested. This solution is expected to reduce the
A Model for Increasing Effectiveness and Profitability of Maintenance. . . 11
“Bearing” failure losses from 25% to only 1% (i.e. 24% of the total losses will be
saved). By this investment, the company will gain about 4% in the first year, see
Table 3. However on the long run, additional gain is also expected.
Failures due to “Raw material quality” demand additional investment of around
17% of the total losses. Since the target value was unknown, therefore a decision
concerning the investment cost-effectiveness could not be made and therefore this
issue has been left to company until the estimation of the target value is done.
6 Conclusions
The model structurally analyses the failure database to identify and prioritize the
failures that affect the company the most. It supports the decision maker to
customize a maintenance plan that suits the company’s special situation. There
was no problem found in applying the failure data to the model, thus for future work
using the model in more case studies could be interesting.
It is important to note that in some cases, one solution could help in solving
multiple problems in different categories, for example, training of the operators (see
Table 3) could help to prevent failures that are not only related to bearing but also to
prevent other failures that are related to, for example, misuse. Using a method,
e.g. Maintenance Function Deployment in Al-Najjar [6], to determine and identify
the interaction and the values shared among different failure areas could be
interesting.
References
1. Al-Najjar B (2003) Modelling the design of the integration of quality, maintenance and
production. In: Proceeding of the 16th international congress COMADEM 2003, Växj€ o,
Sweden, pp 23–31
2. Waeyenbergh G, Pintelon L (2002) A framework for maintenance concept development. Int J
Prod Econ 77(3):299–313
3. Al-Najjar B (2007) The lack of maintenance and not maintenance which costs: a model to
describe and quantify the impact of vibration-based maintenance on company’s business. Int J
Prod Econ 107(1):260–273
4. Sherwin DJ, Al-Najjar B (1999) Practical models for condition monitoring inspection intervals
David. J Qual Maint Eng 5(3):203–221
5. Ahamed Mohideen PB, Ramachandran M, Narasimmalu RR (2011) Construction plant break-
down criticality analysis – part 1:UAE perspective. BIJ 18(4):472–289
6. Al-Najjar B (2011) Maintenance impact on company competitiveness & profit. In: Asset
management: the state of the art in Europe from a life cycle perspective. Springer Science
Business Media B.V., Dordrecht, pp 115–141
7. Al-Najjar B, Jacobsson M (2013) A computerised model to enhance the cost-effectiveness of
production and maintenance dynamic decisions. J Qual Maint Eng 19(2):114–127
8. Hoon Lee J, Chan S, Soon Jang J (2010) Process-oriented development of failure reporting,
analysis, and corrective action system. Int J Qual Stat Reliab 2010:1–8
12 B. Al-Najjar and H. Algabroun
9. Villacourt M, Drive M (1992) Failure reporting, analysis and corrective action system in the us
semiconductor manufacturing equipment industry: a continuous improvement process. In:
Thirteenth IEEE/CHMT international electronics manufacturing technology symposium.
IEEE, pp 111–115
10. Villacourt M, Govil P (1993) Failure reporting, analysis, and corrective action system. Marcel
Dekker, New York
Maintenance Process Improvement Model by
Integrating LSS and TPM for Service
Organisations
Abstract This paper presents an integrated model to provide guidance and support
for those organisations who aim to reach world-class standards in maintenance
processes through continual improvement. A strategic model has been developed
through conceptual integration of three popular process improvement strategies,
which are six sigma, total productive maintenance (TPM) and lean. Lean Six Sigma
can operate in parallel with the TPM strategy and will make it easier to understand
by shop floor operators. The application of the model has been demonstrated using a
case study in maintenance of a fleet of military vehicles. The proposed model is
very generic in nature and can be applied to any service organisations with
maintenance functions to achieve high process performance and overall equipment
effectiveness.
Keywords Lean Six Sigma • TPM Strategy • Integrating Lean Six Sigma and
TPM • Maintenance Process Improvement
List of Abbreviations
A Availability
CTQ Critical to quality characteristics
DMAIC Define, measure, analyse, improve and control
DPMO Defect per million opportunities
LSS Lean Six Sigma
MTBF Mean time between failures
OEE Overall equipment effectiveness
PE Performance rate
PM Preventive maintenance
Q Quality rate
SIOPC Supplier-input-process-output-customer
SMED Single minute exchange of dies
1 Introduction
2 Model Strategies
Lean Six Sigma combines lean methods and Six Sigma, using specific DMAIC
processes to provide companies with better speed and lower variability to increase
customer satisfaction [5]. The first phase in DMAIC process is to define project
objectives and customer needs. The second phase is to measure the current process
performance as well as quantifying the problems. The third phase is to analyse the
Maintenance Process Improvement Model by Integrating LSS and TPM for Service. . . 15
process and find the causes of problems, particularly the root causes. The fourth
phase is to improve the process, i.e. correcting the causes of defects and reducing
process variability. The final phase is to control the process and maintain the
improved performance. These five phases can assist Lean Six Sigma teams to
systematically and gradually develop process rationalisation, starting with defining
the problem and then introducing solutions targeted to the fundamental causes, so
constructing the optimal implementation method and ensuring the sustainability of
solutions [6]. This approach has gained increasing recognition in process improve-
ment practices.
2.2 TPM
This study has proposed an integrated approach of TPM with LSS to reach world
class maintenance performance, with the core model shown in Fig. 1. Lean Six
Sigma forms the basic foundation for the TPM strategy and makes it easier to
understand by shop floor operators who are the most important enablers of success-
ful TPM implementation. Within the five phases of DMAIC, various problems and
sub-processes of the maintenance department are defined, the process performance
is measured, the most important causes of the defects or non-conformities are
identified and analyzed, improvement or corrective actions are then taken with
the improvements sustained by standardisation and continuing process control.
Moreover, the iterative process of DMAIC is used as the main operational approach
for the implementation of this model in order to achieve continual improvement of
maintenance activities and ultimately to reach world class performance in terms of
both sigma level and overall equipment effectiveness. The implementation of this
model or approach will be supported with a rich collection of tools from Six Sigma,
lean, TPM, quality control and problem solving practices.
16 B. Alsubaie and Q. Yang
Table 1 Key activities and tools of implementing the TPM oriented LSS maintenance model
Stage Activities Tools
1. Define • Build process improvement team • SIOPC
• Identify problems & weaknesses of the process • Brainstorming
• Select CTQ characteristics • VOC
• Pareto analysis
2. Measure • Select measuring system • Process map
• Gather information about key maintenance processes • TPM
• Calculate the current OEE • OEE
3. Analyse • Identify root causes of problems • Cause and effect
• Implement basic levels of TPM diagram
• Identify improvement opportunities • TPM
4. Improve • Propose solutions and implement changes for main- • Seven Wastes
tenance improvement • SMED
• Evaluate the process performance • Poka—yoke
• Calculate the new OEE • 5S
• TPM
5. Control • Standardize the best practices • SPC
• Integrate the changes to the organisation knowledge • Performance
base management
• Continual improvement • Education and
training
3 Case study
In order to test the proposed TPM oriented LSS model for the vehicle maintenance,
a case study was performed for the engine maintenance process, since engine for a
vehicle is as vital as the heart of a human being and its maintenance is essential. As
demands for service quality and cost reduction in vehicle maintenance have both
increased in recent years, the effectiveness of a maintenance system for engines has
become an important issue. Engines are subject to deteriorations in relation to both
usage and ageing, which leads to reduced product quality and increased mainte-
nance costs. The maintenance division executes preventive maintenance (PM) on
engines to prevent or slow down such deteriorations.
3.1 Define
Step_D1: The project started with Define phase that gives a clear problem definition
using the supplier-input-process-output-customer (SIPOC) tool. This tool describes
the step-by-step process for the engine maintenance as shown in Fig. 2. The first
process is the engine service or maintenance. The input to this process includes the
engine to be serviced, parts and preventive maintenance program and procedure,
whilst the supplier is the maintenance crew. The output of this process is engine
serviced, the customer is the field service unit. The second process is repair and
replacement of engine. The inputs to this process are operation notification and
work order, the supplier is the field service unit. The output of this process is engine
repaired or replaced and the customer is the field service unit.
Step_D2: The engine preventive maintenance (PM) being analysed is verified to
be significant by the field study. Engine PM cost represents a high percentage of
vehicle PM cost. The team members participate in brainstorming sessions to
identify critical to quality characteristics (CTQ) based on the voices of customer
input. Also, the component(s) failure that results in high machine downtime or cost
(due to machine breakdown) is classified as critical components. Critical engines
failures have been reported for the engines in the field study which causes
significant cost of the PM and also deviations from the customer satisfaction
targets. The project was scoped down to oil and water leakage since they contribute
to about 60% of the total failure cost as determined through the use of Pareto
analysis, as shown in Fig. 3.
3.2 Measurement
Step_M1: To measure the factors that contribute to the process and failures on the
subject engine, a number of tools from the Six Sigma toolbox are used such as
process mapping and fishbone diagram. The process map (Fig. 4) provides a visual
view of all maintenance and operation steps that take place from the time an engine
failure is detected through putting it back to service all the way to operation and
monitoring until it fails again.
Step_M2: Since the CTQ characteristics, i.e. oil and coolant leakage, are iden-
tified in the Define phase, a data collection plan needs to be developed. The
measurement system should be examined prior to data collection. In this case, the
existing service report is used to facilitate the collection of primary data. Monthly
reporting is particularly useful in monitoring the maintenance tasks performed by
the maintenance personnel and calculating the maintenance cost. Also, each vehicle
has its own maintenance history book to record the repairs/replacement done to
it. From these records, the data on the maintenance history of the engines can be
extracted. To quantify the problem, data gathering was initiated on the failures costs
of engines.
Step_M3: For a specific CTQ characteristic, the sigma level can be calculated
from DPMO (defect per million opportunities) as:
Maintenance Process Improvement Model by Integrating LSS and TPM for Service. . . 19
The process capability indices Cpk and the corresponding sigma levels are
summarised in Table 2. The sigma level of a process can be used to express its
capability as to how well it performs with respect to specifications.
3.3 Analysis
Step_A1: To ascertain the root cause(s) of key engine failures, an analysis using the
cause-and-effect diagram is therefore carried out during a brainstorming session of
the LSS team. Figure 5 shows the root causes of the engine failure problems.
20 B. Alsubaie and Q. Yang
3.4 Improvement
Table 4 Implementation of 5S
5S Before After
Sort • Rejected parts are kept inside the • The parts are removed and the space is
workshop. freed.
Set in • Earlier patches on the floor dis- • Patches are filled with cement thus helping
Order turb material movement using smooth material flow.
trolley. • Tools are stored in their respective places
• Tools are placed randomly in identified with labelling.
racks and no labelling is done.
Shine • Work place not very tidy and • Clean and tidy work place.
clean.
Standardize • No operator report is kept. • Writing hourly report is compulsory.
• Operator details are not displayed • Operator details are displayed on the
on the notice board. notice board.
Sustain • Organisation mission and vision state-
ments are displayed in Arabic as well as
English.
• Suggestion scheme stating that whoever
gives the best suggestion will receive a
prize.
• Drivers and maintenance staff will learn how to identify abnormalities as such
during their daily and periodic inspection activities.
(e) Planned Maintenance: It is aimed to have trouble free vehicles without any
breakdown and ensure components at good quality level giving total customer
satisfaction.
(f) OEE is calculated after the implementation. Based on the initial assessment, the
availability has increased to 92%, the performance rate 87% and the quality rate
98.5%, with an overall OEE of 79%. Whilst this is still below the world class
85% performance, it has significantly improved the initial OEE of 66%
(Table 5). Continual improvement is required to reach the world class
performance.
3.5 Control
A new model based on TPM and Lean Six Sigma has been presented to provide
guidance and support for service organisations who aim to reach world-class
standards in maintenance processes through continual improvement. The applica-
tion of the model has been illustrated using a case study in maintenance of a fleet of
military vehicles. The above discussions are largely based on the case study,
particularly the engine repair process due to the water and oil leakage problems.
However, this approach is very generic in nature and can be applied to any other
maintenance or repair process (e.g. engine repair due to heavy friction), and to any
service organisations with maintenance functions. Of course, the complexity of the
proposed model will depend on the application since the nature and the number of
CTQs are application specific. The proposed model as the framework together with
the use of common tools emphasize the process approach and will therefore be
generally applicable in such service organisations. Use of this model will likely
help to achieve high process performance and overall equipment effectiveness. The
model also provides a good framework and methodology to continually improve
the maintenance performance.
References
1. Cassady CR, Murdock WP, Nachlas J, Pohl E (1998) Comprehensive fleet maintenance
management. In: IEEE international conference on systems, man, and cybernetics, vol
5. IEEE, New York, pp 4665–4669
2. Aldairi J, Khan M, Munive-Hernandez J (2015) A conceptual model for a hybrid knowledge-
based Lean Six Sigma maintenance system for sustainable buildings. In: Proceedings of the
World Congress on Engineering, vol 2
3. Milana M, Khan MK, Munive JE (2014) A framework of knowledge based system for
integrated maintenance strategy and operation. In: Applied mechanics and materials. Trans
Tech Publications, Zurich, pp 619–624
4. Uday K et al (2009) The maintenance management framework: a practical view to maintenance
management. J Qual Maint Eng 15(2):167–178
5. Wang H, Pham H (2006) Reliability and optimal maintenance. Springer, Berlin
6. Cheng C, Chang P (2012) Implementation of the Lean Six Sigma framework in non-profit
organisations: a case study. Total Qual Manag Bus Excell 23(3–4):431–447
7. Nakajima S (1989) TPM development program: implementing total productive maintenance.
Productivity Press, Cambridge
8. Chaneski W (2002) Total productive maintenance–an effective technique. Mod Mach Shop 75
(2):46–48
9. Ahuja IPS, Khamba JS (2008) Total productive maintenance: literature review and directions.
Int J Qual Reliab Manage 25(7):709–756
Group Replacement Policies for Repairable
N-Component Parallel Systems
Abstract For a multiple component parallel system, the operation of each compo-
nent is independent, that is, any failure of component will not affect the normal
operation of the system. Due to the inevitable deterioration of the component, it may
fail more frequently when its age increases. In order to reduce the number of failures,
the component should be replaced at a certain age. When a replacement action of the
component is performed, it will incur a replacement cost and a setup cost. Under this
situation, grouping replacement for all components is an important policy in order to
reduce setup cost. This paper derives a formula of the grouping replacement for a n-
component parallel system. Further, an optimal grouping replacement method is
offered, and the optimal replacement time of each group is obtained.
1 Introduction
With advances in technology, customer demand for products have also increased,
making the production system relatively complex. In order to maintain the stable
operation of the system, maintenance policies should be planned and performed,
especially when a system consists of multiple components. In general, multiple
components in a system may be connected in series or parallel. When multiple
components are connected in series, any failure of either one component will cause
system downtime. However, when multiple components are connected in parallel,
the failure of any component will not affect the operation of the other component.
In practice, when a component fails during the operation period, the failed com-
ponent may be rectified by minimal repair, imperfect repair, perfect repair, or
replacement. Due to the inevitable deterioration of the component, it may fail more
frequently when its age increases. In order to reduce the number of failures, the
component should be replaced by a new component at a certain age. In this paper,
two maintenance policies for components are considered: (1) When the component
fails during the operating period, the failed component is corrected by minimal repair
and (2) The component is replaced at a pre-specified time under normal operation.
For a system that consists of multiple components in parallel, all components or
some components may be treated as one group and an optimal replacement policy for
the group can be derived. Under this situation, it is called group replacement policy.
On the other hand, we may focus on each component and find the optimal replacement
policy for each component. It is called individual replacement policy. However, either
way may not be optimal when the maintenance actions of the components are cost-
dependent. For example, minimal repair may be carried out easily, but replacements
should be performed by a professional technician team. When a replacement action is
performed, there is a fix setup cost, including the downtime cost and the cost of hiring
a professional team to install the components. Furthermore, the setup cost may be the
same for replacing a single component or a group of components. In this paper, we
focus on investigating the influences of setup cost on the grouping replacement policy
for a parallel system, which consists of multiple components in parallel. We derive the
grouping models of multiple components and analyze the influences of setup cost for
the number of grouping of multiple components and the replacement time of each
group through numerical examples.
In the past, many research works have been published on the repair and replace-
ment policies of a single component or system. In general, the maintenance policies
for repairable products can be divided into two main categories: repair and replace-
ment. About repair policy, Barlow and Hunter [1] proposed the minimal repair
concept and applied it in the reliability area. Minimal repair refers to a failed device
is restored to its normal operation, and its failure rate remains the same as that just
prior to failure. Nakagawa and Kowada [2] showed that the failure process of a
system during the operating period follows a Non-homogeneous Poisson Process
(NHPP) when the minimal repair is carried out. Chien et al. [3] considered a
preventive replacement model with minimal repair and a cumulative repair-cost
limit by introducing the random lead time for replacement delivery.
About replacement policy, Berg and Epstein [4] gave a rule for choosing the
least costly of the above three policies (age replacement policy, block replacement
policy and failure replacement policy) under conditions specified. Beichelt [5]
introduced a generalized block replacement policy and derived the long-run cost
rate and integral equations of the renewal type for the basic reliability expressions.
Finally, a numerical example is illustrated. Chen and Feldman [6] studied a modi-
fied minimal repair/replacement problem that is formulated as a Markov decision
process and assumed that the operating cost is an increasing function of the age of
the system. Further, a computational algorithm for the optimal policy is suggested
based on the total expected discounted cost. Sheu and Griffith [7] considered a
generalized age-replacement policy with age-dependent minimal repair and ran-
dom leadtime and developed a model for the average cost per unit time. Further,
determination of the minimum-cost policy time is described and illustrated with a
Group Replacement Policies for Repairable N-Component Parallel Systems 27
numerical example. Chien and Sheu [8] supposed a system can have two types of
failure: type I failure (minor) or type II failure (catastrophic) and proposed a
generalization of the age replacement policy for such a system. Finally, some
numerical examples are illustrated and analyzed.
About grouping replacement policy, Shafiee and Finkelstein [9] investigated an
optimal age-based group maintenance policy for a multi-unit series system where
the components are subjected to gradual degradation. Zhao et al. [10] proposed
several approximate models for optimal replacement, maintenance, and inspection
policies. However, most of the literature have not considered the case when the
components are cost-dependent in performing the maintenance actions. This paper
is organized as follows. The grouping models of multiple components are
constructed and the optimal replacement time of each group is obtained in Sect.
2. The Weibull case is illustrated in Sect. 3. In Sect. 4, the influences of setup cost
on the number of grouping of multiple components are illustrated through numer-
ical examples. Finally, some conclusions are drawn in Sect. 5.
2 Mathematical Models
Fig. 1 n-component
parallel system
the failure rate function hi(t) is an increasing function of t. Since a failed component
is rectified by minimal repairs, the failure process of the component within the
operating period is a nonhomogeneous Poisson process with the failure rate function
hi(t). Under this repair policy, the expected repair
cost rate of system within group
Pk P
replacement time T gj is j¼1 C H
i2gj mi i T gj =T gj , for 1 k n.
Xk X X k
Cs
E TC T g1 ; T g2 ; . . . ; T gk ¼ Ai T g j þ ð1Þ
j¼1 i2g
T
j¼1 gj
j
C H T þC
where Ai T gj ¼ 2 T gj ¼ 0.
mi i gj ri
Tj if component i g j; otherwise, A i
The objective is to find the optimal group replacement time Tgj in Eq. (1).
The optimal group replacement time is derive in Sect. 2.2.
From Eq. (1), the excepted average cost rate of the jth group for j ¼ 1, 2,. . ., k and
1 k n group can be obtain as follows.
Ph i
Cmi H i T gj þ Cri þ Cs
h i i2g
E TC T gj ¼ ð2Þ
j
T gj
To find the optimal replacement time T gj , we can take the first derivative of
Eq. (2) with respect to T gj and then the result is obtained as follows.
!
h i Pn h i o
Cmi T gj hi T gj H i T gj Cri Cs
dE TC T gj i2gj
¼ ð3Þ
dT gj T 2gj
Xn h i o
Cmi T gj hi T gj H i T gj Cri ¼ Cs ð4Þ
i2gj
Substituting T ∗
gj into Eq. (4), Eq. (4) can be expressed as
Xh i X
Cmi H i T ∗
gj þ Cri þ Cs ¼ C ∗ ∗
mi gj i T gj
T h ð5Þ
i2gj i2gj
Substituting Eq. (5) into Eq. (2), the expected average cost rate of the jth group
can be rewritten as
P
Cmi T ∗ ∗
g j hi T g j
h i i2g
E TC T ∗ ¼ ð6Þ
j
gj
T∗
gj
From Eq. (6), the expected average cost rate of system can be obtained as
follows.
0P 1
Cmi T ∗ h T∗
h i Xk B
i2gj
g j
i gj
C
∗ ∗ ∗
E TC T g1 ; T g2 ; . . . ; T gk ¼ B C ð7Þ
@ T ∗ A
j¼1 g j
The Weibull case is illustrated for Eq. (5) in the next section.
3 Weibull Case
2 ! 3β1
P
6 Cri þ Cs 7
6 i2gj 7
T∗
gj ¼6 7
6 ð β 1Þ P C α β 7 ð9Þ
4 mi i 5
i2gj
Based on Eq. (9), when n components are divided into 1 group (i.e., k ¼ 1),
the optimal group replacement time T ∗
g1 is
2
P n
3β1
C þ Cs
6 i¼1 ri 7
T∗
g1 ¼6
4 P
n
7
5 ð10Þ
ðβ 1Þ Cmi αiβ
i¼1
When n components are divided into n groups (i.e., k ¼ n), the optimal group
replacement time T ∗
gj for j ¼ 1, 2, . . ., n is
" #β1
Crj þ Cs
T∗
gj ¼ ð11Þ
ðβ 1ÞCmj αjβ
The performance of the optimal replacement policy is evaluated and the prop-
erties of grouping replacement policy are illustrated through numerical examples in
chapter 4 “Low Speed Bearing Condition Monitoring: A Case Study”.
4 Numerical Examples
In this section, the performances of the optimal group replacement policies are
evaluated through numerical examples of Weibull cases. Suppose that 5 compo-
nents are divided into k (1 k 5) groups. The parameter values are considered for
5 components and components are arranged according to mean lifetime μi ¼ (1/αi)Γ
(1 + 1/βi) in Table 3.
For all possible grouping methods of 5 components are showed in Table 4. For
example, 5 components are divided into k ¼ 2 groups. The number of components
of each group is 3 and 2 or 4 and 1. When the number of components of each group
is 4 and 1, the optimal grouping method of components is {P1, P2, P3, P4} and {P5}.
The optimal replacement time of {P1, P2, P3, P4} and {P5} are 9.8 and 56.1, and the
32 C.-W. Chiu et al.
expected total average cost rate of the system is 1002.1. From Table 4, for all
possible grouping methods of 5 components, the optimal grouping method is
2 groups. The number of components of each group is 3 and 2, and the components
of each group is {P1, P2, P3} and {P4, P5}. The optimal replacement time of {P1,
P2, P3} and {P4, P5} are 8.7 and 28.9, respectively and the expected total average
cost rate of the system is 1001.4.
Table 5 shows the optimal replacement time for 1 group and 2 groups under
various setup cost Cs. Observing Table 5, when 800 Cs 1100, the optimal
Group Replacement Policies for Repairable N-Component Parallel Systems 33
1100.00
reduce percentage
1050.00
1000.00 individual
optimal solution k=5,(1,1,1,1,1)
optimal solution k=4,(2,1,1,1)
E[TC(Tg)]
950.00 549.5 optimal solution k=3,(2,2,1)
optimal solution k=3,(3,1,1)
355.5 optimal solution k=2,(3,2)
900.00
106
850.00
28.3
800.00
0 100 200 300 400 500 600 700 800 900
Group Replacement Policies for Repairable N-Component Parallel Systems
5 Conclusions
This paper investigates the group replacement policies for repairable n-component
parallel systems and derives the grouping models when n components are divided
into k (1 k n) groups. Further, the optimal group replacement time of each
group is obtained. Especially, there are closed-form solutions when the lifetime
distributions of components are Weibull. In addition, the influence of the setup cost
on the optimal replacement policy is analyzed by numerical examples. It is
observed that when the setup cost is relatively large, a lower number of groupings
of components should be adopted and performed to minimize the expected total
average cost rate of the system.
References
1. Barlow RE, Hunter LC (1960) Optimum preventive maintenance policies. Oper Res 8:90–100
2. Nakagawa T, Kowada M (1983) Analysis of a system with minimal repair and its application to
replavcement policy. Eur J Oper Res 12:176–182
3. Chien YH, Chang CC, Sheu SH (2009) Optimal periodical time for preventive replacement
based on a cumulative repair-cost limit and random lead time. Proc Inst Mech Eng O J Risk Reliab
223:333–345
4. Berg M, Epstein B (1978) Comparison of age, block and failure replacement policies.
IEEE Trans Reliab 27:25–29
5. Beichelt F (1981) A generalized block-replacement policy. IEEE Trans Reliab 30:171–172
6. Chen M, Feldman RM (1997) Optimal replacement policies with minimal repair and
age-dependent costs. Eur J Oper Res 98:75–84
7. Sheu SH, Griffith WS (2001) Optimal age-replacement policy with age-dependent minimal-
repair and random-lead time. IEEE Trans Reliab 50:302–309
8. Chien YH, Sheu SH (2006) Extended optimal age-replacement policy with minimal repair of a
system subject to shocks. Eur J Oper Res 174:169–181
9. Shafiee M, Finkelstein M (2015) An optimal age-based group maintenance policy for multi-
unit degrading systems. Reliab Eng Syst Saf 59:374–382
10. Zhao X, Al-Khalifa NK, Nakagawa T (2015) Approximate methods for optimal replacement,
maintenance, and inspection policies. Reliab Eng Syst Saf 144:68–73
Low Speed Bearing Condition Monitoring:
A Case Study
Abstract The health condition of worm-wheel gearboxes is critical for the reliable
and continuous operation of passenger escalators. Vibration sessors have been
widely installed in such gearboxes and vibration levels are usually utilized as health
indicators. However, the measurement of vibration levels is not robust in slow
speed bearing condition monitoring. In this paper, the health condition of two slow
speed bearings were evaluated using vibration data collected from sensors installed
in the shaft of a worm wheel gearbox. It has been shown that the vibration level fails
to indicate the bearing health condition. The assessment accuracy can be improved
by combining several simple methods.
1 Introduction
The London Underground (LU) is one of the world’s oldest and busiest metro
systems carrying in excess of one billion passengers each year. The system serves
270 stations and a total 11 train lines. In order to increase accessibility and cope
with high passenger volumes, LU manages 430 escalators within its stations for
expeditious commuter movement during peak periods of travel. These escalators
cope with heavier loading, which is up to 13,000 passengers an hour and run more
than 20 hours a day. Escalator availability is essential for the prompt transport of
passengers to and from platforms. Therefore, it is tracked as a Key Performance
Indicator (KPI) in LU performance reports released by the Mayor of London.
Significant efforts have been made by LU to maintain the network escalator
availability as high as above 95%.
LU escalators are driven by electric motors. Power from the motor is transmitted to
the drive gear, which drives the escalator steps along with the chain, predominantly via
a worm-wheel gearbox. Worm-wheel gearbox generates high friction compared to
other gear types due to the entirely sliding action (as opposed to rolling) between
worm and gear. Gearbox wear occurs during normal operation. Manufacturing toler-
ances, inappropriate installation, poor maintenance and abnormal environmental con-
ditions can accelerate gearbox wear [1, 2]. Hence, condition monitoring, regular
inspection and maintenance are essential in order to obtain maximum usage and
prevent from premature failure.
Preventive maintenance is the predominant mode of LU escalators service plan.
Worm wheel replacement is carried out when wheel pitting damage is observed by
maintenance crews. The assessment of the allowable pitting damage is dependent
on crew experience and hence is subjective; without the establishment of consistent
criteria for the amount of pitting that will render worm wheel change necessary.
Premature replacements normally result in the wastage of otherwise useful worm
wheel hours and additional labour costs. There has been a recent shift towards
condition-based maintenance for greater effectiveness in manpower utilisation and
machine-hour capitalization [3, 4]. Condition monitoring is the process of machine
health assessment while it is in operation. It not only can prevent an unscheduled
work stoppage and expensive repair in the event of catastrophic failures but can also
optimise machine performance, provide effective plans for scheduled maintenance
manpower, and suggest the advance procurement of machine spare parts that need
to be replaced to bring the machine back to health.
Condition monitoring through the use of vibration analysis is an established and
effective technique for detecting the loss of mechanical integrity of a wide range and
classification of rotating machinery. Equipment rotating at low rotational speeds
presents an increased difficulty to the diagnostician, since conventional vibration
measuring equipment is not capable of measuring the fundamental frequency of
operation. Also, component distress at low operational speeds does not necessarily
show an obvious change in vibration signature. Further details can be found in
[5–8]. Furthermore, in most of gears types, defects manifest as periodic impacts in
the form of side-bands around the gear mesh frequencies [9]. However, such distinc-
tive defect symptoms are not obvious for worm-wheel gearboxes due to continuous
sliding interactions [10]. Therefore, diagnostics of worm-wheel gearbox defects with
vibration analysis is challenging. As a case study, vibration data of bearings from two
stations was utilized to assess vibration based condition monitoring method. The
current condition monitoring system provides a vibration level threshold to trigger
an alarm before failure. However, this vibration level based method is not a robust
indicator of bearing heath condition. More advanced signal processing methods, such
as spectral kurtosis kurtogram, envelope analysis and FM4* [11], or/and comprehen-
sive assessments (e.g. acoustic emission [12]) are essential to improve accuracy.
Low Speed Bearing Condition Monitoring: A Case Study 41
A main drive shaft of one of the LU escalators was selected to conduct condition
monitoring. The shaft has two support bearings (SKF 23026 CCK/W33), spherical
roller bearings, cylindrical and tapered bore, located at each end. The main drive
shaft speed operates at 10 rpm. The calculated defect frequencies for the inner ring,
outer ring and rolling element are 2.3, 1.87 and 1.58 Hz, respectively. The operating
condition of the bearings was monitored by measuring vibration levels. Two vibra-
tion sensors (PRÜFTECHNIK VIB 6.127) were located at each end of the shaft, as
shown in Fig. 1.
The velocity overall trends of Station A top shaft left and right side are shown in
Fig. 2a, b, respectively. A threshold value of 2 mm/s is noted on the figures. In
Fig. 2a, the threshold was not exceeded at the shaft left side. However, on the right
side the first observed data exceeding the threshold was noted in Oct. 2014.
Thereafter several peaks above the threshold appeared in Jan. and Feb. 2015, as
shown in Fig. 2b. After Feb. 2015, only a few peaks were observed before the
bearing fault condition was identified in May 2015.
Similar patterns were shown in acceleration overall trends of Station A shaft left
and right side in Fig. 3. It is known that damaged slow speed bearings will operate
continuously irrespective of the level of damaged within the bearing. In some
instances such damaged slow speed bearings have been known to completely grind
away components within the bearing (rollers, cage) and for the shaft to eventually be
supported by the bearing housing, leading to grinding of the shaft itself [13–15]. This
implies that there are situations when the vibration of a damaged bearing will not
increase with time due to increasing severity, on the contrary, the vibration may
decrease. This is one of the reasons why other technologies have been assessed to
give better indications of slow speed bearing condition [16, 17].
Fig. 2 Velocity overall trends of Station A top shaft. (a) Left side vibration sensor; (b) Right side
vibration sensor
Vibration waveform data, typically acquired prior to post processing, from a similar
machine at Station B top shaft left and right side was provided by LU. Figure 4
shows the vibration signal of top-shaft left in Mar. 2013. A total of 307,200 points
are plotted in Fig. 4a. Zoomed data plots of 100,000 to 110,000, and 100,000 to
101,000 data points are shown in Fig. 4b, c, respectively. Figure 4 highlights what
can only be described as transient events superimposed on underling vibration data.
However closer observation of the data reveals the transient impact type events are
Low Speed Bearing Condition Monitoring: A Case Study 43
Fig. 3 Acceleration overall trends of Station A top shaft. (a) Left side vibration sensor; (b) Right
side vibration sensor
not associated with electronic noise but are vibration type responses from the
bearing. Such transient vibration events can be indicative of impending failure.
A time-frequency analysis and Fourier transform were undertaken on the vibra-
tion data. Figure 5 highlights that the transient vibration event contained frequen-
cies of up to 100 Hz whilst Fig. 6 shows the spread of energy across 0–50 Hz, with
the strongest energy concentration at 25 Hz. A FFT of the vibration data (data in
Fig. 4a) is shown in Fig. 7. The shaft rotational speed is identified (0.17 Hz or
10 rpm) as well as several other peaks (0.43 Hz, 0.88 Hz and several multiples of
0.88 Hz). As the authors of this paper do not have information of the exact machine
configuration at Station B, the sources of these vibration peaks cannot be identified.
However, it does suggest that such low frequencies can be measured with the
currently employed sensors.
44 F. Duan et al.
-20
-30
-40
Magnitude (dB)
-50
-60
-70
-80
0 2 4 6 8 10
Frequency (Hz)
0
-5 -4 -3 -2 -1 0 1 2 3 4 5
Velocity (mm/s)
Fig. 8 Top shaft left vibration level distribution. Number of point in the logarithmic scale
6
Mar.
Apr.
5
May
Jun.
4 Jul.
log (Number of Points)
Aug.
0
-5 -4 -3 -2 -1 0 1 2 3 4 5
Velocity (mm/s)
Fig. 9 Top shaft left vibration signal from March to August 2013. Vibration level distribution in
the logarithmic scale
6
Mar.
Apr.
5
May
Jun.
log (Number of Points)
4 Jul.
Aug.
0
-5 -4 -3 -2 -1 0 1 2 3 4 5
Velocity (mm/s)
Fig. 10 Top shaft right vibration signal from March to August 2013. Vibration level distribution
in the logarithmic scale
3 Conclusions
References
1. Elforjani M, Mba D (2008) Monitoring the onset and propagation of natural degradation process
in a slow speed rolling element bearing with acoustic Emissions. J Vib Acoust, Trans ASME
130:041013
2. Elforjani M, Charnley B, Mba D (2010) Observations of a naturally degrading slow speed shaft.
Nondestruct Test Eval 25(4):267–278
3. Mba D, Jamaludin N (2002a) Monitoring extremely slow rolling element bearings: part-I.
NDT & E Int 35(6):349–358
4. Mba D, Jamaludin N (2002b) Monitoring extremely slow rolling element bearings: part II.
NDT & E Int 35(6):359–366
5. Berry JE (1992) Required vibration analysis techniques and instrumentation on low speed
machines (particularly 30 to 300 rpm machinery). In: Advanced vibration diagnostic and
reduction techniques. Technical Associates of Charlotte Inc, Charlotte
6. Canada RG, Robinson JC (1995) Vibration measurements on slow speed machinery. In: Pro-
ceedings of National Conference on Predictive Maintenance Technology (P/PM Technology),
vol 8, no 6. Indianapolis, Indiana, pp 33–37
7. Kuboyama K (1997) Development of low speed bearing diagnosis technique. NKK/Fukuyama
Works, Fukuyama
8. Robinson JC, Canada RG, Piety RG (1996) Vibration monitoring on slow speed machinery:
new methodologies covering machinery from 0.5 to 600 rpm. In: Proceedings of the
Fifth International Conference on Profitable Condition Monitoring-Fluids and Machinery Per-
formance Monitoring, BHR Group Publication (BHR Group, Cranfield), vol 22, pp 169–182
9. Alan D (1998) Handbook of the condition monitoring techniques and methodology.
Chapman and Hall, London
10. Vahaoja P, Lahdelma S, Leinonen J (2006) On the condition monitoring of worm gears.
Springer, London, pp 332–343
11. Elasha F, Ruiz-Cárcel C, Mba D, Kiat G, Nze I, Yebra G (2014) Pitting detection in
worm gearboxes with vibration analysis. Eng Fail Anal 42:366–376
12. Elforjani M, Mba D, Muhammad A, Sire A (2012) Condition monitoring of worm gears.
Appl Acoust 73(8):859–863
13. Jamaludin N, Mba D, Bannister RH (2001) Condition monitoring of slow-speed rolling element
bearings using stress waves. J Process Mech Eng 215(E4):245–271
14. Jamaludin N, Mba D, Bannister RH (2002) Monitoring the lubricant condition in a low-speed
rolling element bearing using high frequency stress waves. J Process Mech Eng 216(E):73–88
15. Mba D, Bannister RH, Findlay GE (1999a) Condition monitoring of low-speed rotating machi-
nery using Stress Waves: part I. Pro Inst Mech Eng 213(E):153–185
16. Mba D (2002) Applicability of acoustic emissions to monitoring the mechanical integrity of
bolted structures in low speed rotating machinery: case study. NDT & E Int 35(5):293–300
17. Mba D, Bannister RH, Findlay GE (1999b) Condition monitoring of low-speed rotating
machinery using Stress Waves: Part II. Pro Inst Mech Eng 213:153–185
Program Control-Flow Structural Integrity
Checking Based Soft Error Detection Method
for DSP
Yangming Guo, Hao Wu, Guochang Zhou, Shan Liu, Jiaqi Zhang,
and Xiangtao Wang
Abstract Nowadays, for the equipment in the outer space, the system safety and
reliability are largely affected by soft errors which are caused by the single high energy
particles. This is frequently reported in DSP and the other memory devices. Thus, the
detection of soft error becomes an interesting research topic. For the purpose of
detecting the occurrence of soft error occurred in storage areas of a DSP program, a
control flow integrity based checking scheme for soft error detection is presented in
this work. In this work, the DSP program implemented in assembly language is mainly
focused. Firstly, the program is divided into a number of basic blocks with
corresponding structure information being stored in a partition table. Then, for each
basic block, a checkpoint is set at the end. The program control flow error can be easily
determined by examining the consistency between the information at runtime and that
recorded information in the partition table. Compared with the signature-based
method, the proposed method is able to achieve almost 100% of error detection
coverage. Furthermore, the proposed detection scheme has better cross platform
portability under almost same detection efficiency and detection overhead.
effect (SEE) in DSP. Consequently the transient or intermittent errors incurred by SEE
are referred to as soft errors. Due to the existence of such errors, the reliability and
security of the system is severely affected [1]. Hence, investigation of the effects of soft
error has drawn the attention of numerous researchers and engineers. Public informa-
tion indicates that more than 70% of soft errors occurring in DSP are likely to incur
control flow errors to the software [1]. Control flow errors may cause the corresponding
program to deviate from the correct execution flow; this is likely to result in incorrect
outputs or even complete breakdown of the system [2]. Therefore, control-flow
checking on the program implemented in the DSP is able to effectively detect whether
soft errors occurred or not, especially in the memory blocks of the program.
The main idea of the existing control-flow error detection approach is signature
based monitoring. To be specific, the program is divided accordingly into several
basic blocks with assigned static signatures. These signatures are then compared
with run-time signatures, which are calculated dynamically with the execution of
the program. The signature based monitoring approach can be implemented by
either hardware or software. The hardware implementation is implemented through
the utilization of a watchdog, aiming to detect the errors that are occurred during the
executing process of the program in real time. The main features of the approach
using a watchdog are high speed and small overhead required. However, this
inevitably introduces extra hardware, which increases the necessary overhead;
this definitely increases the difficulty of the system development.
As the development of human society, there is an increasing demand of devel-
oping embedded system with high performance, low cost and short development
cycle. Due to the reason that the developing process can be greatly accelerated with
the application of Commercial-Off-The-Shelf (COTS), this technique has been
increasingly and widely applied in the development of embedded system. However,
due to the unchangeable property of manufactured COTS architecture, researchers
have to figure out a new technique for the purpose of detecting control-flow errors
on the top of watchdog. Aiming at the limitation of watchdog, a variety of software-
based detection schemes have been proposed for control-flow errors, such as
Enhanced Control-flow Checking using Assertions (ECCA) [3], Control-Flow
Checking by Software Signatures (CFCSS) [4], Enhanced Control-Flow Checking
(ECFC) [5], Edge Control-Flow Checking (ECFC) and Region Based Control-Flow
Checking (RBCFC) [6], and etc. Software-based detection schemes are indepen-
dent on the underlying hardware at the expense of performance to some extent,
which greatly enhances the reliability of COTS-based system. Although the signa-
ture based monitoring approach implemented by software also depends on no
hardware, it still is highly platform-dependent. Due to different instruction set for
different platforms such as ARM, MIPS and x86, the techniques of signature
generation and monitoring for one platform cannot be directly applied to another.
Thus the portability of the software system is greatly limited.
To overcome the shortcomings mentioned above, this paper presents a control-
flow error detection method based on structural integrity checking (SIC). Whether
the control-flow errors occur in the program is determined by checking the execu-
tion integrity of each basic block. The most desirable advantage of the proposed
Program Control-Flow Structural Integrity Checking Based Soft Error. . . 51
scheme is portability, that is, only minor modifications are necessary for adapting
the method for one platform to another. In terms of overhead required for detection,
the proposed technique and signature-based control-flow detection scheme are at
the same order of magnitude. Furthermore, the proposed method based on SIC is
able to reach nearly 100% of error detection coverage.
2 Principle of SIC
Aiming to monitor the state of the DSP in operation, the failure modes and
characteristics of the soft errors incurred by SEE should be analysed first. Then
corresponding program for the DSP is reasonably divided into several modules,
between which the checkpoints can be set.
For the program of a DSP inside a basic block, the instructions are executed
se-quentially. If no control-flow error occurs, the program starts to execute at the
entry port of the block and then jump out at the exit port. Therefore, the instruction
branch error is the main reason for the occurrence of the control-flow error. After
the basic block partition process is completed, all the branch instructions are located
at the end of the basic blocks and all non-branching instructions are inside the basic
blocks. Thus the incorrect execution of a branch instruction is likely to cause the
control flow being transferred to an incorrect address; thus an anticipated control-
flow transition that should have happened does not occur or vice versa.
In addition, soft errors occurring in a DSP may cause data bits in memory cells to
be flipped. If the flip happens inside the storage area for the program, the instruction
code may be changed. It is indicated by the public document [7] and practical analysis
that, the unexpected change in program storage area may cause a non-branching
instruction to be transformed into a branch instruction, or vice versa. Hence, if a
non-branching instruction is converted to a branch instruction, the control-flow error
will definitely occur in the program. Due to the fact that, all non-branching instruc-
tions are inside the basic blocks (including the first instruction of the basic block).
Then the control-flow error incurred by the unexpected conversion is reflected as the
incorrect control-flow shift from the basic block to other blocks.
According to previous analysis, the control-flow errors are being classified into
two categories. The first one is that, the control flow is not transferred where there
should be an anticipated transfer, usually incurred by the incorrect transfer address.
The other one is referred to the scenario that, the control flow is transferred
unexpectedly where there should not be any transfer.
In Fig. 1, five basic blocks are presented for an illustration; here, the rectangles
represent corresponding basic blocks being divided. Here, the black and red arrows
stand for correct and incorrect control-flow transfers respectively. A and B indicate
the type of control-flow errors incurred by the conversion from a non-branching
52 Y. Guo et al.
%DVLF *
' %DVLF
EORFN %DVLF EORFN
EORFN (
&
*
%DVLF )
&
EORFN %DVLF
EORFN
$
)
The delay slot is usually referred to as one or several instructions after the branch
instruction. The instructions in the delay slot are always executed and the result is
submitted before the execution of the branch instruction; this behavior is regardless
of the occurrence of the branch.
The delay slot is used to improve the efficiency of the pipeline on the early
version of platforms which are incapable of predicting the branching. Since modern
processors are capable of predicting the branching, the delay slot is useless.
Nevertheless, aiming to achieve the compatibility of software system, the delay
slot is also kept on the platform like MIPS, PowerPC and DSP. Thus, for the
purpose of reducing the required storage overhead, the proposed signature based
detection method is stored in the branch delay slot [1].
The instructions in the delay slot are supposed to be an independent basic block
according to the definition. Nevertheless, in the function level, the instructions have
a strong correlation with those in former basic block. Hence, the instructions should
be merged into the former basic block. Furthermore, some delay slots are composed
of completely empty operation instructions. Thus, there is nonsense of classifying
the delay slot as a basic block; this is done also for the purpose of reducing the
required detection overhead. To sum up, the delay slot is supposed to belong to the
former basic block in this work.
Program Control-Flow Structural Integrity Checking Based Soft Error. . . 53
The control-flow error detection is largely affected by the basic block partition. For
each basic block, it is a section of code that is executed sequentially. Thus the
control flow can only enter the basic block at the first instruction and leave after the
last instruction is executed. Usually, there are no more control-flow branches in the
basic block except the last instruction; and there is no other control-flow entry into a
basic block except the first instruction.
The basic block partition process can be implemented on two levels: C language
and assembly one [8, 9]. Due to the high-level property of C language, it is
independent of platforms. Therefore, the C program is capable of being executed
on different platforms, after being divided into basic blocks and hardened by
control-flow detection technique. However, the high-level language code has to
go through the compilation and linking process to become executable; otherwise,
the program is unable to be executed run on the machine. In the process of
converting to low-level language, the basic block implemented by C language is
not well mapped to lower level languages.
At present, most basic block partition techniques are implemented based on the
assembly language. Mnemonics are utilized by the assembly instructions to trans-
late the machine instructions. These two kinds of instructions (assembly and
machine ones) correspond to each other well; thus the basic blocks divided on the
assembly level are well mapped to the executable ones consisting of machine
instructions. Hence, the purpose and appeal (adding signature etc.) of the partition
on the language level can be satisfied during the execution of the program. In
addition, the control-flow error detection technique based on SIC is required to get
the structure information of each basic block, including the number of instructions
and the address of the last instruction. Thus, this paper chooses to implement the
basic block partition on the assembly level.
For different instruction sets, they all include the instructions like function calls,
function returns and jumps, regardless of platforms. The basic block partition
process is completed by seeking corresponding positions of the transfer instruc-
tions, such as function calls, function returns and jumps; these instructions are
utilized to locate the boundary of the basic block. Here, the system function calls
and interrupts are not considered as the transfer instructions to dispose. Therefore,
the two types of instruction are both regarded as the ordinary instructions in the
partition process [9].
To realize the SIC based control-flow detection scheme, various information related
with each basic block is required, including the entry port (i.e., the address of the
first instruction), the exit port (i.e., the address of the last instruction), the length of
54 Y. Guo et al.
each block (i.e., the number of instructions included by the block), next address
(i.e., the target address for the jump operation or the address where the first
instruction of the sub function locates if there is a function call operation). In
addition, it is also necessary to back up the instructions and store three copies of
the last instruction in the block, for the purpose of setting checkpoints. Hence, we
need to design the appropriate data structure.
Nevertheless, the number of basic block divided by the program is unable to be
determined as it is dynamic and the process happens at run time. Therefore, the
designed data structure needs to be growing dynamically. In this paper, the block
table is designed as shown in Fig. 2.
Based on the above analysis, the partition rules are listed as follows:
1. Taking the positions of the jump instruction, the function call instruction and the
function return instruction as the exit to the current basic module;
2. Regarding the location of the following instruction (the first instruction after the
delay slot if it exists) after the current jump instruction (call or call return) as the
entrance to the next basic block;
3. Taking the position where the destination instruction of the current jump locates
as the entrance to another basic block.
In fact, the basic block partition process is equivalent to constantly seeking the
position of transfer instructions by scanning the assembly program. These positions
are utilized to locate the boundary of the basic block. Furthermore, for the conve-
nience of the detection, a working pointer is also set for the block table in this work.
The pointer always points to the next node where the information about the block
that is able to be checked completely by SIC.
Integrity checking refers to the final inspection of each basic block did the
current basic block are complete
The SIC refers to check whether the basic block is completely executed at the end.
If the execution is completed, then control-flow errors did not happen in the
execution process of the program in the previous block. The SIC based detection
scheme determines the occurrence of control-flow errors by checking by checking
whether each basic block is completely executed or not.
Program Control-Flow Structural Integrity Checking Based Soft Error. . . 55
'
%DVLF %DVLF %DVLF
EORFN EORFN EORFN
& (
$
%
In order to implement SIC on each basic block, the checkpoints should be set at
the exit of the block. In Fig. 3, when the program is executed in basic block 1, the
control flow will jump from the fifth instruction to the tenth one given the occur-
rence of the control-flow error B. Due to the control-flow error B, the checkpoint
being set at the sixth instruction is missed but the working pointer in the block table
still points to the node. Hence, the structure information of basic block 1 is
obtained. When the program is executed to the eleventh instruction in basic block
2, then the SIC based error detection process starts after the program is running into
the checkpoint being set at basic block 2. Here, the detection module compares the
address of ‘11’ for the current checkpoint with the exit information ‘6’ of the basic
block indicated by the working pointer, and obviously this information is in
different. As a result, the control-flow error occurs between two basic blocks and
this is figured out by the detection module through comparison. Most control-flow
errors can be detected successfully by that address comparison.
The control-flow error occurring inside a basic block cannot be detected by
comparing addresses. As illustrated in Fig. 3, A indicates a control-flow error inside
a basic block. Then after executing the second instruction in basic block 1, the
program transfers to the fourth instruction. After executing the fifth instruction, the
control-flow error B occurs. Thus, the checkpoint being set at the sixth instruction is
unable to detect this error. Hence, a counter with an initial value of 0 is utilized here
when the control flow enters a basic block. The corresponding value of counter is
added by 1 if an instruction is executed.
After the control flow enters the checking module, the address of the checkpoint
is compared with the block exit information in the node pointed by the working
pointer first, and then the value of counter is compared with the length of
corresponding block. If the two kinds of information are consistent, then the basic
block is considered to be executed completely without any control-flow errors.
Nevertheless, if the address information is not the same as the exit information, then
there is a control-flow error between two basic blocks; if the value of counter is not
the same as the block length, this indicates a control-flow error is occurred inside
the basic block.
Hence, the SIC based detection technique is able to detect all control-flow errors
as presented in Fig. 1 except that of C1. Thus, additional fault-tolerant mechanism
needs to be applied to avoid the influence of the error such as C1.
56 Y. Guo et al.
&KHFNLQJ
PRGXOH
In order to realize SIC based detection, the normal execution of program should be
‘interrupted’ by transferring the control flow to the checking module. To ensure the
control flow returning to the original program correctly, we need to save the
breakpoint information after the flow enters the checking module. Then, the
consistency check of address and counter are to be completed by the checking
module. Based on the analysis of the result for the comparison, we can easily figure
out whether the current basic block is executed completely or not, i.e., whether
there is a control-flow error occurred in the block or not. If the error occurs, the
execution of program is paused and the error is reported to the system; otherwise,
the next SIC starts to be initialized, i.e., moves the working pointer in the block
table and assigns the value of counter to be 0.
Then, this indicates that the SIC based checking of the previous block is
completed successfully, and the working pointer must be pointed to the next node
for performing next operation. However, there is no real transfer of the control flow
in the checking module and which block to execute next is unknown. To solve this
problem, we introduce another process, named ‘pre-transfer’.
The ‘pre-transfer’ process is to restore the program breakpoint previously saved,
and then execute three instructions stored in block table. After that, the results are
processed by a voter, namely triple modular redundancy (TMR). According to the
result of voting, the block to execute next is determined, and then the working
pointer repoints to the node where information of next block is stored.
After all the steps are executed, the control flow is still inside the checking
module. Therefore, it is necessary to transfer the control flow back to the program at
the end of the checking module. This can be done just by restoring the breakpoint
saved and performing TMR on the replaced instructions.
4 Performance Analysis
The overhead is defined as the ratio of the execution time of detection mechanism
and that of the original program. In order to reveal the relationship of the overhead
and number of circles, analyses of several scenarios are performed. For the scenar-
ios presented here, different numbers of checkpoints are set. In Table 1, T30
indicates the original executable file; T31 contains one checkpoint and the checking
module is executed 256 times; T32 has two checkpoints and the checking module is
executed 512 times; T33 includes 15 checkpoints and the most of circles have been
optimized; T34 contains 27 checkpoints and there is no optimization. Here, it costs
34 clocks if the checking module is executed for one time. Corresponding simula-
tion results are presented in Table 1. The overhead is mainly influenced by the
number and time of circles, as revealed by the results in Table 1.
For the SIC based detection scheme, four different processes are introduced
listed as breakpoint protection, detection implementation, initialization of next
detection (moving working and reset the value of counter) and return to the original
program. Similarly for the signature-based detection methods, there are also four
processes, listed as breakpoint protection, signature checking, signature update and
return to the original program. Therefore, the overhead of SIC based detection
scheme is roughly the same as that of signature-based detection methods.
5 Conclusions
This paper proposes an SIC based soft error detection method for the DSP. Firstly,
the program for the DSP in assembly language is divided into a number of basic
blocks with structure information stored in a partition table. Then, a checkpoint is
set at the end of each basic block. By examining the consistency of the address
information and the counter value, the program control flow errors can be easily
detected. Compared with signature-based method, the proposed method can reach
almost 100% of error detection coverage. As it is implemented by assembly
language, it has better cross platform portability while detection efficiency and
detection overhead can also be ensured.
References
1. Nodoushan MJ, Miremadi SG, Ejlali A (2008) Control-flow checking using branch instructions.
In: Proceeding of the IEEE/IFIP international conference on embedded and ubiquitous com-
puting (EUC 2008), Shanghai, China
2. Xing KF (2007) Single event effect detection and mitigation techniques for spaceborne signal
processing platform. National University of Defence Technology, Changsha
3. Alkhalifa Z, Nair VSS, Krishnamurthy N et al (1999) Design and evaluation of system-level
checks for on-line control-flow error detection. IEEE Trans Parallel Distrib Syst 10:627–641.
doi:10.1109/71.774911
4. Oh N, Shirvani PP, McCluskey EJ (2002) Control-flow checking by software signatures. IEEE
Trans Reliab 51:111–122. doi:10.1109/24.994926
5. Reis GA, Chang J, Vachharajani N et al (2005) SWIFT: Software implemented fault tolerance.
In: Proceedings of the third international symposium on code generation and optimization
(CGO), San Jose, CA
6. Borin E, Wang C, Wu YF et al (2006) Software-based transparent and comprehensive control-
flow error detection. In: Proceedings of the international symposium on code generation and
optimization (CGO), New York, NY
7. Benso A, Di Carlo S, Di Natale G, Prinetto P (2002) Static analysis of SEU effects on software
applications. In: Proceedings of the international test conference (ITC), Baltimore, MD
8. Goloubeva O, Rebaudengo M, Sonza Reorda M, Violante M (2003) Soft-error detection using
control flow assertions. In: Proceedings of the 18th IEEE international symposium on defect and
fault tolerance in VLSI systems (DFT’03), Boston, MA
9. Huang ZY (2006) Research and implementation of software error detection technique for
on-board computers. Harbin Institute of Technology, Harbin
Research on the Fault Diagnosis of Planetary
Gearbox
1 Introduction
Planetary gearbox is a main component of the wind turbine that transmits heavy
loads from the driving motor to the driven machine with compact structure, large
transmission ratio. Based on the fault statistics of the wind turbine, the outages
caused by planetary gearbox accounts for 5.0%. However its downtime accounts for
a very large share, about 16.3% due to the bad working conditions. Thus the
researches on the fault recognition of planetary gearbox have been obtained more
and more attention to timely discover accurate fault position, predict the trend of
running status [1].
According to the planetary gearbox vibration signal characteristics, the domestic
and foreign scholars have applied various methods to carry on the exploratory
research. During the 1990s, foreign scholars began to study planetary gears fault
diagnosis. For example, the time average method was generalized by McFadden
[2, 3] and Samuel and Pines [4] etc., the calculation method of the vibration
components of the planet wheel and/or the solar wheel by using single sensor and
multiple sensors are proposed respectively. Samuel [5] proposed a constrained
adaptive algorithm to detect the damage of the gearbox based on lifting wavelet
method. Williams and Zalubas [6] used Wigner-Vine distribution to detect the
helicopter planetary gear box fault. Bartelmus and Zimroz [7] applied the cyclic
stationary analysis method to study the modulation characteristics of planetary
gearbox. Barszcz and Randall [8] applied spectral kurtosis to detect the crack of
gear teeth in the wind planetary gearbox. Hu Niaoqing [9, 10] proposed energy
characteristics in the neighborhood of meshing frequency method based on Hilbert
Huang transform (HHT) and gray correlation analysis method to detect the sun gear
fault in the planetary gearbox. Feng Zhipeng [11, 12] proposed the fault vibration
model based on the meshing frequency of the planetary gearbox, amplitude demod-
ulation analysis method and its vibration signal model. The fault diagnosis of
planetary gear box is realized by the frequency demodulation method based on
empirical mode decomposition.
The flow chart of fault diagnosis method that combines maximum correlated
kurtosis deconvolution (MCKD) with frequency slice wavelet transform (FSWT)
of planetary gearbox is shown in Fig. 1. This paper firstly acquires the gear fault
vibration signal via the signal acquisition equipment and selects the appropriate
shock period, venue and the number of iteration steps, while using particle swarm
optimization algorithm to find out the optimal filter length. Then, the fault signal of
gear is denoised by MCKD in time domain. Finally, the gear fault feature is
extracted and the fault resource is detected by applying FSWT method to analyse
the denoised signal.
Research on the Fault Diagnosis of Planetary Gearbox 63
The maximum correlation deconvolution [13] takes related kurtosis as the evalu-
ation index, give full consideration to the signal containing the periodicity of the
impact component. The cycle impact of gear fault will appear in the measured
signal, in order to extract the signal of periodic impact characteristics, the maxi-
mum correlation kurtosis solution convolution method get the signal correlation
kurtosis maximum value by filtering through the optimal filter.
Correlated Kurtosis (CK):
2
P
N Q
M
ynmT
n¼1 m¼0
CK M ðT Þ ¼ max N f ¼ ½ f 1 f 2 . . . f L T ð1Þ
f P 2
yn
n¼1
64 T. Han et al.
In the formula, yn is the periodic function, T is the period of signal yn, f is the
filter coefficient matrix, L is the length of the finite impulse filter, and M is the
shift number.
In MCKD algorithm, the choice of filter length parameter L and reconciliation
periodic convolution parameter T has an important influence on the deconvolution
result. If a parameter remains the same, only take another parameter as the optimi-
zation object to discuss the effects of parameters on the calculation results, the local
search methods ignore the interaction of two parameters, the obtained parameters
are relatively optimal.
Particle swarm algorithm as a swarm intelligence optimization algorithm, can
effectively reduce the influence and has good global searching ability, this paper
use particle swarm optimization algorithm to optimize the two parameters of
MCKD algorithm simultaneously, the adaptive filtering of the filter length para-
meter L and the reconciliation periodic convolution parameter T is realized.
Its principle is as follows, search the dimension Q space which consisting of the
groups of N particles, each correlated particle can be expressed as: xi ¼ (x1, x2,
x3, . . . , x), the velocity of every correlated particle can be expressed as: vi ¼ (vi1,
vi2, vi3, . . . , viQ). Each particle iterated updates its velocity and position by the
local extreme value and population global extreme value, its velocity and position
updating formula in particle swarm optimization algorithm are as follows:
vid kþ1 ¼ wvid k þ c1 ξ pid k xid k þ c2 η gid k xid k
ð2Þ
xid kþ1 ¼ xid k þ vid kþ1
In the formula, pj is a (j) in the normalized form, a (j) is the envelope of the
signal y (j) after Hilbert demodulation.
The main words in all headings (even run-in headings) begin with a capital letter.
Articles, conjunctions and prepositions are the only words which should begin with
a lower case letter.
Research on the Fault Diagnosis of Planetary Gearbox 65
Operate the MCKD to the signal and calculate the entropy in the envelope of the
convolution solution signal with the position of any particle Xi (i.e. parameter L and
T), the entropy Eg (Xi) represents the fitness value of the particle. When the solution
of convolution appears cyclical shocks, the envelope entropy becomes smaller and
the solution the convolution effect becomes less effective and the regularity of
impact becomes less obvious, but envelope entropy is relatively large.
Therefore, take the minimization of convolution signal envelope entropy as the
final optimization goal, the optimization process of determining the reconciliation
periodic convolution parameter T and the filter length parameter L though MCKD
is shown in Fig. 2.
In this paper, the gear fault vibration signal in the NGW11-12.5 planetary gear box
is obtained on the test rig with the DH5927 vibration acquisition instrument. The
test rig contains the motor, planetary gearbox, transmission shaft, magnetic powder
brake and tension controller (loaded), computer and the DH5927 vibration acqui-
sition instrument, as shown in Fig. 3a. Planetary gear box and the left end of the
motor are connected, and the right end is connected to the transmission shaft to
reduce speed. The speed of the motor is 1260 r/min and the secondary planetary
gear transmission ratio is 12.5:1. The frequency parameters of planetary gearbox
are shown in Table 1.
In order to validate the feasibility of the proposed method, sun gear is line cut in
a tooth to simulate tooth breaking fault, as shown in Fig. 3b.
Fig. 3 Planetary gearbox test rig. (a) Planetary gearbox test rig. (b) Tooth broken sun gear
the diagram that the waveform in time domain and frequency domain diagrams is
complex and difficult to distinguish the specific features of signals. Thus, the fault
feature is submerged in the background noise signal. It could be clearly seen that the
second harmonic component of the rotating frequency is prominent and the
meshing frequency also have appeared in the diagram, but the amplitude is not
obvious and the fault feature frequency hasn’t appeared.
In the process of de-noising with the application of MCKD, the parameters are
selected as follow: the impact period T is 43, the shift number M is 5 and the
iteration number is 30.The filter length L is calculated to set 30 by using particle
68 T. Han et al.
Fig. 4 Diagrams of the waveform in time and frequency domain. (a) Waveform in frequency
domain. (b) Waveform in time domain
swarm optimization algorithm when the envelope entropy value of the convolution
signal is minimum.
After determining the parameters, the maximum correlation kurtosis deconvol-
ution is used to reduce the measured waveform of solar gear broken teeth in time
domain, and the original signal and the denoised signal are shown in Fig. 5.
Research on the Fault Diagnosis of Planetary Gearbox 69
Fig. 5 Signals before and after noise de-noising. (a) Input signal. (b) Signal filtered by MCKD
[0,350] Hz.
The FSWT results both before and after signal denoising are presented in Figs. 6
and 7, it is obviously found in the diagrams that after MCKD denoised, the features
of shock is strengthened, and the sun gear fault feature frequency 57.69 Hz and its
2X–5X harmonic generation are shown in the latter diagram and the energy is very
high. In addition, the quadruple rotating frequency of sun gear and the meshing
frequency 249.8 Hz also appear in frequency diagram. So the denoising and fault
diagnosis method for signals in time domain in planetary gearbox are comparably
applicable.
70 T. Han et al.
4 Conclusions
This paper presents a new method for fault diagnosis of the planetary gearbox
combined MCKD with particle swarm optimization algorithm and FSWT. In the
proposed method, MCKD is used to de-noise the raw signal, the features of periodic
impulse components are strengthened after noise reduction. The time-frequency
distribution of the vibration signals could be done using FSWT to identify the gear
fault. Based on above study, the following conclusions are drawn:
For early fault of planetary gearbox, the fault signal feature is relatively weak
and largely effected by environmental noise and transmission path. Thus, the fault
feature extraction is relatively difficult. Therefore, the maximum correlation kur-
tosis deconvolution is used to efficiently prominent periodic impulse component in
the signal and to suppress other noise involved.
The frequency slice wavelet function is introduced to enable the traditional
Fourier transform realize time-frequency analysis function, which makes the signal
filtering and segmentation more flexible and the fault feature of the planetary gear
box can be clearly identify.
References
1. Long Q, Liu Y, Yang Y (2008) Applications of condition monitoring and fault diagnosis to
wind turbines. Modern Electric Power
2. Mcfadden PD (1991) A technique for calculating the time domain averages of the vibration of
the individual planet gears and the sun gear in an epicyclic gearbox. J Sound Vib 144(1):
163–172
3. Mcfadden PD (1994) Window functions for the calculation of the time domain averages of the
vibration of the individual planet gears and sun gear in an epicyclic gearbox. J Vib Acoust
116(2):179–187
4. Pines DJ (2000) Vibration separation methodology for planetary gear health monitoring.
Proc SPIE – Int Soc Opt Eng 3985:250–260
5. Samuel PD, Pines DJ (2009) Constrained adaptive lifting and the CAL4 metric for helicopter
transmission diagnostics. J Sound Vib 319(1–2):698–718
6. Williams WJ, Zalubas EJ (2000) Helicopter transmission fault detection via time-frequency,
scale and spectral methods. Mech Syst Signal Process 14(4):545–559
7. Zimroz R, Bartelmus W (2009) Gearbox condition estimation using cyclo-stationary proper-
ties of vibration signal. Key Eng Mater 413(1):471–478
8. Barszcz T, Randall RB (2009) Application of spectral kurtosis for detection of a tooth crack in
the planetary gear of a wind turbine. Mech Syst Signal Process 23(4):1352–1365
9. Feng Z, Hu N, Cheng Z (2010) Faults detection of a planetary gear based on condition indicator
in time-frequency domain. Mechanical Science & Technology for Aerospace Engineering
10. Zhe C, Niao-Qing HU, Gao JW (2011) Scuffing damage quantitative detection of planetary
gear set based on physical model and grey relational analysis. J Vib Eng 24(2):199–204
72 T. Han et al.
11. Feng Z, Zhao L, Chu F (2013a) Amplitude demodulation analysis for fault diagnosis of
planetary gearboxes. Zhongguo Dianji Gongcheng Xuebao/Proc Chin Soc Elect Eng 33(8):
107–111
12. Feng Z, Zhao L, Chu F (2013b) Vibration spectral characteristics of localized gear fault of
planetary gearboxes. Proc CSEE 33(2):118–125
13. Tang G, Wang X, Energy SO (2015) Adaptive maximum correlated kurtosis deconvolution
method and its application on incipient fault diagnosis of bearing. Zhongguo Dianji Gong-
cheng Xuebao/Proc Chin Soc Elect Eng 35(6):1436–1444
14. Zhong X (2014) Research on time-frequency analysis methods and its applications to
rotating machinery fault diagnosis. Wuhan University of Science and Technology
15. Duan C, Qiang G, Xianfeng XU (2013) Generator unit fault diagnosis using the frequency slice
wavelet transform time-frequency analysis method. Proc CSEE 33(32):96–103
Bridge Condition Assessment Under Moving
Loads Using Multi-sensor Measurements
and Vibration Phase Technology
Abstract This paper presents a bridge condition assessment approach under mov-
ing loads with multi-sensor measurements and vibration phase technology. The
phase trajectories of multi-sensor responses are obtained and a damage index is
defined as the separated distance between the trajectories of undamaged and
damaged structures to identify the damage existence and location. The damage
will induce the local structural stiffness change, and the distance between the
corresponding points on those two trajectories of undamaged and damaged beams
when a moving load travels across the damaged location will also change. Exper-
imental studies demonstrate the proposed approach can be used to successfully
identify the shear connection failure in a composite bridge model.
1 Introduction
Dynamic responses of bridge structures subjected to moving load could be used for
assessing structural conditions [1]. In practical applications, the properties of the
moving vehicle could not be obtained and thus they are usually assumed as
unknown parameters. Zhang et al. [2] proposed a method for simultaneous identi-
fication of moving masses and structural damage from measured responses. Zhu
and Law [3] performed the simultaneous identification of the moving forces and
H. Hao • J. Li (*)
Center for Infrastructural Monitoring and Protection, Curtin University, Kent Street, Bentley,
WA 6102, Australia
e-mail: [email protected]; [email protected]
W. Zhang
Department of Mechanics, Taiyuan University of Science and Technology, Taiyuan, Shanxi
030024, China
e-mail: [email protected]
H. Ma
College of Science and Engineering, Jinan University, Guangzhou, Guangdong 510632, China
e-mail: [email protected]
damage. However, this has not been fully explored and tested, particularly when
multi-type vibration measurements are acquired from bridge structures under
moving loads and used for damage detection.
For convenience, (u0, v0, a0) and (ud, vd, ad) denote the spatial coordinates on the
phase trajectory from the undamaged and damaged beam models, respectively. The
damage will induce the local structural stiffness change, and the distance between
the corresponding points on those two trajectories of undamaged and damaged
beams when a moving load travels across the damaged location will also change.
The phase trajectory of the undamaged beam is considered as the baseline, a
damage index based on the change in the phase space trajectories is used to identify
the structural damage.
With the above mentioned bridge-vehicle system, the time axis could be nor-
malized as the location of the moving load by x ¼ ct, where c and t are the travelling
speed and time, respectively. In a general case, the location of the moving load is
often expressed as the normalized location by x’ ¼ x/L, in which L is the length of
the beam. The damage index is defined as the distance between the spatial coordi-
nates of the normalized moving load locations from the undamaged and damaged
phase trajectories
qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
0 0
0 0 2 0 2 0 2
DI ¼ uxd ux0 þ vxd vx0 þ axd ax0 ð2Þ
0 0 0 0 0 0
where uxd ; vxd ; axd and ux0 ; vx0 ; ax0 denote the spatial coordinates on the
damaged and undamaged phase space trajectories of the normalized locations of the
moving load. It is noted that by using the normalized locations of the moving load
on the bridge to define the damage index, the proposed approach can be used for
76 H. Hao et al.
scenarios with different travelling speeds of the moving loads before and after
damage.
To overcome the impact effect when the vehicle starts and stops as well as the
measurement noise effect, a simple moving average (SMA) [14] process is used to
smooth the responses. It is the unweighted mean of a specific number of data points.
SMA is commonly used with time series data to smooth out short-term fluctuations
and highlight longer-term trends. Supposing x is a time series, SMA is written as
1 X
N 1
x s ði Þ ¼ xði þ jÞ ð3Þ
N j¼0
where xs and x represent the smoothed result and the original series, respectively,
and N is the number of points used to calculate the average. SMA is used to process
the dynamic vibration responses and then the phase trajectory is plotted.
Multiple types of vibration responses are used to construct the multi-dimensional
vibration phase space and plot the trajectory, such as a three dimensional phase
space “displacement-velocity-acceleration”, or a two dimensional phase space
“displacement-acceleration”. By analyzing the phase space trajectories and calcu-
lating the damage index in the three dimensional or two dimensional phase space,
the local maximum value could be used to indicate the damage location
3 Experimental Verifications
Fig. 1 Experimental
testing model
Bridge Condition Assessment Under Moving Loads Using Multi-sensor. . . 77
screwing into metric nuts which were cast in the slab were used to link the slab and
girder together. The nuts were welded onto the reinforcement bar in the slab before
pouring. Each girder has 16 connectors with an equal space of 200 mm and there are
32 in total in the bridge model, which are denoted as SC1~SC16 and SC10 ~SC160 ,
as shown in Fig. 2. The bridge model is placed on two steel frames which are fixed
to the laboratory strong floor. Figure 3 shows the experimental “bridge-vehicle”
system. The vehicle model is simplified as a steel beam carrying two concrete
blocks. The distance between the front axle and rear axle of the vehicle is 1 m. A
roller system is designed and the crane is used to pull the vehicle to travel on the top
of the bridge model with a constant speed as the pulling force of the crane is steady
and stable. A fabricated track is placed and clamped on the top of the concrete slab
to make sure the vehicle is moving on the predetermined traveling path. Three
transverse connecting plates are located at the two ends and the center of the track to
ensure the stability and rigidity of the track, as shown in Fig. 1. Subjected to the test
conditions, the velocity response is not measured in the experimental tests, and the
displacement and acceleration responses are used for the analysis. Only a displace-
ment sensor and an accelerometer are used, and their locations are marked in Fig. 2.
The sampling rate is set as 2000 Hz.
78 H. Hao et al.
When all the connectors are tight into the nut, the bridge is defined as the
undamaged state. If a specific connector is removed from the nut, it indicates the
failure of the particular shear connector which also shows that the bridge is under
the damaged state. In this study, two damage scenarios and their associated
damaged shear connectors are shown in Table 1. The regions of those introduced
damages in these two damage scenarios are also listed. The measurements from
both the undamaged and damaged states of the bridge are taken. Figure 4 shows the
measured displacement and acceleration responses from the undamaged model.
The time instants when the vehicle starts and stops can be observed from the
measured responses. In addition, there are some unexpected oscillations in accel-
eration responses. They are probably induced by the impact on the track when the
front and rear axles of the vehicle are crossing the connecting transverse plate in the
centre of the track. The Fourier spectrum of the acceleration response from the
undamaged state and Damage Scenario 1 are shown in Fig. 5. The fundamental
Fig. 5 Fourier spectrum of the acceleration response from the undamaged state and Damage
Scenario 1
Fig. 6 Comparison of the phase trajectories between the undamaged and damage states. (a)
Scenario 1 and (b) Scenario 2
The travelling path of the bridge-vehicle system, the distance of the front and
rear axles of the vehicle and the number of shear connectors are shown in Fig. 9.
Damage scenario 1 is associated with the removal of shear connectors SC15 and
SC150 , and Damage scenario 2 associated with shear connectors SC2 and SC20
being removed. Only the front axle will pass along the region of removed shear
connectors in Damage scenario 1, and only the rear axle will pass the area of
removed shear connectors in damage scenario 2. The highest damage index as
shown in Fig. 7 indicates the location of damage 1 when the front axle moves on the
top of damage. It is roughly located at the normalized location 0.91, which matches
Bridge Condition Assessment Under Moving Loads Using Multi-sensor. . . 81
well with the true introduced damaged location at 2.8m/3 m ¼ 0.93, indicating that
the damage can be identified effectively for the Damage Scenario 1 with the
proposed approach. The other two peak damage index values at the normalized
locations of 0.3 and 0.8 are corresponding to the instants when the front and rear
82 H. Hao et al.
Fig. 9 The schematic diagram of the bridge-vehicle system and travelling path
axles impact the track at the central transverse plate. This is also observed in the
acceleration measurements as shown in Fig. 4. For Damage Scenario 2, the detec-
tion result is shown in Fig. 8. Besides the high damage index values associated with
the locations of damage 1, the instants of the front and rear axles moving across the
central transverse plate of the track as mentioned above, one more relatively high
damage index value at the normalized location 0.1 corresponds to the instant when
the rear axle moves on the top the location of damage 2, which is at 0.2m/
3 m ¼ 0.67. This can be used to identify the location of damage 2.
When the time instants of the identified damaged locations as shown in Figs. 7
and 8 are obtained, the damaged locations could be determined based on the vehicle
speed from the measurement record. It is worthy to note that the real travelling
distance of the vehicle is about 2 m instead of 3 m because the distance between the
two axles of the vehicle is 1 m, as shown in Fig. 9. In this regard, damage 1 can be
detected by using the response when the front axle crosses this region, and damage
2 may be detected by the rear axle. Supposing c and t1 are the moving speed of the
vehicle and the identified time instant when the front axle is crossing the location of
damage 1, the damage region can be obtained as Lx1 ¼ ct1 + Lv, in which Lv is the
distance between the two axles of the vehicle. For damage 2, the location can be
calculated as Lx2 ¼ ct2 with t2 identified as the time instant of the rear axle crossing
the location of damage 2. The speed of vehicle c can be obtained as the average
speed when the vehicle moves along the bridge. The total time during the vehicle
travelling on the top of the bridge model is 49.44 s, the vehicle speed is obtained as
0.04 m/s. Table 2 shows the true and identified damage locations in the
Bridge Condition Assessment Under Moving Loads Using Multi-sensor. . . 83
4 Conclusions
References
1 Introduction
Production processes has over the years become more complex, applying produc-
tion philosophies like just in time (JIT) and involving automatic and advanced
production techniques, [1–3]. This, together with harsh competition has increased
the severity of downtime in the production processes. This makes the importance of
maintenance greater due to its role in keeping and improving availability, perfor-
mance efficiency, quality, on-time deliveries, environment, safety and total plant
productivity, see for example [4–9].
The positive financial effects of maintenance are widely recognized in the
scientific community, for example [7, 10–13]. In [7] the authors conducted a
study in north east of England during 1997 and 1998 including 298 companies
followed by a smaller study among 23 companies and they found that companies
systematically adopting best practices in maintenance do achieve higher perfor-
mance. Also, among companies it is more and more realized that effective main-
tenance can contribute significantly to companies profitability, even if it
traditionally has been considered as a non-core function, see for example
[4, 14]. Yet, a study conducted among Swedish industry gave as a result that 70%
is considering maintenance only as a cost [15].
Several authors have pointed out the importance of decision making in mainte-
nance management, e.g. what maintenance policy to use and when to stop produc-
tion for conducting maintenance actions [16–19]. However, to be able to make
efficient decisions in maintenance require, among other things, knowledge about
maintenance technical and economic performance. When following up mainte-
nance performance maintenance costs are traditionally divided into direct and
indirect costs. Some of these costs, such as spare parts and direct labour, can
comparably easy be related to maintenance. Also some of the indirect costs are
also relatively easy to relate to maintenance, see for example [20]. However,
savings and profit generated by maintenance are often neglected, and are not easy
to find in current accounting systems. In order to assess the economic impact of
maintenance its impact on life cycle cost/profit needs to be considered [18, 21]. This
is usually not an easy task. Due to the complex nature of maintenance it is not
possible to measure maintenance impact only in the maintenance function. Also its
impact on other areas such as quality and production must be considered.
Consequently, in order to be able to use maintenance as a cost-effective tool to
reduce production cost and increase company’s profitability a system based on a
holistic view of maintenance for monitoring its technical and economic impact,
considering both costs and savings of maintenance and providing both an overview
of maintenance performance and possibility to trace causes behind deviations in
maintenance performance, is required, [16, 22]. Such a system will enable cost-
effective decisions regarding maintenance, make it possible to follow up actions and
support continuous improvement of the maintenance function [23]. In [24] is reported
that maintenance initiatives often fail to deliver because the supportive systems such
as management and information systems are not in place. Thus, the main purpose of
this study is to develop and test a new system for monitoring and assessing the
economic performance of the production process with respect to maintenance.
monitoring techniques and advanced methods for prognosis and prediction, see for
example [18, 26, 27]. In [28] the authors summarized the strategically elements of
maintenance decisions into structural decision elements regarding capacity, facil-
ities, technology and vertical integration and into infrastructural decision elements
regarding organization, policy and concept, planning and control systems, human
resources, modification and performance measurement and reward systems. Deci-
sions in each of these areas will have an impact on the technical and financial
performance of the company. The technical consequences can be for example in
form of changed availability and reliability. Furthermore, some of these effects are
delayed and may only be seen over time. These technical effects of maintenance all
have financial consequences, e.g. low availability, reliability and quality rate lead to
increased production costs, increased costs for buffers and penalties for late deliv-
ery. In addition, the low internal effectiveness that can be a result from not having
proper maintenance may lead to decreased market shares, profitability, etc. Several
researchers have described maintenance impact on productivity and profitability of
a company, for example [15, 29, 30].
In order to assess the real impact from maintenance the life cycle cost (LCC)
concept need to be applied [18]. LCC is defined as the sum of research and
development costs, production and construction costs, operation and maintenance
support costs and retirement and disposal costs [31]. In this definition it is obvious
that maintenance influence LCC. The maintenance cost includes several mainte-
nance related cost factors, e.g. direct labour and spare parts but also costs due to
unavailable, inefficiency, delivery delay penalties and redundant equipment. In
addition to LCC it is often also necessary to assess the life cycle income (LCI) in
order to estimate the economic importance of maintenance, which can be difficult.
However, instead it is possible to assess the savings that can be gained from
maintenance, e.g. reduced downtime, increased quality, and reduced capital tied
in inventory and spare parts [10]. A prerequisite to apply cost-effective mainte-
nance is a well-functioning maintenance performance measuring system [32]. The
maintenance performance measuring system should assess the total economic
impact of maintenance as well as indicating where to invest in improvements.
This allows following up maintenance performance measures more frequently,
thereby be able to intervene on deviations earlier and thereby avoid unnecessary
costs. In [33] the authors discuss how to implement maintenance performance
measuring system, and describe how different performance indicators (PI) can be
constructed. The described PI:s are focused on the overall cost of maintenance,
e.g. maintenance cost/production output. In a review of state-of-the art in mainte-
nance performance metrics, authors in [34] refer to several frameworks for mea-
suring maintenance performance, but it seems that none of these framework directly
link decreased number of disturbances or improved quality to maintenance eco-
nomic impact.
88 A. Ingwald and B. Al-Najjar
There are mainly two reasons for measuring performance that are to follow up and
improve performance. This is also reflected in the requirements of a system,
EcoCon, for monitoring economic performance of maintenance. Consequently,
the output from EcoCon should provide the user with information regarding the
economic impact, including costs, losses and savings from maintenance at opera-
tive as well as strategic level. In addition to this the system should also provide
more detailed information regarding causes of deviation in maintenance perfor-
mance. Then the output will be useful for both control purposes and also enable
continuous improvements. These abilities of EcoCon may be achieved by showing
aggregated values and allow the user to bring up the data that can be found under
these aggregated values, see for example [22, 35].
Furthermore, to provide the manager with an overview of maintenance perfor-
mance it is necessary to translate technical maintenance performance into economic
terms. This makes it possible to compare and prioritise where improvements are
required and how much to invest in improvements. However, translating
non-economic measures into economic is not without problems. The authors of
[36] cite several sources that points out the difficulties in expressing operative
performance into economic terms, e.g. difficulties in putting a correct economic
value on downtime and those operational changes will not immediately occur as
profits in the accounting system. Problems with assessing the consequences of
downtime are also reported by [37]. Maintenance actions are performed at operative
level and influence the technical performance of a production system or a machine,
i.e. failures, short stoppages and quality rate. To show the economic impact from
maintenance all costs, losses and savings due that can be related to technical
performance of maintenance must be considered. This means that when
implementing EcoCon in a certain situation it is necessary to find the links between
operational maintenance performance and impact on tactical and strategic levels,
see Fig. 1.
The basis for measuring performance is that it is improvements in non-economic
measures that create economic improvement, see for example [38]. Consequently,
in order to assess the economic impact from an investment or action in maintenance
first, maintenance technical performance at operative level need to be determined,
i.e. number and length stoppages, length of short stoppages and quality level. Next,
EcoCon: A System for Monitoring Economic and Technical Performance of. . . 89
F1) Savings or F2) Savings or F3) Savings or F4) Savings or F5) Savings or losses
losses due to losses due to losses due to losses due to in other maintenance
changed number of changed average changed short changed quality related economic
failures (S1) failure time (S2) stoppage time (S3) level (S4) factors (S4 to S14)
Time
t0 t1 t2
Previous period Current period
1. (F1 to F5) The first two factors that are monitored and assessed are regarding
losses due to failures. S1: Savings or losses due to changed number of failures,
and S2: Savings or losses due to changed average stoppage length.
Two factors are used to describe the losses; due to the number of failures and
stoppage length.
Savings or losses in S1 and S2 are calculated as:
∗
S1 ¼ ðY yÞ∗ L1 PR∗ PM ð1Þ
∗
S2 ¼ ðL1 l1 Þ∗ y PR∗ PM ð2Þ
where
Y: Number of failures during previous period,
y: Number of failures during current period,
L1: Average failure time during previous period,
l1: Average failure time during current period,
EcoCon: A System for Monitoring Economic and Technical Performance of. . . 91
where
B: Number of short stoppages during previous period,
b: Number of short stoppages during current period, and
L2: Average short stoppage length.
S4: Savings or losses due to changed quality level. S4 can be assessed by:
where
p: Production of good quality during current period,
P: Production of high quality during previous period,
WH: work hours per day, and
WD: work days per year.
Maintenance economic impact in other identified maintenance related economic
factors S5 to S14, is assess by subtracting the cost for the current period with the
cost for the previous period. This method of assessing maintenance economic
impact is developed and described in [16, 21, 22, 39].
2. Assess maintenance total economic impact by adding S1 to S14.
3. Monitor maintenance performance by following the development of mainte-
nance total economic impact and the development in the maintenance related
economic factors, S1 to S14.
4. If a deviation occurs in either maintenance total economic impact or in any of the
factors S1 to S14 a deeper analyze of S1 to S4 using more detailed information is
required to find the causes of the deviations. For more detailed information, see
for example [35, 40].
5. Because the dynamic nature of a production system the system should regularly
be reviewed and updated to ensure that it is always reflecting the current
situation.
92 A. Ingwald and B. Al-Najjar
3.3 EcoCon
To develop a prototype of EcoCon Microsoft Excel was used, see Fig. 4. In order to
make it easy to understand we decided that it was important to show as much as
possible of the data used to assess the losses and saving due to maintenance. This
was an important issue because it makes it easier to understand for persons involved
in testing EcoCon. For the prototype of EcoCon data was gathered from two
sources: a system for registering disturbances in a production process, MUR
developed by Adductor. From MUR EcoCon fetch data regarding failures and
short stoppages. All other required data was entered manually in the prototype.
Below is a description of the most important features of EcoCon as shown in Fig. 4.
1. Setup: The data regarding failures and short stoppages that come from MUR are
mapped to S1 Savings or losses due to changed number of failures, S2 Savings or
losses due to changed average stoppage length and S3 Savings or losses due to
short stoppages. Because MUR is a measuring system and registers all stoppages
and disturbances in a production process. It was also necessary to associate each
disturbance type from MUR with maintenance or any other working area,
e.g. disturbance type number 5, Repair, can be connected to Maintenance
while disturbance type number 4 can be associated classified as Logistic and
so on. Also during setup it is necessary to manually enter the economic conse-
quences of downtime. During setup maintenance related economic factors S5 to
S14 is defined. Depending on the type of production these can vary and have to be
decided for each implementation.
Fig. 4 EcoCon
EcoCon: A System for Monitoring Economic and Technical Performance of. . . 93
Fig. 6 Prioritization of losses per working area and per failure type
2. Update: To monitoring maintenance impact Update is used. First the user have
to enter start and end date for the current period, then data related to stoppages
and short stoppages are fetched from MUR for the current period. Data related to
quality level, S4, and additional economic factors, S5 to S14, are also entered
now. When the required data is entered EcoCon immediately shows the result in
the excel sheet, see Fig. 4. The savings (profit) or losses gained because of
improved maintenance compared with last review period in each of the moni-
tored economic factors, S1 to S14, and total.
3. Prioritize: If significant deviation in any of the monitored maintenance related
economic factors occurs it is possible to further analyze data to find the cause,
see Fig. 5. If S1 or S2 deviates significantly it is possible to directly look at the
impact from individual failure types. It is also possible to prioritize according to
working area, i.e. to see if deviations can be related to maintenance or if there are
other reasons such as logistics or organizational causes behind the deviation.
This requires that the failure types are associated with working areas, see Fig. 6.
EcoCon have been tested using data from the mail terminal in Alvesta Sweden.
More than one million letters pass through and are sorted at the terminal each
24 hours. For this test historical data regarding stoppages and quality, covering a
period for approximately 3 years, from one sorting machine which handles about
94 A. Ingwald and B. Al-Najjar
100,000 letters each 24 hours was used. The data are stored in an Access data base.
Also, a few other economic factors that were possible to relate to stoppages and
quality were identified, i.e. extra costs for transportation when the sorting were not
finished on time, spare parts and direct maintenance cost. However, the historical
data did not allow us to include these factors in this test. Table 1 describe the data in
the access database and how it is associated with the maintenance related economic
factors in EcoCon. It could be argued that jam should not be associated with S1 and
S2 because the stoppage times are short. However, it was decided to do that because
this stoppage type and stoppage time was properly registered, while technical
failures include several different types of short stoppages without proper registra-
tion of the causes. In this test only S1 to S4 was considered. All collected data are
related to maintenance.
Three adaptations of EcoCon had to be made for this test:
1. Since the mail terminal have no influence over the amount of letters they have to
sort each day and do not sell a product it was not possible to use profit margin to
measure the economic consequences of lost production. Instead it was decided to
use the cost of using extra resources, in this case the sorting machine and the
operators. However, this is an area that needs to be further researched in order to
get an accurate cost.
2. Because the number of letters varies from day to day, the savings or losses
compared to previous period was shown both as normalized figures, i.e. savings
or losses per two million sorted letters, and as the actual saving or loss.
3. Also, because the number of letters varies from day to day the formula used for
assessing S4 savings or losses due to quality level, see equation 4, had to be
modified to compare based on a normalized number of letters entered.
Table 2 shows the result from using EcoCon on data covering a period from
2005-03-21 to 2008-05-28. Please note that the figures in the table are masked.
EcoCon: A System for Monitoring Economic and Technical Performance of. . . 95
From Table 2 it can be seen that savings have been made due to decreased number
of failures, S1, and also due to reduced average stoppage time, S2. EcoCon can show
that the savings made in S1 and S2 are mostly due to reduced numbers of letters
getting stuck in the machine, i.e. the machine has become more reliable. On the
other hand costs have increased due to increased short stoppages and reduced
quality level.
The result was discussed with the maintenance manager at the mail terminal and
the improvements made to S1 and S2 is the result of continuous improvement
activities in maintenance. The increased losses shown in S3 and the increased losses
in S4, between periods P3 and P4 may be due to introduction of new technology in
the sorting machine. Even if this was an initial limited test and we only considered
S1 to S4, EcoCon provided the maintenance manager with a new view of mainte-
nance performance by showing financial losses and savings that can be associated
with maintenance activities. The test of EcoCon will continue including additional
economic factors that can be traced to maintenance performance of this machine.
The main result presented in this paper is a system, EcoCon, for monitoring
maintenance economic performance. EcoCon present maintenance total increased
or decreased maintenance performance in economic terms and also the contribution
from the underlying maintenance related economic factors are presented in eco-
nomic terms. This allows both for monitoring maintenance technical and economic
impact. Furthermore, EcoCon can provide more detailed information to use for
analysing causes behind deviations in maintenance performance which support
cost-effective continuous improvement of maintenance.
By expressing maintenance performance in economic terms, EcoCon makes it
easier for maintenance managers to communicate with higher management. A test
that was performed showed that EcoCon can provide useful financial information
regarding maintenance performance. Also, during the test it was emphasized that
determine the cost of down time can be a problem.
A system for measuring economic maintenance performance relies heavily on
finding an adequate economic measure of downtime, which can be difficult. The
96 A. Ingwald and B. Al-Najjar
latter was revealed during the test of this system as well as previously being pointed
out by several authors. Consequently this is an area that needs further research.
Additionally the mechanism that makes savings that are gained on operational level
e.g. due to reduced number of failures and short stoppages, turn up as profit in the
accounting system needs to be further investigated in order to increase the confi-
dence in this type of systems.
Acknowledgements The authors would like to acknowledge Adductor for letting us use MUR for
this project, and also Postterminalen in Alvesta.
References
19. Waeyenbergh G, Pintelon L (2002) A framework for maintenance concept development. Int J
Prod Econ 77:299–313
20. Cavalier M, Knapp GM (1996) Reducing preventive maintenance cost error caused by
uncertainty. J Qual Maint Eng 2(3):21–35
21. Al-Najjar B (2007) The lack of maintenance and not maintenance which costs: a model to
describe and quantify the impact of vibration-based maintenance on company’s business. Int J
Prod Econ 107(1):260–274
22. Al-Najjar B (2009) A computerised model for assessing the return on investment in mainte-
nance; Following up maintenance contribution in company profit. In: Proceedings of the fourth
World Congress on Engineering Asset Management (WCEAM) 2009, Greece, Athens, 08 Sep
2009. Springer, London, pp 137–145
23. Al-Najjar B, Kans, M, Ingwald, A, Samadi, R (2003) Decision support system for monitoring
and assessing maintenance financial impact on company profitability. In: Proceedings of
business excellence I, performance measures, benchmarking and best practices in new econ-
omy, Portugal
24. Tsang AHC (2002) Strategic dimensions of maintenance management. J Qual Maint Eng 8
(1):7–39
25. EN 13306:2001 (2001) Maintenance terminology. Comite Europeen de Normalisation
26. Al-Najjar B (1997) Condition-based maintenance: selection and improvement of a cost-
effective vibration-based policy for rolling element bearings. Doctoral thesis, Lund University,
Lund Sweden
27. Kelly A (1997) Maintenance strategy. Butterworth Heinemann, Oxford
28. Pinjala SK, Pintelon L, Vereecke A (2006) An empirical investigation on the relationship
between business and maintenance strategies. Int J Prod Econ 104(1):214–229
29. Alysouf I (2007) The role of maintenance in improving companies’ productivity and profit-
ability. Int J Prod Econ 105(1):70–78
30. Kutucuoglu KY, Hamali J, Irani Z, Sharp JM (2001) A framework for managing maintenance
using performance measurement systems. Int J Oper Prod Manag 21(1):173–195
31. Fabricky WJ, Blanchard BS (1991) Life cycle cost and economic analysis. Prentice-Hall,
Englewood Cliffs
32. Pintelon L (1997) Maintenance performance reporting systems: some experiences. J Qual
Maint Eng 3(1):4–15
33. Stenstr€om C, Parida A, Kumar U, Galar D (2013) Performance indicators and terminology for
value driven maintenance. J Qual Maint Eng 19(3):222–232
34. Kumar U, Galar D, Parida A, Stenstr€om C, Berges L (2013) Maintenance performance
maetrics: a state-of-the-art review. J Qual Maint Eng 19(3):233–277
35. Pintelon L, Van Puyvelde F (1997) Maintenance performance reporting systems: some
experiences. J Qual Maint Eng 3(1):4–15
36. Caplice C, Sheffi Y (1995) A review and evaluation of logistics performance measurement
systems. Int J Logist Manag 6(1):61–74
37. Moore WJ, Starr AG (2006) An intelligent maintenance system for continuous cost-based
prioritisation of maintenance activities. Comput Ind 57(6):595–606
38. Kaplan RS, Norton DP (2007) Using the balanced scorecard as a strategic management system.
Commun Manag 85(7):150–161
39. Ingwald A, Al-Najjar B (2006) System for monitoring machines financial and technical
effectiveness. In: Proceedings for the 19th international congress of Condition Monitoring
and Diagnostic Engineering Management (COMADEM 2006), pp 219–229
40. Martorell S, Sanchez A, Mu~noz A, Pitarch JL, Serradell V, Roldan J (1999) The use of
maintenance indicators to evaluate the effects of maintenance programs on NPP performance
and safety. Reliab Eng Syst Saf 65(2):85–94
41. Ljungberg Ö (1998) Measurement of overall equipment effectiveness as a basis for TPM
activities. Int J Oper Prod Manag 18(5):495–507
An Adaptive Power-Law Degradation Model
for Modelling Wear Processes
R. Jiang
Abstract The power-law model is often used as the mean value function of the
non-homogeneous gamma process for describing nonlinear degradation behaviour
of a product population. Since there is unit-to-unit variability among individual
degradations, it would be inaccurate to use the power-law process of the population
to predict the time to a degradation limit for a given individual. To address this
issue, we present an adaptive power-law degradation model in this paper. The scale
parameter of the power-law model is dynamically updated on the basis of the recent
degradation observation. A method that is somehow like the exponential smoothing
approach is adopted to modify the scale parameter. A real-world example that deals
with a non-homogeneous wear process is included to illustrate the appropriateness
of the proposed model. A graphical method is also presented to intuitively display
the relative health level of a condition monitored item.
1 Introduction
R. Jiang (*)
Changsha University of Science and Technology, Changsha, China
e-mail: [email protected]
[6–8]). In the literature, adaptive approaches are also applied to update maintenance
decision rules (e.g., see [9–12]).
In this paper, we consider the power law wear process and assume that the shape
parameter of the power-law model is a constant but the scale parameter is dynam-
ically updated based on the recent wear observation. A method, which is somehow
like the exponential smoothing approach, is developed to modify the scale param-
eter. A graphical method is presented to intuitively display the evolution trend and
relative health level of a wear process. A real-world example is used to illustrate the
proposed approach. The applications of the resulting model are also discussed. As
such, the main contributions of this paper are: (a) a new method to dynamically
update the scale parameter of the power-law model and (b) a graphical method to
show the relative health level of an item.
The paper is organized as follows. The proposed adaptive model is presented in
Sect. 2. Section 3 deals with the model parameter estimation. A numerical example
is presented in Sect. 4. The applications of the resulting model are discussed in
Sect. 5. The paper is concluded in Sect. 6.
2 Proposed Model
A wear process generally undertakes three phases: the rate of wear first increases
quickly, then roughly maintains in a constant and finally increases quickly again.
This implies that the mean wear curve is inverse S-shaped with an inflection point.
That is, the rate of wear is bathtub-shaped. Jiang [13] presents a three-parameter
model that is suitable for representing the whole wear process. Its mean value
function is given by
where ti0 ¼ 0, wi0 ¼ 0, and wij is the cumulative wear amount measured at time tij for
the ith wear process. The mean value function of W(t) is given by the power-law
model:
where β is the shape parameter and assumed to be unchanged; η is the scale para-
meter and can vary with time. Specifically, for time interval (ti0 , ti1), the scale
parameter is η0 for any wear process. At time tij, the j-th observation is obtained.
Based on this observation and Eq. (3), the scale parameter can be estimated as
1=β
sij ¼ tij =wij : ð5Þ
4
Mean value function
3
w(t)
Observed wear
2
0
0 10 20 30 40 50 60 70
t
ηij sij ¼ α η0 sij , ηij η0 ¼ ð1 αÞ sij η0 ð7Þ
The power-law mean value function represents non-linearity of the wear process.
For a given value of t, we assume that W(t) follows the gamma distribution with the
scale parameter v and shape parameter u(t)¼w(t)/v. The mean and variance of W(t)
are given respectively by [16]
Equation (8) indicates that the scale parameter reflects the variability of the
wear process.
The proposed non-homogeneous gamma process model has four parameters:
β , η0 , α and v. Since the scale parameter is a function of tij and wij, w(t) is state-
dependent and hence the increment is no longer independent.
We use the maximum likelihood method (MLM) to estimate the model parameters.
Consider the i-th wear path given by Eq. (3). The wear increment ΔWij ¼ Wij Wi ,
j 1 ( j ¼ 1 , 2 , . . .) is a random variable with mean given by
!β !β
tij ti, j1
μij ¼ E ΔW ij ¼ , ηi, 0 ¼ η0 : ð9Þ
ηi, j1 ηi, j1
An Adaptive Power-Law Degradation Model for Modelling Wear Processes 103
Under the assumption that the wear increment ΔWij follows the gamma distri-
bution with shape parameter uij ¼ μij/v and the scale parameter v, the likelihood
function for path i is given by
Y
ni
Li ¼ g Δwij ; uij ; v , Δwij ¼ wij wi, j1 ð10Þ
j¼1
X
n
ln ðLÞ ¼ ln ðLi Þ: ð11Þ
i¼1
4 Illustration
The data shown in Table 1 come from Bocchetti et al. [1], and deal with accumu-
lated wear amounts of 33 cylinder liners of 8-cylinder SULZER RTA 58 engines. In
the table, tij (in 1000 h) is thejth inspection time or operating age ( j ¼ 1, 2, . . ., ni)
for the ith liner (i ¼ 1, 2, . . ., 33), and wij is the corresponding accumulated wear
amount in mm. There is no wear record for the 32nd liner. The maximum admis-
sible wear is 4 mm. The 12th, 25th and 28th liners were replaced at the last inspec-
tion due to a large accumulated wear.
The data have been modelled using different models. For example, Bocchetti
et al. [1] fit the data to stochastic process model with a log-linear mean function and
a power-law mean function, respectively; Giorgio et al. [3] fit the data to three
different stochastic models, one of which is state-dependent homogeneous Markov
chain; Giorgio et al. [4] fit the data to a parametric Markov chain model, in which
the transition probabilities between states depend on both the current age and wear
level; Guida and Pulcini [5] model the data using a non-stationary inverse Gamma
process, which is state-dependent.
104 R. Jiang
The maximum likelihood estimates (MLEs) of the model parameters are shown in
the second row of Table 2; and the MLEs for case α ¼ 1 are shown in the last row.
From the table we have the following observations:
• The smoothing parameter α > 1, implying that the scale parameter is adjusted
towards the opposite direction of sij relative to η0. As mentioned earlier, this is
allowed.
An Adaptive Power-Law Degradation Model for Modelling Wear Processes 105
5 Applications
In this section, we present two applications of the resulting model. One deals with
prediction of time to the wear limit, and the other deals with evaluation of
relative health level.
The distribution function of time to the wear limit (L ¼ 4 mm) for the population is
given by
where G(x; u, v) is the gamma cdf evaluated at x. For the special case α ¼ 1, we have
Using Eq. (13) into Eq. (12), the cdf of time to the wear limit can be directly
evaluated, and the pdf can be numerically evaluated. For small value of Δt, the pdf
is approximated by
For the current example, the pdf evaluated by Eq. (14) is shown in Fig. 2 (the
solid line), and can be approximated by the gamma distribution with the parameters
shown in the second row of Table 3. Its plot is also shown in Fig. 2 (the dotted line).
For the general case α 6¼ 1, we first infer the time to the wear limit for each
process using the last observation before the process reaches the wear limit. The
inference is based on the mean value function (this can lead to an underestimate of
dispersion). This yields a sample of time to the wear limit (which is sometimes
106 R. Jiang
0.07
0.06 Pdf from pseudo
0.05 failure times
0.04
f(t)
0.03
Pdf from Eq.
0.02
(14) with α = 1
0.01
0
0 20 40 60 80 100 120
t
called pseudo-failure time). Then, we fit the sample to a standard distribution using
the MLM. The results are shown in the last three rows of Table 3. As seen, the
Weibull distribution provides the best fit in terms of log-likelihood. The fitted Wei-
bull distribution is also shown in Fig. 2.
A possible application of the population distribution is to determine a preventive
replacement age for the population. This needs a decision model such as the cost
model (when the cost parameters for preventive and failure replacements are
known) or the tradeoff BX life B∗ X (when the replacement costs are unknown, see
[17]). For the current example, the values of B∗ X from different models are shown in
the last column of Table 3. As seen, the values of B∗ X are very close to each other
though they come from different models.
We now apply the adaptive model to make the maintenance decision for Liners
12, 25 and 28. The time to the wear limit can be inferred using linear interpolation
or extrapolation based on the last two records. The results are shown in the
second row of Table 4. Actual maintenance types are shown in the third row.
The decision times td are shown in the fourth row. The values of scale parameter
are shown in the fifth row, and the times for the mean wear curves to reach the wear
limit are shown in the sixth row. As seen, the predicted failure times are close to the
actual failure times for Liners 12 and 28 but larger than the actual failure times for
Liner 25.
The seventh row shows the tradeoff BX lifetimes of individual liners. As seen
from the table, these liners would have been preventively replaced at B∗ X if the liners
were preventively replaced at age B∗ X . Of course, some useful lifetimes would
be lost.
An Adaptive Power-Law Degradation Model for Modelling Wear Processes 107
A
Wear limit
B
C
D
w(t)
B X* B X* m B X*
t
The actual maintenance decision is based on decision rules. For the current
example, the decision rules can be defined as below:
• Rule 1: a corrective replacement is immediately performed if the observed wear
is larger than or equal to the wear limit (see Point A in Fig. 3).
• Rule 2: a preventive replacement (PR) is immediately performed if the observed
wear is smaller than the wear limit and the time to the previously predicted PR
time (say B∗X ) is smaller than a certain critical value τm (e.g., 1000 h) (see Point B
in Fig. 3).
• Rule 3: a PR can be performed at B∗ X or opportunistically performed between
B∗ ∗
X τM and BX τm if the observed wear is smaller than the wear limit and the
time to the PR time is larger than τm and smaller than the inspection interval τM
(e.g., 5000 h) (see Point C in Fig. 3).
• Rule 4: the next inspection is scheduled if the observed wear is smaller than the
wear limit and the time to the PR time is larger than τM (see Point D in Fig. 3).
108 R. Jiang
The power-law model given by Eq. (4) can be linearized under the following
transformations:
Since η varies with time and can be different for different wear processes,
parameter A is a random variable.
Applying the transformations given by Eq. (15) to the data in Table 1, the
observed data can be transformed as
xij ¼ ln tij , yij ¼ ln wij : ð17Þ
Using Eq. (17), the data and resulting model can be visualized. Figure 4 shows
the plot of the transformed data. As seen, the relationship between x and y is roughly
linear, implying that the power-law model can be appropriate for fitting the data.
To evaluate the relative health level of a specific liner, we may build equal-
health lines on x-y plane [19]. An h-equal-health line is given by
The straight line yh(x) divides all the m data points into two parts: those data
points that are above the line and the other data points that are on or below the line.
Let k [m k] denote the number of data points that are above [on or below] the line.
The relative health level is defined as h ¼ (k 0.3)/(m + 0.4). If an observation point
falls on an h-equal-health line, its relative health level is h. The larger h, the
healthier the process.
An Adaptive Power-Law Degradation Model for Modelling Wear Processes 109
2
y=ln(L)
h=0.3
0
y
0 1 2 3 4 5
h=0.1
-1
h=0.7
h=0.9 h=0.5
-2
x
0.8
0.6
G(a)
0.4
0.2
0
-3 -2.5 -2 -1.5 -1 -0.5 0
a
0.8
0.6
h
0.4
0.2
0
1 3 5 7 9 11 13 15 17 19 21 23 25 27 29 31 33
Liner
Liner 13
3
w
2
Liner 19 Liner 17
1
0
0 5 10 15 20 25 30 35
t
1 X ni
hi ¼ hij : ð20Þ
ni j¼1
Figure 6 displays the relative health levels of wear processes of all the liners.
As seen from the figure, the three worst processes are from Liners 13, 17 and 19.
Their wear processes are shown in Fig. 7. As seen, they have similar wear trends
and high initial wear rates. Therefore, the special attention should be paid on them.
An Adaptive Power-Law Degradation Model for Modelling Wear Processes 111
6 Conclusions
Acknowledgments The research was supported by the National Natural Science Foundation of
China (No. 71371035).
References
1. Bocchetti D, Giorgio M, Guida M, Pulcini G (2009) A competing risk model for the reliability of
cylinder liners in marine diesel engines. Reliab Eng Syst Saf 94:1299–1307
2. van Noortwijk JM (2009) A survey of the application of gamma processes in maintenance.
Reliab Eng Syst Saf 94:2–21
3. Giorgio M, Guida M, Pulcini G (2009) Stochastic processes for modeling the wear of marine
engine cylinder liners. In: Erto P (ed) Statistics for innovation. Springer, Berlin, pp 213–230
4. Giorgio M, Guida M, Pulcini G (2010) A parametric Markov chain to model age- and state-
dependent wear processes. In: Mantovan P, Secchi P (eds) Complex data modelling and
computationally intensive statistical methods. Springer, New York, pp 85–97
5. Guida M, Pulcini G (2013) The inverse Gamma process: a family of continuous stochastic
models for describing state-dependent deterioration phenomena. Reliab Eng Syst Saf 120:
72–79
6. Laurenciu NC, Cotofana SD (2013) A nonlinear degradation path dependent end-of-life
estimation framework from noisy observations. Microelectron Reliab 53(9-11):1213–1217
7. Si XS (2015) An adaptive prognostic approach via nonlinear degradation modeling: appli-
cation to battery data. IEEE Trans Ind Electron 62(8):5082–5096
8. Xu W, Wang W (2012) An adaptive gamma process based model for residual useful life pre-
diction. In: 2012 Prognostics and system health management conference (PHM-2012 Beijing),
MU3019
9. Do P, Voisin A, Levrat E, Iung B (2015) A proactive condition-based maintenance strategy
with both perfect and imperfect maintenance actions. Reliab Eng Syst Saf 133:22–32
10. Fouladirad M, Grall A (2011) Condition-based maintenance for a system subject to a
non-homogeneous wear process with a wear rate transition. Reliab Eng Syst Saf 96:611–618
11. Li H, Deloux E, Dieulle L (2016) A condition-based maintenance policy for multi-component
systems with Lévy copulas dependence. Reliab Eng Syst Saf 149:44–55
12. Zhu W, Fouladirad M, Bérenguer C (2015) Condition-based maintenance policies for a
combined wear and shock deterioration model with covariates. Comput Ind Eng 85:268–283
112 R. Jiang
13. Jiang R (2010) Optimization of alarm threshold and sequential inspection scheme. Reliab Eng
Syst Saf 95:208–215
14. Guida M, Pulcini G (2009) Reliability analysis of mechanical systems with bounded and
bathtub shaped intensity function. IEEE Trans Reliab 58(3):432–443
15. Gardner ES (2006) Exponential smoothing: the state of the art – part II. Int J Forecast 22(4):
637–666
16. Jiang R (2015) Introduction to quality and reliability engineering. Springer, Berlin
17. Jiang R (2013) A tradeoff Bx life and its applications. Reliab Eng Syst Saf 113:1–6
18. Zhang X, Zeng J (2015) A general modeling method for opportunistic maintenance modeling of
multi-unit systems. Reliab Eng Syst Saf 140:176–190
19. Jiang R, Jardine AKS (2008) Health state evaluation of an item: a general framework and
graphical representation. Reliab Eng Syst Saf 93(1):89–99
A Comparison Study on Intelligent Fault
Diagnostics for Condition Based Maintenance
of High-Pressure LNG Pump
Abstract High Pressure Liquefied Natural Gas Pump (HP-LNG Pump) is impor-
tant equipment in the LNG receiving terminal process, and plays a key role in
determining the total supply capacity of natural gas in LNG terminal. Therefore, the
condition monitoring, fault diagnostics and prognostics technologies are applied to
implement the CBM (Condition Based Maintenance) strategy for HP-LNG pumps,
which can support for appropriate maintenance decision. Currently, a number of
valuable diagnostic models and methods have been proposed in machine diagnos-
tics. However, most intelligent diagnostic models have generally validated using an
experimental data, and still not focus on industrial real data because of the complex
nature in real industry. In this paper, intelligent fault diagnostic performances using
three conventional classification algorithms such as SVM (Support Vector
Machines), k-NN (k-Nearest Neighbor) and Fuzzy-ARTMAP have been evaluated
for the CBM decision of HP-LNG pump. Comparative results indicate that the fault
classification using SVM provides more accurate performance compared to other
classification methods for HP-LNG pump.
1 Introduction
The LNG receiving terminal receives Liquefied Natural Gas (LNG) from LNG carrier
ships, stores the liquid in special storage tanks, vaporizes the LNG and then delivers the
natural gas through distribution pipelines. The receiving terminal is designed to deliver
a specified gas rate into a distribution pipeline and to maintain a reserve capacity of
LNG. LNG takes up six hundredths of the volume of natural gas at/or below the boiling
temperature (162 C), which is used for storage and easy transportation [1].
Figure 1 shows key facilities and process overview in an LNG receiving terminal.
As shown in Fig. 1, the unloaded LNG from vessels is transported to ground storage
tanks via pipeline using cargo pumps on the LNG carrier vessel. In an LNG receiving
terminal, primary cryogenic pumps that are installed in the storage tanks which
supply the LNG to HP-LNG pumps with pressure around 8 bar. The HP-LNG pumps
boost the LNG pressure to around 80 bar for evaporation and delivery of the highly
compressed natural gas via a pipeline network across the nation.
CBM is a management philosophy that posits repair of replacement decisions on
the current or future condition of assets; it recognizes that change in condition
and/or performance of an asset is the main reason for exciting maintenance. Thus,
Horner et al. noted that the optimal time to perform maintenance is deter-mined
from actual monitoring of the asset, its subcomponent, or part. They also noted that
condition assessment varies from simple visual inspections to elaborate automated
inspections using a variety of condition monitoring tools and techniques [2].
In the history of development of machine diagnostics, signal processing method
has become a central issue for CBM technology. However, most studies in the field
of diagnostics have only focused on the experimental data given artificial fault.
Hwang [3] has shown a novel approach for applying the fault diagnosis of rotating
machinery. To detect multiple faults in rotating machinery, a feature selection
method and support vector machine (SVM) based multi-class classifier are
constructed and used in the faults diagnosis. Kim et al. [4] has developed a new
fault detection method based on vibration signal for rotor machinery. Ahn et al. [5]
has shown that signal processing and experiment under AE system is performed to
evaluate the performance complex bearing fault classification.
In this paper, the comparative evaluations for intelligent fault diagnostics of
HP-LNG pump have been performed using real industrial data including various
A Comparison Study on Intelligent Fault Diagnostics for Condition Based. . . 115
vibration signal processing techniques such as Hilbert Transform (HT) and Discrete
Wavelet Transform (DWT) before extracting of dominant fault features from time
domain and frequency domain signals. In other to estimate the abilities of fault
classification techniques, three methods such as SVM (Support Vector Machine),
k-NN (k-Nearest Neighbor) and Fuzzy-ARTMAP are used. To extract useful
features, Distance Evaluation Technique (DET) has been applied in this study.
This paper is organized as follows: Sect. 2 briefly introduces the HP-LNG pump
in LNG terminal. Section 3 describes the pump fault data used in this study through
historical failure data analysis of HP-LNG pump. The proposed fault diagnostics
methodology including feature extraction/selection techniques is explained in Sect.
4. Finally comparative evaluation results of intelligent fault diagnostics have been
addressed in Sect. 5.
2 HP-LNG Pump
In LNG receiving terminal, HP-LNG pumps play a key role in determining the total
supply capacity of natural. These HP-LNG pumps are submerged and operate at
super-cooled temperatures. They are self-lubricated at both sides of the rotor shaft
using LNG. Due to the low viscous value (about 0.16 cP) of LNG, the two bearings
of the HP-LNG pump are poorly lubricated and bearing must be specially designed.
Therefore, vibration and noise of HP-LNG pumps are regularly monitored and
managed based on CBM (Condition based maintenance) techniques. Normally
condition monitoring and fault diagnostic technology are applied for HP-LNG
pump to evaluation of machine health and early detection before the catastrophic
pump failures [1].
Table 1 shows the pump specifications. HP-LNG pump is installed in pump
vessel with integrated shaft of the pump and motor. Two ball bearings are installed
at top and bottom side of motor to support the entire dynamic load. The HP-LNG
pump schematic, vibration measuring points are presented in the Fig. 2.
Vibration data are generally collected using two different condition monitoring
techniques. One is on-line monitoring using two accelerometers installed motor
housing and the other is off-line monitoring method using potable vibration data
collection device from pump top plate.
Rubbing 35%
Shaft Bending 9%
In this research, the historical fault data have been analyzed to identify the main
fault types of HP-LNG pump by using over 20 year’s maintenance records. Figure 3
shows the statistical analysis result of HP-LNG pump faults. From this result, three
predominant faults have been defined as main faults of HP-LNG pump during pump
operation such as bearing fault, rubbing of wearing part and rotor bar fault as shown
in Fig. 3.
To estimate the proposed methodology, the historical condition monitoring data
collected on top plate points (Horizontal, Vertical, Axial) from nine HP-LNG
A Comparison Study on Intelligent Fault Diagnostics for Condition Based. . . 117
pumps are used in this work. As shown in Table 2, total 80 vibration samples of four
types (Rotor Bar Fault, Rubbing, Bearing Fault and Normal Condition) are obtained
as described in Sect. 2. Each fault condition has 324 features which have calculated
from frequency domain and time domain vibration data.
In this paper, the proposed intelligent fault diagnostics for HP-LNG pump follows
the typical procedure of intelligent fault diagnosis consisting of condition monitor-
ing, signal processing, feature extraction and fault classification. The conventional
feature-based diagnostics framework is illustrated in Fig. 4.
For signal processing of vibration monitoring data, mainly two pre-processing
techniques such as DWT and HT have been employed for this research because
these two methods has been extensively used in vibration signal analysis for fault
conditions of machine. The statistical parameters form vibration time and fre-
quency domain were calculated after pre-processing of vibration signals. Total
324 features from each condition of pump were calculated and used in this study.
For outstanding performance of fault classification and reduction of computational
effort, effective features for pump fault conditions were selected using the DET to
evaluate the employed classification algorithms.
Fig. 4 Conventional
feature-based diagnostics
framework
1 XN
d i, j ¼ pi, j ðmÞ pi, j ðnÞ ð1Þ
N ðn 1Þ m, n¼1
1 XM
d0i, j ¼ pai, m pai, n ð2Þ
M ðM 1Þ m, n¼1
0
αi ¼ d ai =dai ð3Þ
The total 24 features were selected using distance evaluation technique (DET),
these features were mainly made up of more dynamic and variable parameters such
as mean square frequency (f14), root variance frequency (f17) and mean average
deviation frequency (f18) among number of features which were calculated after
signal processing techniques such as DET and HT from pump vibration data.
4.3.1 SVM
Several methods have been proposed regarding Multi Class Support Vector
Machines (MCSVMs) such as one-against-one (OAO), one-against-all (OAA)
and one-acyclic-graph (OAG). Among these methods, OAA method is employed
in this research because this method is a simple and effective for multi-class
classification, and has been successively demonstrated with its high classification
performance. In addition, selection of appropriate kernel function is very important
to classify in feature space. In various kernel functions, the radial basis function
(RBF) is the most popular kernel type for fault diagnostics because it finds a set
weights for a curve fitting problem in complex faults feature set as stated below:
4.3.2 k-NN
Fig. 5 Comparison result of SVM, k-NN and Fuzzy-ARTMAP for fault diagnostics
6 Conclusions
In this paper, three classification techniques for the intelligent fault diagnostics of
HP-LNG pump were evaluated using real industrial data. For signal processing of
vibration monitoring data, mainly two pre-processing techniques such as Discrete
Wavelet Transform and Hilbert Transform have been employed in this research. To
improve classification performance, distance evaluation technique (DET) has been
used to select the features that reflect the dynamic characteristics of each fault
condition.
The comparison results indicate that SVM has a better capability for HP LNG
pump fault diagnostics than Fuzzy-ARTMAP and k-NN relatively. Especially,
SVM provides more accurate performance for fault conditions such as rotor bar
fault(RBF) and bearing fault(BF) compared to rubbing(RB) and normal condition
(NOR) for the intelligent fault diagnostics of HP-LNG pump.
Acknowledgments This research was conducted by support of Korea Gas Technology Corpora-
tion (KOGAS-Tech).
122 H.-E. Kim and T.-H. Jeon
References
1. H. E. Kim and J. S. Oh (2014) Comparative study of intelligent fault diagnostics for LNG pump
failure. In: WCEAM 2014
2. Byron A. Ellis (2008) Condition based maintenance. In: TJP, November 10, 2008, pp 1–5
3. Yu W, Hwang W (2003) Fault diagnosis of machinery using multi-class support vector
machines. In: KSNVE annual fall conference 2003, pp 537–543
4. Kim DH, Shon SM, Kim YW, Bea YC (2014) Rotating machinery fault diagnosis method on
prediction and classification of vibration signal. In: KSNVE annual fall conference 2014, pp
90–93
5. Ahn BH, Kim YH, Lee JM, Ha JM, Choi BK (2014) Image fault classification for AE complex
fault signal. In: KSNVE annual spring conference, pp 417–419
6. Huang HZ, Qu J, Zuo MJ (2009) Genetic-algorithm-based optimal apportionment of reliability
and redundancy under multiple objectives. IEEE Trans 41:287–298
7. Knerr S, Personnaz L, Dreyfus G (1990) Single-layer learning revisited: a stepwise procedure
for building and training a neural network. Neurocomputing 68:41–50
8. Yang BS, Widodo A (2010) Introduction of intelligent machine fault diagnosis and prognosis.
Nova Science, New York
9. Carpenter GA, Grossberg S, Markuzon N, Reynolds JH, Rosen DB (1991) Fuzzy ARTMAP: a
neural network architecture for incremental supervised learning of analog multidimensional
maps. IEEE Trans Neural Netw 3:698–713
Rotating Machine Prognostics Using
System-Level Models
Abstract The prognostics of rotating machines is crucial for the reliable and safe
operation as well as maximizing usage time. Many reliability studies focus on
component-level prognostics. However, in many cases, the desired information is
the residual life of the system, rather than the lifetimes of its constituent compo-
nents. This review paper focuses on system-level prognostic techniques that can be
applied to rotating machinery. These approaches use multi-dimensional condition
monitoring data collected from different parts of the system of interest to predict the
remaining useful life at the system level. The working principles, merits and
drawbacks as well as field of applications of these techniques are summarized.
1 Introduction
Rotating systems such as gas turbines and compressors are widely used due to their
high performance and robustness. However, many kinds of failures may occur
during the operation of the machine. Those failures will cause unplanned downtime
and economic losses as well as reduced reliability. One way to minimize the
negative influence of these failures is to make maintenance strategies more predic-
tive by using automated condition monitoring. Condition-based maintenance
(CBM) is a preventive maintenance strategy that seeks to improve the reliability
of engineering systems based on condition monitoring information [1]. CBM
enables the diagnosis of impending failures and the prognosis of the future health
state and remaining useful life (RUL) of a system [2]. Prognostic programs based
on condition monitoring data provide a potent tool for practitioners in making
The prognostic approaches reviewed in this paper can be divided into three cate-
gories: (1) statistical methods, (2) artificial intelligence methods, and (3) similarity-
based methods (see Fig. 1).
For statistical methods, we review models based on Bayesian theory and pro-
portional hazard models. These models predict the RUL based on past observed
information and statistical models in a probabilistic manner. Therefore, a probabil-
ity density function (PDF) of the RUL is formulated for uncertainty management.
Moreover, in statistical methods, machine lifetime data might be required in
addition to monitoring data for failure prediction.
When lifetime data is scarce or non-existent, artificial intelligence methods that
make predictions using only monitoring data can be considered. These models can
be treated as a nonlinear function approximator, which aims to determine depen-
dencies in a training data set such that predictions of the outputs (e.g., RUL) can be
made when new inputs are available. Unlike statistical methods, most artificial
intelligence methods do not provide a PDF of the RUL. For artificial intelligence
methods, we review models based on neural networks and support vector machines
(SVMs).
Similarity-based methods are fundamentally different from the techniques in the
first two categories because they do not perform trending or extrapolation of the
degradation process. Instead, they construct a probabilistic health indicator, which
characterizes the system health state via trajectories. Then, predictions are made
based on evaluating similarities between the trajectories.
The working principles, merits, drawbacks and the applications of these tech-
niques are discussed in the following sections.
We assume the system model based on Bayesian theory can be defined as:
Where xt refers to the unobservable health state of the system under study at time t,
yt refers to the observed information at time t. p(xt| xt 1) is the state equation, and p
(yt| xt) is the observation equation. Then, the prognostic tasks can be divided into
two sequential stages: the estimation stage and the prediction stage. The purpose of
the estimation stage is to find a state estimate p(xt| Yt) given the observation history
up to time t, where Yt denotes the observation history up to time t. The prediction
problem is to predict the health state p(xt + k| Yt) at a given prediction time t þ k. The
RUL can be computed by extrapolating the predicted health state to a pre-set failure
threshold [4].
Because most engineering facilities in practice undergo a non-linear and time-
varying deterioration process, the parameters of the prediction model cannot remain
constant over the entire prediction process (otherwise the predictions might be not
accurate). Therefore, Bayesian updating methods are often adopted to jointly
update the model parameters and the system health state when new observations
are available. Three widely used Bayesian updating tools are particle filtering
[4, 17], semi-stochastic filtering [18] and dynamic Bayesian updating [19]. To
address systems with multiple sensory inputs, sensor fusion techniques, such as
principal component analysis [20], independent component analysis (ICA) (Wang
and Hussin [21]), linear regression [4] and path model [19] are often employed to
merge the multivariate measurements into a one-dimensional variable—the system
health indicator. Then the equations defined in the first paragraph can be used to
map health indicator values to the RUL. Some common Bayesian theory techniques
will be discussed below.
1. Particle Filtering: The main concept of particle filtering is to represent the
required posterior distribution of the state variables, such as p(xt| Yt), using a
set of particles with associated weightings. These particles evolve and adapt
recursively when new information becomes available [22]. Then, the
unobserved health state is estimated based on these weighted samples.
In health state prediction, particle filter has three advantages: (a) it can be applied to
nonlinear process with non-Gaussian noise; (b) it provides probabilistic results
which is helpful to manage the prognostic uncertainties; (c) it allows information
fusion such that data from multidimensional sensors could be employed collec-
tively [23, 24]. However, one limitation of particle filtering is that a large
number of samples might be required to accurately approximate the future
state distributions, which may cause the filtering system to collapse. A good
Rotating Machine Prognostics Using System-Level Models 127
prefer this type of estimate because it can avoid the loss of accuracy caused by the
assumption of a parametric distribution [32]. However, in practice, the baseline
hazard function is often assumed to be a parametric distribution, such as the
Weibull or exponential distribution [11]. Such assumptions might not be reasonable
in many cases because of the confusing effects of different covariates [31].
To apply the PHMs, the sensory measurements and lifetime data are combined to
fit the model. The PHM can identify the important risk factors from all input
variables and their relative influence on the failure of the equipment [29]. Then,
the system failure distribution at time service t can be estimated. Finally, the failure
time at time t can be predicted according to the estimated probability distribution.
PHMs have been applied to many non-linear and non-stationary machinery
prognostic problems. Jardine et al. [33] developed a PHM and employed it to
estimate the remaining useful life of aircraft engines and marine gas turbines. The
baseline hazard function was assumed to be a Weibull distribution and was esti-
mated using lifetime data. The levels of various metal particles (such as Fe, Cu and
Mg) in the oil were used as the covariates in both cases. The influence of the
condition monitoring variables on the equipment RUL has been properly
interpreted by the developed PHM. The authors also applied the PHM to estimate
the RUL and optimize the maintenance decisions of haul truck wheel motors in
[29]. The key covariates related to failures were identified from 21 monitored oil
analysis variables using the developed PHM. The results showed that significant
savings in maintenance costs could be achieved by optimizing the overhaul time as
a function of lifetime data and oil analysis variables.
The above models are based on the assumption that the system under study is
subject to a single failure mode. However, in practice, most complex mechanical
systems consist of multiple sub-systems with various failure modes [28]. Therefore,
prognostic models for determining only one type of failure mode cannot properly
estimate the overall system failure time. Recently, Zhang et al. [28] proposed a
mixed Weibull proportional hazard model (MWPHM) for complex mechanical
system reliability assessment. In this model, the overall system failure probability
density is determined by a mixture of failure densities of various failure modes. The
influences of multiple monitoring signals on different failure modes are integrated
using the maximum likelihood estimation algorithm. Real data from a centrifugal
water pump were combined with lifetime data to test the robustness of the model.
The main problem with applying PHMs for failure prediction is that they require
a large amount of lifetime data to determine the parameters of the baseline hazard
function and the weighting of covariates [27], which may limit the applications of
PHMs because the amount of lifetime data might be insufficient in many cases due
to various reasons, such as missing or no records or transcription mistakes
[34]. Another drawback of PHMs is that they rely on the choice of the failure
threshold for RUL prediction. The threshold must be continuously updated when
system maintenance is conducted [28].
130 X. Li et al.
Artificial neural networks (ANNs) have recently been widely used in modelling
degradation processes. An ANN is a computing system that is able to capture,
represent and compute mapping from the input multi-variable space to the output
space [35]. ANNs have three layers: an input layer, one or more hidden layers and
an output layer. ANNs are comprised of a large number of processing elements
(known as neurons) that are connected to each other by weighted interconnections
[36]. These neurons are organized into distinct layers, and their interconnections are
determined through a training process. The network training involves presenting
data sets collected from the degradation process. Then the network parameters are
adjusted to minimize the errors between the model output and desired output
[35]. Once the training is finished, ANNs process new input data to make pre-
dictions about the outputs (RUL).
Network architectures that have been used for prognostics can be classified into
two types: feed-forward and recurrent networks [37]. In feed-forward networks, the
signals flow in one direction; therefore, the inputs to each layer depend only on the
outputs of the previous layer. However, applications in signal processing and
prognostics should consider the system dynamics. Recurrent networks provide an
explicit dynamic representation by allowing for local feedbacks [38]. Two types of
networks: multi-layer perceptron (MLP) and recurrent neural networks (RNNs)
(Fig. 2 shows the architecture of a simple RNN), which have been applied exten-
sively by researchers, will be discussed below [35].
1. Multi-Layer Perceptron (MLP): MLPs are one of the most popular feed-forward
neural networks used for prognosis [11]. MLPs utilize back-propagation
(BP) learning technology for training. After the training, the MLP is capable
of classifying the fault and predicting the RUL based on new measurements
ANNs can represent and build mappings from experience and history measure-
ments to predict the RUL and then adapt it to unobserved situations. The strong
learning and generalization capabilities of ANNs render them suitable for model-
ling complex processes, particularly systems with nonlinear and time-varying
dynamics [41]. ANNs are superior in capturing and depicting relationships between
many variables in high-dimensional data space [42, 43]. RNNs are suitable for
approximating dynamic dependencies [40]. These distinct characteristics make
ANNs promising candidates for modelling degradation processes in rotating
machinery.
Xu et al. [44] successfully employed RNNs, support vector machines (SVMs)
and Dempster-Shafer regression to estimate the RUL of an aircraft gas turbine.
Echo state network (ESN), which is a variant of the RNNs, was employed by Peng
et al. [45] to predict the RUL of engines using NASA repository data. The results
indicated that the ESN significantly reduced the computing load of the traditional
RNNs. ANNs have also been used in combination with Kalman filters and Extended
Kalman filters in [46] and (Felix [47]) to perform failure predictions of aircraft
engines.
Although ANNs have been shown the superior power in addressing complex
prognostic problems which have multivariate inputs, there are some limitations. For
example, the majority of the ANN based prognostic models aim to assume a single
failure mode and do not relate lifetime data with the machine RUL. Moreover, the
models rely on a large amount of data for training. The prognostic accuracy is
closely dependent on the quality of the training data [44]. Furthermore, ANNs allow
for few explanatory insights into how the decisions are reached (also known as the
black box problem), which has become a concern to modelers because the causal
relationship between the model variables is essential for explaining the fault
evolution [48]. Attempts to solve the black box problem can be found in [49]. More-
over, ANNs lack a systematic approach to determine the optimal structure and
parameters of the network to be established [12, 13].
132 X. Li et al.
Previously, SVMs were mainly used for pattern recognition problems and have not
been used for time series forecasting until the introduction of the Vapniks insensi-
tive loss function [8]. SVM-based machine learning starts with a number of input
variables x(i) , i ¼ 1 , 2 , 3 , . . . , N, and the corresponding target values y(i) ,
i ¼ 1 , 2 , 3 , . . . , N. The idea is to learn the dependency of y(i) on x(i) and to define
a function over the input variables. Then, predictions of y(i) can be made given
unseen x(i) [50]. When applying SVMs to nonlinear prognostics, model inputs are
first mapped onto a higher dimensional feature space by using a kernel function.
The most commonly employed kernel function is the radial-based function (RBF)
[51]. Then a linear model is constructed in the feature space to make estimation.
Figure 3 shows the architecture of a simple SVM based prognostic model. SVMs
are excellent in addressing prognostic problems regarding complex rotating
machinery because they have no limitations on the dimensionality of the input
vectors and have relatively a low computational burden [12, 13]. Besides, SVMs
can achieve highly accurate results with nonlinear inputs.
Several different prognostic models based on SVMs have been used in model-
ling nonlinear and non-stationary dynamic systems.
1. Relevance Vector Machine (RVM): RVM has an identical function form as a
SVM. Hu and Tse [52] proposed a model based on RVM for RUL prediction of a
pump. This model was proved to be accurate when dealing with non-stationary
vibration signals.
2. Particle Swarm Optimization and SVM (PSO-SVM): Garcı́a Nieto et al. [53]
developed an RUL framework based on the PSO-RBF-SVM technique. This
model combines a SVM with particle swarm optimization (PSO) to enable the
parameter adjustment of the RBF kernel function. The results show that the
the monitoring data collected from a pump with six distinct degradation modes.
The normalized cross correlation was employed to determine the optimal
matching trajectory segments, which were then used to estimate the RUL.
3. Similarity Model Based on PCA and K-NN: Mosallam et al. [62] employed
principle component analysis (PCA) and an empirical mode decomposition
(EMD) algorithm to construct health indicators from turbofan engine deteriora-
tion simulation data. Then, the k-nearest neighbor (K-NN) classifiers were used
to determine the most similar HIs for RUL prediction.
4. Similarity Model Based on Belief Functions: An improved technique based on
belief functions was proposed by Ramasso and Gouriveau [60] and Ramasso
et al. [63]. In this method, the authors only match the last points of the
trajectories in the library with tested ones because the last points are more likely
to be closely related to the degradation state. One of the main contributions of
this method is its ability to manage labels (indicating degradation states) that
would have been incorrectly assigned to the sensory data.
5. Similarity Model Based on Linear Regression and Kernel Smoothing: Wang
et al. [56] proposed a prognostic model in which the health indicator is
constructed from multiple sensors using linear regression. The best matching
instances were selected by examining the Euclidean distance between the test
and stored instances. This method was applied to data provided by the 2008
PHM Data Challenge Competition to predict the RUL of an unspecified system.
6. Similarity Model Based on Regression Vector Machine (RVM): Wang et al. [64]
improved the previous model by incorporating the uncertainty information into
the RUL estimation. The degradation curves of health indexes were estimated
using RVM. The Challenge data was employed again to test the effectiveness of
this method.
7. Hybrid Similarity Model: Hu et al. [65] proposed an ensemble prognostic model
which combines five individual algorithms (i.e. similarity-based approach with
the SVM, RVM and Exponential fitting, a Bayesian linear regression with the
quadratic fitting and a RNN) with a weighted sum formulation. The integrated
model shows higher accuracy in RUL prediction compared to any single
algorithm.
Fig. 5 Type of data used in Gas turbine engines All types of reviewed rotating machines
studies regarding (a) Gas
turbine engines only (blue),
16
and (b) All kinds of 14
reviewed rotating machines
Number of publications
(orange) 12
10
8
6
4
2
0
Simulated data Field data
studies regarding (a) gas turbine engines only and (b) all types of reviewed rotating
machines.
According to the reviewed articles, the RUL estimate of gas turbine engines is
the main application field. However, the proportion of studies using simulation data
is higher than those using field data. The reason for this is the simplicity of using
simulation programs and the difficulty of obtaining sufficient field data from
operating machines.
Most existing prognostic techniques were originally developed for a single failure
mode. To predict RUL for systems with multiple failure modes, several models
must be separately constructed for each failure mode. For example, Daigle et al. [6]
developed a distributed method for failure prediction of a four-wheel rover. This
method first decomposes the system-level prognostic problem into independent
sub-failure problems through structural model decomposition. Thereafter, the
Kalman filter and physical model are used to perform individual failure prognostics.
Finally, the local prognostic results are merged to form a system-level result.
However, the correlation between different failures may be overlooked via this
approach. To solve this problem, several frameworks for reliability analysis in
general engineering systems with competitive multiple failures have been pro-
posed. Ahmad et al. [66] developed a failure analysis approach by integrating the
failure mode effect and criticality analysis (FMECA) and PHM. This method was
validated in a cutting process system with two failure modes. FMECA was applied
to classify the censored and uncensored data based on the severity of the different
failure modes. The FMECA output was then used in PHM to determine the
138 X. Li et al.
reliability of the system. Huang and Askin [67] proposed a method to analyze the
reliability of an electronic device with two types of competing failure modes. Based
on the competing failure rule, the mean time-to-failure of the device was estimated
by jointly considering the failure rate of both failure modes. Bichon et al. [68]
proposed a surrogate-based approach based on the Gaussian process model and
physical laws. This method was used to analyze the failure probability of a liquid
hydrogen tank with three failure modes. However, most of the frameworks
discussed above have not yet been used to predict the RUL for rotating machinery.
Therefore, efforts can be made to extend them to rotating machine prognostics.
5 Conclusions
This paper has explored prognostic models for predicting the remaining useful life
of rotating machines at the system level. The reviewed prognostic models make
predictions based on the multi-dimensional condition monitoring signals collected
from the sensors distributed over the studied system. The relevant theories were
discussed, and the advantages and disadvantages of the main prognostic model
classes were explored. Examples were given to explain how these approaches have
been applied to predict the RUL of rotating systems. The reviewed approaches
generally require a large amount of historical data (condition monitoring data or
lifetime data) to obtain accurate estimates. In addition, the implementation of the
reviewed models in industry is still in the nascent stage and more work should be
conducted to apply them in real-world operating machines. Moreover, most of the
reviewed techniques were originally designed for a signal failure mode. Therefore,
several frameworks for reliability analysis of general engineering systems with
multiple failure modes were examined. In the future, more work should be
conducted to apply these frameworks to rotating machinery prognostics.
References
29. Jardine AKS, Banjevic D, Wiseman M, Buck S, Joseph T (2001) Optimizing a mine haul truck
wheel motors’ condition monitoring program use of proportional hazards modeling. J Qual
Maint Eng 7(4):286–302
30. Cox RD (1972) Regression models and life tables (with discussion). J R Stat Soc 34:187–220
31. Bendell A (1985) Proportional hazards modelling in reliability assessment. Reliab Eng 11
(3):175–183
32. Li Z, Zhou S, Choubey S, Sievenpiper C (2007) Failure event prediction using the Cox
proportional hazard model driven by frequent failure signatures. IIE Trans 39(3):303–315
33. Jardine AKS, Anderson PM, Mann DS (1987) Application of the weibull proportional hazards
model to aircraft and marine engine failure data. Qual Reliab Eng Int 3(2):77–82
34. Tsang AHC, Yeung WK, Jardine AKS, Leung BPK (2006) Data management for CBM
optimization. J Qual Maint Eng 12(1):37–51
35. Rafiq MY, Bugmann G, Easterbrook DJ (2001) Neural network design for engineering
applications. Comput Struct 79(17):541–552
36. Rodrı́guez JA, Hamzaoui YE, Hernández JA, Garcı́a JC, Flores JE, Tejeda AL (2013) The use
of artificial neural network (ANN) for modeling the useful life of the failure assessment in
blades of steam turbines. Eng Fail Anal 35:562–575
37. Atiya AF, El-Shoura SM, Shaheen SI, El-Sherif MS (1999) A comparison between neural-
network forecasting techniques-case study: river flow forecasting. IEEE Trans Neural Netw 10
(2):402–409
38. Gençay R, Liu T (1997) Nonlinear modelling and prediction with feedforward and recurrent
networks. Phys D 108(1–2):119–134
39. Ahmadzadeh F, Lundberg J (2013) Remaining useful life prediction of grinding mill liners
using an artificial neural network. Miner Eng 53:1–8
40. Liu J, Djurdjanovic D, Ni J, Casoetto N, Lee J (2007) Similarity based method for manufactur-
ing process performance prediction and diagnosis. Comput Ind 58(6):558–566
41. Senjyu T, Takara H, Uezato K, Funabashi T (2002) One-hour-ahead load forecasting using
neural network. IEEE Trans Power Syst 17(1):113–118
42. Wang S (2003) Application of self-organising maps for data mining with incomplete data sets.
Neural Comput Appl 12(1):42–48
43. Zhang S, Ganesan R (1997) Multivariable trend analysis using neural networks for intelligent
diagnostics of rotating machinery. J Eng Gas Turbines Power 119(2):378–384
44. Xu J, Wang Y, Xu L (2014) PHM-oriented integrated fusion prognostics for aircraft engines
based on sensor data. Sensors J 14(4):1124–1132
45. Peng Y, Wang H, Wang J, Liu D, Peng X (2012) A modified echo state network based
remaining useful life estimation approach. IEEE, Denver
46. Peel L (2008) Data driven prognostics using a Kalman filter ensemble of neural network
models. IEEE, Denver
47. Felix OH (2008) Recurrent neural networks for remaining useful life estimation. IEEE, Denvor
48. Olden JD, Jackson DA (2002) Illuminating the “black box”: a randomization approach for
understanding variable contributions in artificial neural networks. Ecol Model 154
(1–2):135–150
49. Sussillo D, Barak O (2013) Opening the black box: low-dimensional dynamics in high-
dimensional recurrent neural networks. Neural Comput 25(3):626–649
50. Saha B, Goebel K, Christophersen J (2009) Comparison of prognostic algorithms for estimat-
ing remaining useful life of batteries. Trans Inst Meas Control 31(3–4):293–308
51. Huang H-Z, Wang H-K, Li Y-F, Zhang L, Liu Z (2015) Support vector machine based
estimation of remaining useful life: Current research status and future trends. J Mech Sci
Technol 29(1):151–163
52. Hu J, Tse P (2013) A relevance vector machine-based approach with application to oil sand
pump prognostics. Sensors 13(9):12663–12686
Rotating Machine Prognostics Using System-Level Models 141
53. Garcı́a Nieto PJ, Garcı́a-Gonzalo E, Sánchez Lasheras F, de Cos Juez FJ (2015) Hybrid PSO–
SVM-based method for forecasting of the remaining useful life for aircraft engines and
evaluation of its reliability. Reliab Eng Syst Saf 138:219–231
54. Li X, Qian J, Wang G (2013) Fault prognostic based on hybrid method of state judgment and
regression. Adv Mech Eng 5(0):149562–149562
55. Liao L, Kottig F (2014) Review of hybrid prognostics approaches for remaining useful life
prediction of engineered systems, and an application to battery life prediction. IEEE Trans
Reliab 63(1):191–207
56. Wang T, Yu J, Siegel D, Lee J (2008) A similarity-based prognostics approach for remaining
useful life estimation of engineered systems. IEEE, Denver
57. Xue F, Bonissone P, Varma A, Yan W, Eklund N, Goebel K (2008) An instance-based method
for remaining useful life estimation for aircraft engines. J Fail Anal Prev 8(2):199–206
58. Baurle RA, Gaffney RL (2008) Extraction of one-dimensional flow properties from
multidimensional data sets. J Propuls Power 24(4):704–714
59. Malinowski S, Chebel-Morello B, Zerhouni N (2015) Remaining useful life estimation based
on discriminating shapelet extraction. Reliab Eng Syst Saf 142:279–288
60. Ramasso E, Gouriveau R (2014) Remaining useful life estimation by classification of pre-
dictions based on a neuro-fuzzy system and theory of belief functions. IEEE Trans Reliab 63
(2):555–566
61. Zhang Q, Tse PW-T, Wan X, Xu G (2015) Remaining useful life estimation for mechanical
systems based on similarity of phase space trajectory. Expert Syst Appl 42(5):2353–2360
62. Mosallam A, Medjaher K, Zerhouni N (2014) Data-driven prognostic method based on
Bayesian approaches for direct remaining useful life prediction. J Intell Manuf 27
(5):1037–1048
63. Ramasso E, Rombaut M, Zerhouni N (2013) Joint prediction of continuous and discrete states
in time-series based on belief functions. IEEE Trans Cybern 43(1):37–50
64. Wang P, Youn BD, Hu C (2012b) A generic probabilistic framework for structural health
prognostics and uncertainty management. Mech Syst Signal Process 28:622–637
65. Hu C, Youn BD, Wang P, Taek Yoon J (2012) Ensemble of data-driven prognostic algorithms
for robust prediction of remaining useful life. Reliab Eng Syst Saf 103:120–135
66. Ahmad R, Kamaruddin S, Azid IA, Almanar IP (2012) Failure analysis of machinery compo-
nent by considering external factors and multiple failure modes–a case study in the processing
industry. Eng Fail Anal 25:182–192
67. Huang W, Askin RG (2003) Reliability analysis of electronic devices with multiple competing
failure modes involving performance aging degradation. Qual Reliab Eng Int 19(3):241–254
68. Bichon BJ, McFarland JM, Mahadevan S (2011) Efficient surrogate models for reliability
analysis of systems with multiple failure modes. Reliab Eng Syst Saf 96(10):1386–1395
Study on the Poisson’s Ratio of Solid Rocket
Motor by the Visual Non-Contact
Measurement Teleoperation
1 Introduction
As one of the basic parameters of materials, the measurement of Poisson’s ratio has
a very important significance for the performance analysis of materials. Poisson’s
ratio is the ratio of the material to the axial strain and the transverse strain of the
material under uniaxial tension or compression. In the case of the linear elastic
material, the ratio of the transverse strain and the longitudinal strain is constant, so
the Poisson’s ratio is constant. But in the viscoelastic material, Poisson’s ratio is a
parameter related to the time, frequency and temperature [1, 2]. The measurement
of Poisson’s ratio of viscoelastic material cannot use simple mechanics of elasticity
to characterize the relationship between modulus and strain [3]. The results
obtained by this method are complex and require a large amount of computation.
The influence of contact measurement on viscoelastic material is very large, for
example, the data obtained by the contact method such as strain gauge are not
consistent with the actual data [4]. Compared with the contact measurement,
non-contact measurement has the advantages of high precision and little influence
on the performance of the material [5–7].
In this paper, the method of non-contact measurement is used to analyse the
displacement of the specimen after speckle processing [8–10]. The calculated
Poisson’s ratio of the transverse strain and relaxation modulus is obtained from
the experimental measurements [11, 12].
Under the action of constant load (or stress), the process of strain increasing with
time is called creep [14]. And the ratio of strain to constant stress is called creep
compliance, which is denoted as D(t) and given as follow.
εð t Þ
DðtÞ ¼ ð1Þ
σ0
The process of stress relaxation under constant strain is called stress relaxation
[14]. And the ratio of stress and constant strain is called relaxation modulus, which
is denoted as E(t) and given as follow.
σ ðt Þ
EðtÞ ¼ ð2Þ
ε0
The above Eqs. (1) and (2) can be expressed by the following step function H(t).
1, ðt 0Þ
H ðt Þ ¼ ð3Þ
0, ðt < 0Þ
Zt
∂σ ðtÞ
εðtÞ ¼ σ 0 DðtÞ þ Dðt τÞ dτ ð4Þ
∂τ
0
Zt
∂εðtÞ
σ ð t Þ ¼ ε 0 Eð t Þ þ Eð t τ Þ dτ ð5Þ
∂τ
0
εx ðtÞ
μðtÞ ¼ ð6Þ
εy ðtÞ
where, εx(t) is the transverse strain, and εx(t) is the longitudinal strain.
In the case of the viscoelastic material in the longitudinal direction, the response
lags behind the longitudinal deformation history. Therefore, the ratio of transverse
longitudinal strain cannot be directly used to calculate the Poisson’s ratio of
viscoelastic material, and the corresponding calculation formula is obtained in the
Laplace domain [15]. Poisson’s ratio expression in Laplace domain can be
converted to the formula as follow.
εx ð s Þ
μ ðsÞ ¼ ð7Þ
εy ð s Þ
The Eqs. (8) and (9) are substituted into (7), and we can obtain
146 Y.-B. Li et al.
s
μ ðsÞ ¼ εx ðsÞE ðsÞ ð10Þ
σ0
By using the convolution theorem, the formula (10) is used to obtain the
Poisson’s ratio in time domain as follow [15],
2 t 3
Z
14 ∂εx ðtÞ
μðtÞ ¼ Eð t τ Þ dτ þ EðtÞεx ð0Þ5 ð11Þ
σ0 ∂τ
0
where, E(t) is the relaxation modulus, and εx(t) is the creep strain.
Using the formula (11) as the theoretical basis, the relevant experimental test is
carried out in this paper.
3 Test
Samples are made from a certain model of solid rocket motor grain. According to
the special requirements of the test bench, the tensile sample is processed with
height 50.00 mm, width 11.00 mm, and thickness 3.60 mm. In order to ensure the
universality of the experiment, a total of 5 samples were prepared, and the specific
size of the sample is shown in Table 1.
In order to meet the accuracy requirements, sample processing is completed in a
high temperature environment (50 C) for 24 h, eliminating the residual stress
generated during processing.
The EPLEXOR dynamic mechanical thermal analyser and the VIC-3D visual
test bench produced by the American Solutions Correlated company are used in the
experiment. The dynamic mechanical thermal analysis instrument is used to mea-
sure the tensile relaxation modulus and shear creep compliance. VIC-3D is used to
collect the transverse strain of the sample in the test, as shown in Fig. 1.
For the measurement error, measuring accuracy of dynamic mechanical thermal
analyser GOBA is +0.1%; the accuracy of VIC-3D is consistent with the 0.5 level
standard. Both kinds of test data obtained in the accuracy of more than six after the
decimal point, fully meet the accuracy required by the test.
When using the VIC-3D, visual bench displacement of the sample is measured,
and diagram speckle is sampled and shown in Fig. 2.
In order to obtain the corresponding parameters in the formula (11), a total of two
sets of tests have been done. Each group has been repeated five times under the
same conditions using different samples. The results obtained from the test with
abnormal data removed are as the final test result, and then follow-up calculation.
The two sets of tests are tensile relaxation and tensile creep, as shown in Fig. 3.
During the test, the test process is kept at a constant temperature of 27 C. In the
process of the test, the temperature change is controlled in the range of 0.5 C to
meet the accuracy requirements.
In the relaxation test, each test time is set to 104 s, and the relaxation modulus is
obtained. The test results are shown in Fig. 4.
148 Y.-B. Li et al.
The creep test of each test time is 104 s. The VIC-3D measure displacements of
Point A and Point B under two coordinates with time. The positions of Point A and
Point B are shown in Fig. 5.
The Point A in the test process in the horizontal coordinates of the record is
denoted as Ia(t). The Point B of the horizontal axis is denoted as Ib(t). The initial
time point displacement are recorded as Ia(0) and Ib(0). The Ia(t) and Ib(t) get
displacement data as shown in Fig. 6a, b.
In the experiment, the visual test bench is only on the displacement of two
points. Therefore, the strain of the test process can be obtained by a simple
calculation, and the formula is as follows,
The transverse strain εx(t) is calculated by the formula (13), as shown in Fig. 7,
which is further in the Eq. (11) for Poisson’s ratio calculation.
Through the above two experiments, the relaxation modulus and the transverse
strain in creep are obtained respectively. In order to improve the accuracy of the
test, the test abnormal data need to be excluded. The application of Grubbs criterion
in the choice (set reliability α ¼ 95%) can delete abnormal data [18].
150 Y.-B. Li et al.
In order to obtain the stable relaxation modulus and lateral strain, the five sets of
data are calculated using MATLAB, and the corresponding average data are
obtained.
The relaxation modulus and transverse strain are measured by Prony series
fitting using MATLAB platform. In order to ensure that there is sufficient fitting
precision, the Prony series are set to the nine class number.
In the fitting of the data in the MATLAB, in order to ensure the parameters have
good accuracy in time domain on the whole, the nine coefficients are in different
number of basis coefficients, such as τi in Table 2. This coefficient setting can not
only ensure the accuracy of the initial segment, but also can guarantee the accuracy
of small amplitude change in the long time course.
It is found that the instability of the initial transient load has great influence on
the acquisition of the experimental data and the result of the initial time. Therefore,
there is an abnormal in the data measured at the initial time of the test. The [0,1]s
interval segment approximation curve of the data is fitted by the least square fitting
before the test time in 1s.
Study on the Poisson’s Ratio of Solid Rocket Motor by the Visual Non-Contact. . . 151
Fig. 8 (a) E(t) in natural coordinate system; (b) E(t) in logarithmic coordinate system
After the process of experimental data obtained, Prony series fitting. Prony series
fitting formula for relaxation modulus is given as follow,
X
9
t
EðtÞ ¼ a0 þ ai exp ð15Þ
i¼1
τi
The coefficients in the formula are shown in Table 2. The fitting results are
shown in Fig. 8a, b. Due to the selection of coordinate system, the date changes
significantly in the front part in natural coordinates, and the date changes signifi-
cantly in the rear part in logarithmic coordinates. Therefore, the obtained data are
used in two coordinates expression, which are separately presented in figures of the
natural coordinates system and after logarithmic coordinate system.
In the same way, the coefficients of the Prony series of εx(t) are shown in
Table 3. The fitting results are shown in Fig. 9a, b. The Prony series fitting formula
for εx(t) is given as,
X
9
t
ε x ð t Þ ¼ b0 þ bi exp ð16Þ
i¼1
φi
152 Y.-B. Li et al.
Fig. 9 (a) εx(t) in natural coordinate system; (b) εx(t) in logarithmic coordinate system
Fig. 10 (a) μ(t) in natural coordinate system; (b) μ(t) in logarithmic coordinate system
The above obtained two fitting results by MATLAB are substituted into the
Eq. (11) with the Poisson like Fig. 10a, b. The calculated Poisson’s ratio function is
fitted to Prony series, and the results are shown in Table 4.
Study on the Poisson’s Ratio of Solid Rocket Motor by the Visual Non-Contact. . . 153
X
9
t
μðtÞ ¼ c0 þ ci exp ð17Þ
i¼1
ωi
4 Conclusion
Acknowledgements This work was supported by the National Natural Science Foundation of
China under the contract number 11262014.
References
1. Hilton HH, Yi S (1998) Significance of (an)isotropic viscoelastic Poisson ratio stress and time
dependencies. Int J Solids Struct 35(23):3081–3095
2. Lee HS, Kim JJ (2009) Determination of viscoelastic poisson’s ratio and creep compliance
from the indirect tension test. J Mater Civil Eng 21(8):416–425
3. Lakes RS, Wineman A (2006) On Poisson’s ratio in linearly viscoelastic solids. J Elast 85
(1):45–63
4. Zhang JB, Ju YT, Meng HL et al (2012) Research on measurement method of surface of
double base propellant grain. J Ball 24(2):62–65
5. Aushev AA, Barinov SP, Vasin MG et al (2015) Alpha-spectrometry and fractal analysis of
surface micro-images for characterisation of porous materials used in manufacture of targets
for laser plasma experiments. Quantum Electron 45(6):533–539
6. Kabeer S, Attenburrow G, Picton P et al (2013) Development of an image analysis technique
for measurement of Poisson’s ratio for viscoelastic materials: application to leather. J Mater
Sci 48(2):744–749
7. Zhou GD, Li HN, Ren L et al (2007) Study on influencing parameters of strain transfer of optic
fiber Bragg grating sensors. Eng Mech 24(6):169–173
8. Edmundson K, FraserC S (1998) A practical evaluation of sequential estimation for vision
metrology. ISPRS J Photogramm 53(5):272–285
9. Sun W, He XY, Xu M et al (2007) Study on the tension test of membrane materials using
digital image correlation method. Eng Mech 24(2):34–38
10. Sun W, He XY, Huang YP et al (2008) Experimental study on identification of modal
parameters of cable. Eng Mech 25(6):88–93
11. Kassem E, Grasley Z, Masad E (2013) Viscoelastic Poisson’s ratio of asphalt mixtures. Int J
Geomech 13(2):162–169
12. Keramat A, Kolahi AG, Ahmadi A (2013) Waterhammer modelling of viscoelastic pipes with
a time-dependent Poisson’s ratio. J Fluid Struct 43:164–178
13. Klompen ETJ, Govaert LE (1999) Nonlinear viscoelastic behaviour of thermorheologically
complex materials. Mech Time Dep Mater 3(1):49–69
14. Yang TQ, Luo WB, Xu P et al (2004) Viscoelastic theory and application. Science Press,
Hubei
15. Zhao BH (1995) An investigation on viscoelastic Poisson’s ratio and dynamic complex
Possion’s ratio. J Prop Tech 3:1–7
16. Fu YM, Li PE, Zheng YF (2005) Nonlinear dynamic responses of viscoelastic symmetrically
laminated plates. Eng Mech 22(4):24–30
17. Zheng JL, Lu ST, Tian XG (2008a) Viscoelastic damage characteristics of asphalt based on
creep test. Eng Mech 25(2):193–196
18. Shalom BY, Li XR, Kirubarajan T (2001) Estimation with application to tracking and
navigation. Wiley, New York
19. Tian SP, Lei YJ, Li DK et al (2007) Incremental viscoelastic three-dimensional fem based on
herrmann variational principle and its application. Eng Mech 24(7):28–32
20. Zhang HL, Zhou JP (2001) Viscoelastic stochastic finite element simulation of solid propellant
grain with random Poisson’s ratio. J Prop Tech 22(3):245–249
21. Zheng JL, Guoping Q, Ronghua Y (2008b) Testing thermal viscoelastic constitutive relation of
asphalt mixtures and its mechanical applications. Eng Mech 25(1):34–41
Reliability Allocation of Multi-Function
Integrated Transmission System Based
on the Improved AGREE Method
Qi-hai Liang, Hai-ping Dong, Xiao-jian Yi, Bin Qin, Xiao-yu Yang,
and Peng Hou
1 Introduction
An integrated transmission system is a key device that decides the mobility of tanks
and armored vehicles, and undertakes the task of transmitting mechanical energy to
crawler traveling device, and makes tanks to generate movements [1]. Reliability is
an important characteristic, which directly affects the design quality and operation
Suppose that R∗ s ðtÞ is the reliability index of a system at operating time t. For a
series system composed of the subsystems with exponentially distributed lifetimes,
it is known that the greater number of basic component in a subsystem, the more
complex the subsystem is. Suppose that each basic component in the series system
Reliability Allocation of Multi-Function Integrated Transmission System. . . 157
has the same impact on the reliability of the series system, so based on the AGREE
method [6], the allocated reliability Ri(t) of subsystem iat time t can be expressed as
ni
1=N ni =N
Ri ð t Þ ¼ R∗
s ðt Þ ¼ R∗
s ðtÞ ð1Þ
P
n
Where, ni is the number of basic components in subsystem i; N ¼ ni is the
i¼1
total number of basic components in the system.
Moreover, the impact of different subsystem on the system reliability is differ-
ent, and the different impact can be described by importance factor. The importance
factor of subsystem i represents the probability that the system will fail if subsystem
i fails, and Eq. (1) can be written as
ni =N
ωi 1 eλi ti ¼ 1 R∗
s ðtÞ ð2Þ
The main words in all headings (even run-in headings) begin with a capital letter.
Articles, conjunctions and prepositions are the only words which should begin with
a lower case letter.
For a simple series and parallel system, it is easy to determine the importance factor
of a subsystem, but for a multi-function complex system, the importance factor of a
subsystem cannot be easily determined, so the traditional AGREE method can only
be applied in the reliability allocation of a simple structure system (e.g., a series
system or a series–parallel system), and cannot be directly applied in reliability
allocation of a multi-function complex system. To solve this problem, a heuristic
algorithm is adopted to calculate the subsystem importance factor in this paper.
158 Q.-h. Liang et al.
ωk, i ¼ 1, i ¼ 1, 2, . . . , mk ð5Þ
nk , i
ak , i ¼ , i ¼ 1, 2, . . . , mk ð6Þ
N k ωk, i tk, i
max β
s:t: R∗
k ð t Þ Rk ð t Þ ð9Þ
β>0
0
λk, i ¼ ak, i β, i ¼ 1, 2, . . . , mk ð10Þ
0
0
Rk, i ðtÞ ¼ exp λk, i tk, i , i ¼ 1, 2, . . . , mk ð11Þ
0
(6) Calculate the new importance factor ωk, i of each subsystem i for function k.
0 ∂Rk ðtÞ
ωk , i ¼ 0 , i ¼ 1, 2, . . . , mk ð12Þ
∂Rk, i ðtÞ
Reliability Allocation of Multi-Function Integrated Transmission System. . . 159
Where, Rk ðtÞ ¼ f ðRk, 1 ðtÞ; Rk, 2 ðtÞ; ; Rk, mk ðtÞÞ is the calculated reliability of
function k.
0
(7) Compare Rk, i ðtÞ with Rk , i(t)
0
If any
Rk, i ðtÞ Rk, i ðtÞ
ε, then let
8 0
>
< ωk , i ¼ ω k , i
0
λk, i ¼ λk, i , i ¼ 1, 2, . . . , mk ð13Þ
>
: R ðtÞ ¼ R0 ðtÞ
k, i k, i
0
And go to step (2); otherwise, stop and Rk, i ðtÞ is the initial allocation result for
function k.
In the above allocation process, the constraint f ðRk, 1 ðtÞ; Rk, 2 ðtÞ; ; Rk, mk ðtÞÞ
R∗k ðtÞ will always be met. Under the constraint of system function reliability
goal, we aim to determine the most reasonable and allowable reliability for sub-
systems through reliability allocation because a higher reliability level generally
requires a more cost and more difficult design. The purpose of reliability optimi-
zation allocation is to find a reasonable allocation result that not only satisfies the
requirement of function reliability goal, but also with the minimum cost and the
simplest design.
Because some subsystems not only exist in a function, but also exist in two or more
functions, then the same subsystem in different functions may be allocated to different
reliability value. So the reliability allocation values of such subsystems need to be
adjusted. In order to meet reliability requirements
of all functions, we select the strictest
reliability allocation value, i.e., max Rk1 , ix ðtÞ; Rk2 , ix ðtÞ; . . . ; Rkj , ix ðtÞ , where k1 ,
k2 , . . . , kj are functions 1 to j in which the subsystem ix is involved. For other
subsystems, if we keep the initial allocation values unchanged, the reliability of each
function will be larger than its reliability requirement, and the more design resources
will be needed than what is required. So, after the reliability of subsystems which are
reused in multiple functions is determined, the reliability of other subsystems which are
not reused will be reallocated in the following section.
160 Q.-h. Liang et al.
For function k, suppose that the number of subsystems reused in multiple functions
is pk, and their reliability values can be expressed as R∗ ∗ ∗
k, 1 , Rk, 2 , . . . , Rk, pk , then the
reliability of other subsystems which are not reused can be reallocated as follows:
(1) Suppose the initial reliability importance factor of subsystems which are not
reused is 1
max β
R∗k ðt Þ R k Þ
ð t
s:t: Rk, 1 ðtÞ ¼ R∗ k , 1 ðt Þ
Rk , 2 ð t Þ ¼ R∗ ð15Þ
k , 2 ðt Þ
...
Rk, pk ðtÞ ¼ R∗ k, pk ðtÞ
0
λk, i ¼ ak, i β, i ¼ pkþ1 , pkþ2 , . . . , mk ð16Þ
0
0
Rk, i ðtÞ ¼ exp λk, i tk, i ð17Þ
(4) Calculate the new importance factor of subsystems which are not reused in
function k
0 ∂Rk ðtÞ
ωk , i ¼ 0 , i ¼ pkþ1 , pkþ2 , . . . , mk ð18Þ
∂Rk, i ðtÞ
0
(5) Compare Rk, i ðtÞ with Rk , i(t)
0
If any
Rk, i ðtÞ Rk, i ðtÞ
ε, then let
Reliability Allocation of Multi-Function Integrated Transmission System. . . 161
8 0
>
< ωk , i ¼ ωk , i
0
λk , i ¼ λk , i , i ¼ pkþ1 , pkþ2 , . . . , mk ð19Þ
>
: R ðtÞ ¼ R0 ðtÞ
k, i k, i
0
And go to step (2); otherwise, stop and Rk, i ðtÞ is the final allocation result of each
subsystem.
The reliability indexes which are often used in function reliability allocation are
Mean Time Between Failures, reliability or failure rate. In this paper, reliability is
selected to be allocated for a certain type of integrated transmission device based on
the following assumption [7]:
1. There are only two states as failure and success for the system and each
subsystem;
2. There are no common cause failures, which mean the failures of different
subsystems are independent of each other;
3. Times to failures all follow exponential distributions.
straight
swerve braking
driving
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19
Assume that the function reliability requirements in this system for straight running,
swerve and braking at time 1000 h are 0.9, 0.8, 0.7 respectively, and 0.0001 is the
allocation accuracy requirement. According to the calculation process in Sect. 2.2,
reliability indexes of the three functions can be allocated to each subsystem. The
complexity of each subsystem, i.e. the number of basic components involved in
each subsystem, and reliability values under different function states are shown in
Table 1.
Some subsystems may not only exist in a single function, but in two or more
functions, so the same subsystem involved different functions may be allocated
different reliability values. For example, the allocated reliability of body parts in
straight running function is 0.99623, while its allocated reliability in the steering
task is 0.99167. The body parts are not involved in braking function, so the
reliability of braking function is not allocated to the body parts. In order to meet
the reliability requirements of all functions, we choose the maximum reliability
allocation value for such reused subsystems, i.e., body parts’ reliability value is
Reliability Allocation of Multi-Function Integrated Transmission System. . . 163
4 Result Comparison
In order to verify the rationality and accuracy of the reliability allocation method
proposed in this paper for multi-function systems, take the same integrated trans-
mission device whose reliability index is allocated by the traditional AGREE
method as a comparison. The reliability requirement for the three functions: straight
running, swerve, braking is still assumed as 0.9, 0.8, 0.7 respectively, and the
reliability importance factor of each subsystem is still set to be 1. We choose the
maximum allocation value for those subsystems which are reused in different
functions, and the reliability allocation results are shown in Table 4.
According to the reliability allocation results of all subsystems in Tables 2 and 3,
we can obtain the expected reliability level of three functions: straight running,
swerve and braking, is 0.900035, 0.800009, 0.700032, respectively, a little larger
than the required. However, we also can obtain the reliability of three functions
based on AGREE method from Table 4: straight running, swerve, braking are
0.900091, 0.895405, 0.790525, obviously, larger than the required, especially the
last two results. This result shows that the reliability allocation results based on
Reliability Allocation of Multi-Function Integrated Transmission System. . . 165
Table 4 Reliability allocation results of the integrated transmission device with traditional
AGREE method
R/Straight
No. Name running R/Swerve R/Braking Rmax
1 Body parts 0.99624 0.99177 NA 0.99624
2 Middle bracket 0.99624 0.99177 NA 0.99624
3 Torque converter assembly 0.99437 0.98768 0.98203 0.99437
4 Transmission assembly 0.99064 0.97955 NA 0.99064
5 Hydraulic Torque converter 0.98881 0.97551 NA 0.98881
6 Planet before shift gear 0.98881 0.97551 NA 0.98881
7 Auxiliary drive NA 0.98361 NA 0.98361
8 Hydraulic gear Retarder NA NA 0.95279 0.95279
9 Hydraulic retarder control valve NA NA 0.98203 0.98203
10 Left side cover 0.99437 NA 0.98203 0.99437
11 Right side cover 0.99437 NA 0.98203 0.99437
12 Fan drive assembly NA NA 0.97023 0.97023
13 Liquid viscous clutch assembly NA NA 0.95279 0.95279
14 Oil pump Group 0.98881 0.97551 0.96438 0.98881
15 Couplet of pump motor NA NA 0.97611 0.97611
16 Oil Supply System 0.98881 0.97551 0.96438 0.98881
17 Hydraulic control system 0.99063 0.97955 0.97023 0.99063
18 Manipulation of the electronic con- 0.99063 0.97955 0.97023 0.99063
trol system
19 Overall gearing 0.99250 0.98361 NA 0.99250
Notes: NA means that the subsystem is not used in the corresponding function
improved AGREE method is more accurate and can save more resource and cost
under the condition of required reliability.
5 Conclusions
References
1. Zheng MQ, Feng CZ, Lan ZY (2003) Tank and armored vehicle. Beijing Institute of Technol-
ogy Press, Beijing
2. Guo SW (2016) Study on the allocation method for the power-shift steering transmission on
balance of reliability and maintainability index with the goal of availability. Beijing Institute of
Technology, Beijing
3. Yi XJ, Lai YH, Dong HP et al (2015b) A reliability optimization allocation method considering
differentiation of functions. Int J Comput Methods 13(4). doi:10.1142/S0219876216410206
4. Nabil N, Mustapha N (2005) Ant system for reliability optimization of a series system with
multiple-choice and budget constraints. Reliab Eng Syst Saf 72:1–12
5. Yi XJ, Hou P, Dong HP et al (2015a) A reliability optimization allocation method for systems
with differentiation of functions. In: Proceedings of ASME 2015 international mechanical
engineering congress and exposition, IMECE 2015-52928
6. Ebeling CE (2009) An introduction to reliability and maintainability engineering. Waveland,
Long Grove, IL
7. Li RY, Wang JF, Liao HT et al (2015) A new method for reliability allocation of avionics
connected via an airborne network. J Netw Comput Appl 48:14–21
A Study of Sustained Attention Improving by
Fuzzy Sets in Supervisory Tasks
Abstract In supervisory task, the operator needs not do any thing over prolonged
periods of time, but must keep a certain level of attention to detect more serious
signals for confirming the problems of the system. When the activity is to be
performed for a continuous period of time, the ability to maintain attention may
be severely impaired. The purpose of this study is to develop a quantitative
sustained attention measuring model by fuzzy sets, which integrates the concepts
of hit, false alarm and response time. Then, an alarm system is developed for
improving performance of the sustained attention in supervisory task. The exper-
imental results showed that the effect of the system is good for improving perfor-
mance of sustained attention.
1 Introduction
As systems are integrated and automated, the role of the human operator has
changed from that of an active controller to a decision maker and manager, a
shift from active to supervisory control. The supervisory task can be automatically
started when the automation is turned on or the operation mode is set to auto mode.
The supervisory responsibility includes monitoring the automated operating sys-
tem, the start or shut off of the system, and restoring the order. This duty empha-
sizes judgment and solution [1]. Therefore, the efficiency carrying out and the back
coupling do not mostly come directly from automation but is by the panel. As a
result, automation monitoring work is mostly boring work, such as: the system-
monitoring supervisor in a nuclear power plant. The task is much more difficult if
In sustained tasks, there will be more sensory or neural activity in the brain when a
signal is present than when it is absent [14]. We refer to the quantity X as the
evidence variable when the external stimulus generate human neural activity. If
there is enough neural activity, X exceeds the critical threshold X1 of preliminary
attention, and the operator will decide “yes.” If there is too little, the operator will
decide “no,” as shown in Fig. 1. After a period of time, the signal gradually
strengthens until X exceeds criterion X1, and the operator will say “yes.” Some-
times, even when no signal is present, X will exceed the criterion X1 because of
random variations in the environment and the operator’s own “baseline” level of
neural firing, and the operator will say “yes” (generating a false alarm). When the
operator detects an unusual or infrequent condition in the preliminary vigilance,
then he/she must keep sustained attention to detect any resulting emergency signals.
Evidence
variable X Response Threshold X2
Response Threshold X1
Time
Emergency
response point
Signal
Amount occurring
of energy
from a
signal Dangerous
off Warning
Fig. 1 The change in the evidence variable x caused by signal in supervisory task
170 C.-L. Liu et al.
If the system approaches a dangerous situation and there is enough neural activity,
then X exceeds the critical threshold X2 of the sustained attention, and the operator
will decide “Yes,” otherwise, decide “No.”
When the human neural activity in preliminary and sustained attention is
understood, a fuzzy logic inference can then be used to develop a fuzzy sustained
attention measuring model (FSAMM). The fuzzy logic inference can be seen as a
heuristic and modular way to define nonlinear, table-based measurement and
control [15]. A fuzzy logic is a set of IF-THEN rules. The linguistic statements of
the IF-part are obtained by fuzzyfication of numerical input values, the statements
of the THEN-part are defuzzyficated to numerical output values [15, 16]. Assume
the fuzzy rule consists of N rules as follows:
3 Experiment
subjects. When any of these signals flashed, the subject should press the switch to
stop flashing.
3.2 Subjects
There are 16 Subjects requested to monitor and feed cool water satisfying the SG
situations, and are requested to maintain safety: If he/she detects emergency status,
it should be controlled immediately. Each subject must accept the training (at least
three times) to learn how to control the simulated system.
There are two independent variables “Event interval” and “Event rate” tested in a
two-factor replicate experiment.
1. Event interval (factor A): refers to the total time of the emergence situation
(i.e. event). There are three levels to be considered: 5 min. (A1), 10 min. (A2),
and 20 min. (A3).
2. Event rate (factor B): refers to the frequency of the unusual situation (i.e. event).
There are two levels: 1/per half hour (B1) and 5/per hour (B2).
There are four dependent variables (performance variables) used to record the
performance of the participants, including:
1. The water level at the time of the participant’s response.
2. “Hit” rate. The “Hit” condition is defined as follows. If the participant detects the
water level of any SG to be below the normal water level and presses the MFS
valve button between 5.9 and 6 m, the response is defined as a “Hit” in the
preliminary vigilance. If not, it is defined as a “Miss.” In the extended vigilance,
the “Hit” is defined as the interval between 4.9 and 5 m.
3. “False Alarm” rate. If the alarm signal appears when the water level is at 6.3 m
(a noise in the preliminary vigilance), the participants must press the AFS
(Automatic feed-water system) valve button to cancel the alarm before it reaches
6.1 m, otherwise it will be a “False Alarm”. In the extended vigilance, the alarm
signal appears at 5.3 m and must be cancelled before it reaches 5.1 m.
4. Fuzzy value y* of sustained attention performance.
174 C.-L. Liu et al.
In the FSAMM, we defined y ¼ 50 (%) as good status, and in the first experiment
the average performance of preliminary vigilance is 47.31 (%). Therefore, the
performance value 50 (%) of y* was defined as the threshold for producing an
alarm signal. If the value of y* was smaller than 50 (%), the system would produce
an alarm signal to warn the participant to maintain his/her sustained attention.
In the sustained attention experiment, we found that the preliminary attention
performances of participants #4 and #11 to #18 were poor, so the fuzzy alarm
activated. A differences of means test showed that these three participants’
sustained attention performances were significantly better than in the preliminary
A Study of Sustained Attention Improving by Fuzzy Sets in Supervisory Tasks 175
attention. However, we also found that the means of the sustained attention
performance of some participants was poor than in the preliminary attention
when the FSAMM not working. These findings indicated that the effect of an
alarm for improving the sustained attention performance is significant when the
performance of preliminary attention of participants is poor.
5 Conclusions
In this study, a fuzzy sustained attention measuring model (FSAMM) with relation
to human decision-making in supervisory task was proposed and verified by
conducting experiments. The main findings of this study are as follows:
1. In general, sensitivity parametric index d’ and bias parametric index β are used
to evaluated the performance of sustained attention. These indices only used
observer’s response “yes” or “no” to explain and evaluate sustained attention.
However, for some supervisory tasks such as unlimited hold tasks (e.g., super-
visors in nuclear plant), the reaction time is longer. These indices are unable to
respond to this characteristic, but the FSAMM could.
2. The FSAMM could consider the characteristics of the difference between Type I
error (miss), Type II error (false alarm) and reaction time in unlimited supervi-
sory control task. Using fuzzy sets to evaluate the domain of decision response,
the differences of performance could be evaluated. It could expand the tradi-
tional two values logic to continuous multiple-valued performance evaluation,
then, clearly describe difference of the judgment in the monitoring “also this also
other” work.
3. The fuzzy logic model and d’ are more efficient in response sensitivity than β in
this study. The correlation coefficient between d0 value and the fuzzy model is
0.569. However, if the automation is in high reliability, the effect of d0 value on
zero miss will be over estimated. The fuzzy model will not.
By using fuzzy variables and fuzzy rules, the fuzzy logic measuring model
combined the “signal detection ability” and “false alarm status” to evaluate
sustained attention. According to the results, then, we can take one step ahead to
analyze and design when and how to call operator’s sustained attention at the right
time to set an adapted attention performance alarm to reduce the probability of
human decision-making error.
References
1. Szalma JL, Taylor GS (2011) Individual differences in response to automation: the big five
factors of personality. J Exp Psychol Appl 17:71–96
176 C.-L. Liu et al.
2. Warm JS, Jerison H (1984) The psychophysics of vigilance. In: Warm JS (ed) Sustained
attention in human performance. Wiley, England
3. Smit AS, Eling TM, Coenen ML (2004) Mental effort causes vigilance decrease due to
resource depletion. Acta Psychol 115:35–42
4. Hancock PA (2013) In search of vigilance: the problem of iatrogenically created psychological
phenomenon. Am Psychol 68:97–109
5. Kenneth RB, Lloyd K, James PT (1986) Handbook of perception and human performance, vol
2. Wiley, New York
6. Wickens CD (1992) Engineering psychology and human performance. HarperCollins,
New York
7. Grier JB (1971) Nonparametric indexes for sensitivity and bias: computing formulas. Psychol
Bull 75:424–429
8. Hodos W (1970) A nonparametric index of response bias for use in detection and recognition
experiments. Psychol Bull 74:351–354
9. Parasuraman R (1979) Memory load and event rate control sensitivity decrements in sustained
attention. Science 205:924–927
10. Parasuraman R, Davies DR (1977) A taxonomic analysis of vigilance performance. In: Mackie
RR (ed) Vigilance: theory, operational performance, and physiological correlates. Plenum,
New York
11. Egan JP, Greengerg GZ, Schulman AI (1961) Operating characteristics, signal detectability
and the method of free response. J Acoust Soc Am 33:993–1007
12. Silverstein SM, Light G, Palumbo DR (1998) The sustained attention test: a measure of
attentional disturbance. Comput Hum Behav 14:463–475
13. Szalma JL, Warm JS, Matthews G, Dember WN, Weiler EM, Meier A (2004) Effects of
sensory modality and task duration on performance, workload, and stress in sustained atten-
tion. Hum Factors 46:219–233
14. Wickens CD, Lee JD, Liu Y, Becker SG (2004) An introduction to human factors engineering,
2nd edn. Person Education, New Jersey
15. Driankov D, Hellendoorn H, Reinfrank M (1996) An introduction to fuzzy control. Spring,
New York
16. Zadeh LA (1973) Outline of a new approach to the analysis of complex systems and decision
processes. IEEE Trans Syst Man Cybernet 3:28–44
17. Pattyn N, NeytX HD, Soeten E (2008) Psychophysiological investigation of vigilance decre-
ment: boredom or cognitive fatigue? Physiol Behav 93:369–378
Waukesha 7044 Gas Engine Failure
Investigation
Abstract Three engine failures occurred from March to July in 2015. The failures
were due to engine overloading and in each of the three instances caused severe
scuffs and cracking on pistons and liners, particularly on #4 cylinders. Operating
condition changes were identified as the root cause for the recent engine overload
failures. Several maintenance issues, such as incorrect stepper motor settings,
lubrication oil over filling and failed spark, also contributed to the failures. These
factors combined and caused the engines to be overloaded and triggered the failure
mechanisms which lead to the engine failures. To avoid failures in the future, the
recommendations are provided, such as reduce the engine load by increasing the
compressor clearance; monitor the engine load when operating condition changes
and provide appropriate engine load awareness training to engineering and field
personnel.
1 Introduction
Since March 2015, there have been a number of engine cylinder failures on one
APA compressor station. The compressor station consists of three Enerflex pack-
ages which comprise of a Waukesha 7044 GSI gas engine and Ariel JGK/4 single
stage gas compressor [3]. The failure history is as follows:
• Severe scuffing was found on Unit 2 Right Bank (RB) cylinder 4 piston and liner
on 10/3/15 at 636 h operating hours.
• Severe scuffing was found on Unit 2 Left bank (LB) cylinder 4 piston and liner
on 14/04/2015.
X. Liu (*)
APA Group, Level 3, 121 Wharf Street, Spring Hill, QLD 4000, Australia
e-mail: [email protected]
A. Newbury
APA Group, Suite 1, 60-62 Gladstone Street, Fyshwick, ACT 2620, Australia
• Split liner and piston damage were found on Unit 1 RB cylinder 4 and in addition
severe scuffing on LB cylinder 4 piston and liner on 05/07/2015.
All failures occurred on left and right bank #4 cylinders. An initial investigation
found that several factors might have contributed to the failures:
• According to the ESM (OEM engine system management computer)/SCADA
data:
– Engines were found to be overloaded prior to failures.
– Cylinder temperature deviations were observed, these were higher on the
bank that had the failure.
– Manifold banks appear to have been imbalanced, the manifold air pressure
was higher on the bank that failed.
– The RB has far more alarms and Emergency Shut-downs (ESDs than the LB,
this varies from a factor of three times up to 127 times that of the LB, refer to
Table 2 for more details.
– LB and RB gas/air ratios and (Inlet Manifold Absolute Pressure) IMAPs were
found to be too high on unit 3.
• Oil level controller issues caused oil overfilling the frame and sump on unit 2 for
the second failure which can increase engine load substantially.
• A couple of cylinder spark failures were also spotted which may have
overloaded other cylinders.
• It appears that #4 cylinders, either left bank or right bank, are the most vulner-
able cylinders.
Figure 1 shows a failed piston from Unit 2, on which scuffs and cracks are evident.
This is a typical engine failure at the compressor station.
Research by Waukesha has found that the scuffs and cracks are caused by the so
called wear load. Wear load is defined as a measure of frictional power due to
asperity contact per unit area with units of MW/m2. This frictional energy flux will
dissipate heat into the piston and liner. Apparently, this frictional power is related to
pressure of metal to metal contact and sliding velocity. The heat generated from this
process can cause the oil viscosity to drop leading to further asperity contact and
more frictional heating. This thermal instability will eventually lead to scuffing and
even seizure [1].
Waukesha 7044 Gas Engine Failure Investigation 179
decrease. It reaches maximum load at 8500 kPa suction pressure when discharge
pressure is 13,000 kPa. As the discharge pressure increases, the engine load always
rises, see Fig. 2. Figure 2 is acquired using the Ariel performance software.
From January 2015, the suction pressure at the compressor station has been
operating in the range from around 7800 to 10,000 kPa. The discharge pressure has
been fluctuating mostly within the range from 10,800 up to 14,800 kPa.
Under certain operating conditions, especially at the high discharge pressures
(up to 14,800 kPa); it is likely that the compressor demands more power than the
engine can generate. The engine will be overloaded when this occurs. Please refer
to the dotted section of the engine power curves. At 1200 rpm the power required is
1680 hp, while the driver can only provide a maximum of 1575 hp which equates to
106.7% of sustainable available power before considering power deterioration due
to tuning issues or engine condition. Figure 3 shows an engine overloading situation
possibly caused by decreased suction pressure and elevated discharge pressure.
Engine tune imbalance issues were observed in the historical data at compressor
station and contributed to engine failures. The identified issues included cylinder
and bank imbalance issues, typically cylinder temperature deviation and uneven LB
and RB gas/air ratio.
Waukesha 7044 Gas Engine Failure Investigation 181
Stepper motor malfunctioning can cause serious engine imbalance issues. Cylinder
temperature deviations were observed and manifold banks Gas/Air ratios were also
found to be out of range. The cylinder temperatures were also found uneven when
Unit 1 engine failed [2]. In this instance high cylinder temperature was found on the
failed LB #4 cylinder.
Gas/Air Ratio and Engine Intake Manifold Absolute Pressure (IMAP) Issues
On Unit 3, the LB and RB gas/air ratios and engine intake manifold absolute
pressures (IMAP) were found to be too high, see Table 1. It can be seen from the
table that Unit 3 was burning rich gas fuel which might have contributed to the
temperature abnormality, while the IMAP is a direct reading of engine load. The
IMAP is above the specification which shows the engine is in overloading
condition.
182 X. Liu and A. Newbury
Table 2 shows statistics of alarms and shutdowns on Unit 1 since reset, from this
data it is apparent that the engine is in a severe bank imbalance situation. The alarm
and shutdown count is substantially higher on the right bank than the left bank, the
exhaust temperature high alarm shows the greatest ratio of 1:127.5.
There are several maintenance issues that also contributed to engine overloading
failures, these are described below.
Oil Overfilling
When Unit 2 failed on 14th of April, both the compressor frame and engine sump
were found to be overfull with lubricating oil. Upon further investigation it was
learned that the oil level controllers were malfunctioning which caused the issue.
The high oil level creates parasitic loads on the engine and compressor crankshafts
as they work to displace the oil during each rotation. This causes extra friction
within the engine and compressor thus increasing engine load. A field test was
conducted and proved the assumption, see Table 3.
It is recommended that the Kenco engine lube oil level controllers be replaced.
The vent tubing from Kenco to engine crankcase should be minimum ½” tube,
continuously sloping towards Kenco and should be as short as possible.
Waukesha 7044 Gas Engine Failure Investigation 183
Ignition/Spark Failure
There were also several cases when failed ignition was found. If one cylinder has a
failed spark, the remaining cylinders will need to work harder to compensate for the
power loss, which causes overloading. For a 12 cylinder engine, each cylinder
contributes 8.33% power. Therefore, 131.25 hp was lost for each spark failure. One
known issues is that the engine wiring harness was damaged by rodents shortly after
commissioning and repaired locally; the harness integrity has been compromised.
Therefore upgrading of the engine wiring harness is recommended to reduce
ignition failures and other reliability issues.
The fact that all the failures are on #4 cylinder, either left bank or right bank,
indicates #4 cylinder is likely to be in the most vulnerable location. Although no
evidence was found to support this hypothesis, the vulnerability may be caused by
insufficient lubrication and/or heat transfer. However, historic operation of these
engines indicates that the #4 cylinders would be less vulnerable if the engine is
tuned and loaded properly.
2.2.5 Vibration
Concerns were raised by field personnel that the failures might be triggered by
vibration issues as high vibration was reported during commissioning. Enerflex was
contacted and the original commissioning document was reviewed. The vibration
was found not to be an issue.
It is understood that over time vibration levels may change, there still lack of
evidence that there is a correlation between engine failure and vibration in these
failures. Further investigation may be required if more vibration failure evidence
becomes available.
184 X. Liu and A. Newbury
2.2.6 Lubrication
The engine lubrication oil was changed from Castrol Duratec L-MG to Castrol
Duratec L. Although the Kinematic Viscosity of Duratec L is slightly lower than
that of Duratec L-MG, both of them meet Waukesha engine lubrication require-
ments. The reason for the change was that Duratec L-MG40 became obsolete and
no longer available from the vendor. Duratec L was the vendor replacement
product. The change was reviewed and approved under management of change
procedures. It is therefore unlikely that the oil change contributed to the engine
failures.
However, upon review of the oil analysis reports, it appears that the oil sampling
is conducted on a run hour basis. Considering the actual frequent starts/stops which
could contribute to accelerated wear, It is recommended that fluid sampling be
conducted at a minimum monthly interval. The oil sampling and analysis program
should be reviewed for alignment to OEM requirements for each piece of
equipment.
Valve clearance spacers can be applied to reduce the compressor power require-
ment and therefore, the engine load. However as a result, the gas flow rate will be
reduced. To minimize flow rate losses, the clearance used need to be optimized.
Based on Ariel performance modeling, two valve spacers at suction side are
recommended for 14,500 kPa discharge pressure, and four spacers are
recommended if discharge pressure of 15,000 kPa is anticipated. The Variable
Volume Clearance Pocket (VVCP) also needs adjustment to maximize gas flow
rate, depending on the suction pressure. Figure 4 shows the VVCP settings under
14,500 kPa discharge pressure.
0%VVCP
25%VVCP
50%VVCP
75%VVCP
100%VVCP
0%VVCP
50%VVCP
100%VVCP
Fig. 4 Recommended VVCP operating envelope (green line) at 1200 rpm, under 14, 500 kPa
• Introduce weekly ESM fault logging data capture, reporting to Engineering and
clearing.
• Revise maintenance documentation to record essential tuning parameters every
2000 h (Gas/air pressure left and right bank and pressure difference, stepper
position left and right and difference, IMAP left and right and difference, throttle
position) as well as log basic operating parameters for a period of no less than 1 h
following every service.
• Recommend replacing Kenco engine lube oil level controllers, confirm that vent
tubing from Kenco to engine crankcase is minimum ½” tube, continuously
sloping towards Kenco and as short as possible.
4 Conclusion
Operating condition changes were found to be the root cause for the failures, in
particular the elevated suction and discharge pressures created exceeded available
engine power. In addition, several maintenance issues such as incorrect stepper
motor settings, lubrication oil compartment over filling and failed spark contributed
to the failures. These factors combined and caused the engines to be overloaded and
triggered the failure mechanisms which led to the engine failures.
186 X. Liu and A. Newbury
5 Recommendations
• Reduce the engine load by increasing the compressor clearance (install valve
spacers). Increasing the compressor clearance will reduce the power require-
ment. As a result the capacity might be affected. VVCP can be adjusted to
optimize compressor performance under changing operating conditions.
• Monitor the engine load when operating condition (especially the suction and
discharge pressure) changes; implement engine load alarms and trips to protect
the engine from being overloaded for prolonged period of time.
• Review the operating and maintenance practices.
• Provide appropriate engine load awareness training to engineering, field person-
nel, engineers and controllers; make sure engine tuning procedure is in place and
comply to the OEM requirements and provide field personnel with engine tuning
training to ensure any tuning issues can be identified and rectified properly.
References
1. Donahue RJ, Draege WA, Schaefer NP (2009) Robust design of a large industrial natural gas
engine piston. In: Proceedings of the ASME internal combustion engine division 2009 spring
technical conference ICES2009, Milwaukee, Wisconsin, USA
2. Liu X, Newbury A (2016) Waukesha 7044 gas engine failure investigation, Technical report,
February 2016, APA Group, Australia
3. MPA Consulting Engineers (2008) Wallumbilla compressor station project design basis man-
ual, May 14, 2008, APA Group, Australia
Construction of Index System
of Comprehensive Evaluation of Equipment
Maintenance Material Suppliers
1 Introduction
X. Liu (*)
Academy of Armored Forces Engineering, Beijing, 100072 Beijing, China
Equipment Academy, Beijing, 101416 Beijing, China
e-mail: [email protected]
J. Liang
Equipment Academy, Beijing, 101416 Beijing, China
e-mail: [email protected]
2.1 Definition
The EMMS is the general designation of partners (corporations) who lawfully estab-
lish supply needed productions (or service) to the army and locate the top of the
equipment maintenance supply chain [2]. The basic mission of EMMS is to supply
material according to purchasing and maintenance need to the army and ensure the
army’s equipment maintenance material need either in peacetime or in wartime.
The military consumers have higher demand of quality and after service than
general consumers. They demand not only higher cost performance, but also
service based on whole life circle management provided by suppliers. That service
based on whole life circle management, which consist acquirement service, exer-
cise service and disposition service, is the service including the whole process of
production exploitation, design, manufacture, sell, logistics, training, using, main-
tenance, rejecting and recovery.
Comprehensive evaluation and choice of EMMS are different from other suppliers
in purpose and process of evaluation, choice of indexes and indexes weigh value
coefficient and evaluation method design.
In the past time, the war industry scientific research task of the military was
absolutely occupied by ten war industry group corporations which belonged to
the national defence science industry committee. The participation chance of
private enterprises was few. Along with the continual deepening of marketization
reform of military industry, the military industry market opened to private enter-
prises gradually and some enterprise groups having advanced technology and
abundant strength became the key cooperation partners. Evaluation of EMMS
must be especially careful because that most cooperation items belong to advanced
military core technic field, the productions and technic are proprietary, the demand
of security and safety is extraordinary strict and items relate to the creation of
military battle effectiveness and national military superiority. Therefore, the eval-
uation to EMMS should stand on the core competitive power and future develop-
ment capability of enterprises, including technic standard, development potential,
strategic cooperation capability, commerce credit, etc., to guarantee the success of
strategic cooperation.
purchasing, such as price, quality, delivery date and history performance, and so on
[3]. Evaluation index system under the condition of supply chain should add
indexes relating to strategy, sustainable development, complementarity, coopera-
tion compatibility, and so on. Evaluation indexes such as employee quality, man-
agement standard, history performance, enterprise credit, innovation capability of
productions and technics, and so on, are emphases of evaluation indexes. Further-
more, evaluation and examination should be done to construction level of informa-
tion system, consciousness of circumstance protection and so on, to confirm
suppliers chosen can carry out of responsibility of not only supply raw and
processed materials, accessories and parts to the army with high-quality, duly,
well and truly but also take part in the military’s activities such as new production
exploitation, stock control, quality improvement, service betterment, and so on with
active attitude, well diathesis and enough ability.
The EMMS evaluation based on supply chain management differs from traditional
military equipment supplier’s evaluation in index system. Even if the index was
same, the enactment of index weigh value is different. An example is the index of
price. Price is an very important index with a biggish weigh value in the evaluation
of traditional military equipment suppliers, but its enactment of weigh value
descend more in evaluation of EMMS based on supply chain management because
that the military customers pay more attention to other indexes, including quality
management ability, delivery ability, reaction flexibility, innovation ability, devel-
opment potential, diathesis and credit, whose enactment of indexes weigh value are
more bigger than the index of price’s.
The military supply chain management, whose core is to satisfy the arm’s need, is
not only an operation integration process which begins from the army (the end of
the chain) and ends at military equipment suppliers (the beginning of the chain), but
also a process of realizing the best join of every parts and steps on the chain. In
order to control the risk of military supply chain effectively, enterprises whose
competitive power of production is powerful, capability of core operation is
Construction of Index System of Comprehensive Evaluation of Equipment. . . 191
Explanations only about relational indexes which are different from common
suppliers are given in the following subsections.
Shorter the goods order advanced period is, better the supply chain’s respond ability
to the military customer’s requirement is, and fewer stock in trade is needed. Under
the condition of cooperation of supply and demand, the military customer wants the
order batch output and purchasing advanced period as smaller as possible, while the
supplier wants the military customer to increase order batch output. In the proces-
sion of real support, the military customer usually allows to increase order batch
output to satisfy the need of speedy reaction support and shorten the purchasing
advanced period. So the supplier uses the purchasing advanced period discount as
an instrument to attract the military customer increase order batch output. By this
way, the supplier’s total cost is reduced and the military customer’s support cost is
reduced at the same time, that is to say, both sides of supply and demand win the
income by the cooperation decision-making.
Together with unit production quote price, the military purchasing cost which is the
index should be consulted when the military analyses the cost is the gross of the
price. The index of military purchasing cost has a macroscopical point of view, so
the military can choose the supplier form the aspect of total price easily. It involves
review cost, palaver cost, goods order cost and goods delivery cost of EMMS. In the
period of T, assuming the production current rate of the supplier by P, the stock
quantity by N, goods order cost spent by C0, the purchasing cost of unit production
is:
192 X. Liu and J. Liang
Table 1 (continued)
One level indexes Two level indexes Three level indexes
Equipment status Facility advanced level
Facility using
Facility time using rate
Facility disrepair rate
Status of manufacture and Total throughput ability & total output
production Productivity of entire personnel labour
Ability of market control Market share
Market overlap
Market strain capacity
Quality system Quality system satisfaction
percentage
Quality engineer percentage
Informationization Outfit of information Integrative information equipment scale
degree equipment Improvement level of integrative infor-
mation equipment scale
Outfit of Information Full-time information personnel rate
personnel Improvement level of full-time infor-
mation personnel rate
Using of information Integrative using scale of information
equipment equipment
Improvement level of integrative using
scale of information equipment
Factors of person Integrated diathesis expo-
nential of high-grade
manager
Training expense per person
Training time per person
Universities & colleges per-
sonnel proportion
Staff demission rate
Environment of Geography position/km
enterprise Economy technique level
Cooperation ability Amalgamation of enterprise
of enterprise strategic target
Compatibility of enterprise
culture
Information share degree
Consciousness of
safeguarding the army
Characters of Credit Credit of repaying loan
supplier Keeping appointment scale
Plan realization scale
Bad credit record scale
(continued)
194 X. Liu and J. Liang
Table 1 (continued)
One level indexes Two level indexes Three level indexes
Kind of enterprise National property percentage
Private property percentage
Foreign property percentage
Security Security accident record
Security accident record in a year
Record of economy expense by security
accident
Difficulty economy expense by diffi-
culty security accident
Secrecy Blow-the-gaff affair record
Blow-the-gaff affair record in recent
5 years
Carrying out of secrecy system
Related reports of national secrecy
department
P N þ C0
C¼ ð1Þ
N
The index of delivery flexibility, which is the ability to adjust and plan the delivery
plan by the supplier when external condition changes, can reflect the supplier’s
reaction ability changing with the military’s require. Apparently, the faster the
reflection speed is, the more advantageous the military gains [4].
Qmin Qmax
FOQ ¼ ð2Þ
Q
The index of variety flexibility, which can be expressed by the production’s variety
amount synchronously produced by EMMS, reflects the ability of changing the
variety of production by EMMS.
In order to complete support mission, the supplier should fleetly design and produce
new type equipment fitting the campaign demand or ameliorate existing equipment.
The index of smart reaction reflects the reaction ability of EMMS when variety of
production needed by the military changes. The stronger the reaction ability is, the
shorter time needed is.
The index of feedback degree of military demand reflects the recognition degree of
EMMS to military customer’s demand. Military’s needs will change when cam-
paign circumstance changes. Military customer’s problems will be settled in time if
EMMS pay enough attendance to their demand. The index of feedback degree of
196 X. Liu and J. Liang
military demand expressed by the time quantum starting from new demand is put
forward by the military customer and ending in settle program is sent to purchasing
decision-making department of the military.
5 Tag
References
1. Mao Z (2007) Suppliers evaluation and choice in the management of suppliers chain. Master
Thesis of Hefei Industry University
2. Peng Y, Wu Q (2007) Establish process of military material suppliers partnership. China
Logistics and Purchasing, vol 24, p 40
3. Wang J (2005) Proposition of military suppliers evaluation index system construction. National
Defence Technology Base 5:1–4
4. Yu F (2007) Research on suppliers evaluation and choice based on supply chain management.
Master Thesis of Wuhan Institute of Technology
Addressing Missing Data for Diagnostic
and Prognostic Purposes
1 Introduction
proposed over the years [1, 2]. Condition Based Maintenance and Prognostics and
Health Management (CBM/PHM) [3], which is founded on the diagnostics and
prognostics principles, has been increasingly popular over the recent years. It
consists of three steps [3–6], and is presented in Fig. 1.
The success of CBM/PHM is heavily dependent on the quality of information
used due to its sequential structure (Fig. 1), and missing data is one of the major
issues that affect it. It is a frequent phenomenon in industry that can manifest in
various ways with the most dominant being sensor failure [7]. While the amount of
measurements recorded increases, so is the probability of missing data occurring.
The presence of missing values may compromise the information contained within
a set, introducing bias or misleading results [8, 9]. Consequently, as noted in [9]), it
is important to address this matter. To deal with this problem, various methodolo-
gies to recover such data, called imputation techniques, have been developed. To
the authors’ knowledge, none of them has yet been applied to centrifugal compres-
sor data. The purpose of this paper is to apply the most widely used imputation
techniques in order to identify the most suitable technique.
2 Literature Review
with the same neurons (Best Matching Unit—BMU) or their neighbours [30]. The
map is constructed with the available information. Then the data with the missing
values are fed to the map to calculate their BMUs. The missing measurements are
estimated as their corresponding BMU values of the respective weight vectors.
K-Nearest Neighbours
Assuming a variable within a set contains a missing value, KNN imputation
(method 9) uses the K other variables that don’t have a missing value at the same
time stamp and are most similar to that variable, to estimate the missing sample
([20]; [31]) The number of neighbours (K ) used affects strongly the performance of
this method.
Bayesian Principle Components Analysis
PCA is the linear projection (scores) of data where the retained variance is maxi-
mum (principal components) [35]. PPCA [36], is an extension where the principal
components are assumed to have a prior distribution. BPCA (method 10) [35], is an
extension where the optimum number of principal components is selected
automatically.
Multiple Imputation
MI (method 11) ([14]; [37]; [24]) takes into account the uncertainty caused by the
existence of missing data, by creating m complete sets ([17]; [23]; [24]; [25]).
Usually, a small number of sets is adequate m ¼ 3 5. The backbone of MI is data
augmentation algorithm ([24]), a two-step iterative procedure where missing values
are simulated.
There are three mechanisms of missing data [10–12, 14, 16, 21, 23–25, 29]:
(i) missing completely at random, (ii) missing at random (MR), (iii) missing not
at random. Most methods, can perform only under the assumption of the first two
types [11, 14, 16, 21]. In this work MR type was assumed.
The information employed for this study was taken from an operational indus-
trial centrifugal compressor. After a preliminary examination, it was observed that
in 92% of the sets, the missing data had the form which can be found in Fig. 2. For a
specific variable within a set there is a single group of missing values with observed
data before and after it. This type of missing data is referred to in this paper as
continuous missingness and was the focus of the project.
For analysis, a complete set of 474 samples containing 25 variables was selected.
Five percentages of missing data were simulated: 1, 5, 15, 25 and 50%. At any time,
only one variable within the set contained missing values, since as stated above, at
any time only one variable presented missing data. For each percentage and the
chosen variable, a sliding window with span equal to the percentage of missingness
was translated into samples; started from the beginning of the data set, the window
Addressing Missing Data for Diagnostic and Prognostic Purposes 201
slid across the signal and removed the respective samples, creating new sets. This
way, the effect of the position of missing information on the quality of imputation
was also considered. Hence, for each variable the new sets created were: 470 for
1%, 451 for 5%, 404 for 15%, 356 for 25% and 238 for 50%. In total, the number of
sets created and analysed was (470 þ 451 þ 404 þ 356 þ 238)∗25 ¼ 47975. In
order to benchmark the performance of the imputation techniques the normalised
root mean square error (NRMSE) between the predicted and actual values was
employed, as given in [38]).
The results regarding the NRMSE for 50% of missingness are presented in
Fig. 4. In the x-axis, ranging from 1–25, are the variables within the set while in
the y, ranging from 1–11, are the methods used for imputation. The graph is
separated in a number of boxes, which are the combination of each variable and
each method. For example, box [1,9] corresponds to the results of applying method
1 to variable 9. Within each box there are several lines, indicating the location of the
missing values. For 50%, as stated previously, there are 238 sets for each variable.
From top to bottom, the 1st line (top) corresponds to a set where the missing data
are found in the beginning, while the 238th line (bottom) corresponds to a set where
the missing data are found in the end. These locations can be seen in Fig. 3. The
lines are colour coded based on their NRMSE value. Although NRMSE ranges
from 1 (no fit) to 1 (perfect fit), for scaling reasons the results range from 1,
dark blue indicating poor performance, to 1, dark red yielding perfect estimation.
For box [1,9], it can be observed that method 1 performs inadequately as it is filled
with a variety of blue lines corresponding to low NRMSE values.
Going through the results, it is evident that multivariate methods were superior
to univariate techniques employed as indicated by the high NRMSE values (red
lines in the colour map of Fig. 4). It can be seen that regardless the percentage or the
202 P. Loukopoulos et al.
5 Conclusions
References
1. Kothamasu R, Huang SH, VerDuin WH (2006) System health monitoring and prognostics—a
review of current paradigms and practices. Int J Adv Manuf Technol 28(9–10):1012–1024.
doi:10.1007/s00170-004-2131-6
2. Lee J, Wu F, Zhao W, Ghaffari M, Liao L, Siegel D (2014) Prognostics and health manage-
ment design for rotary machinery systems—reviews, methodology and applications. Mech
Syst Signal Process 42(1–2):314–334. doi:10.1016/j.ymssp.2013.06.004
3. Vachtsevanos G, Lewis F, Roemer M, Hess A, Wu B (2006) Intelligent fault diagnosis and
prognosis for engineering systems. Wiley, Hoboken, NJ. doi:10.1002/9780470117842
4. Jardine AKS, Lin D, Banjevic D (2006) A review on machinery diagnostics and prognostics
implementing condition-based maintenance. Mech Syst Signal Process 20(7):1483–1510.
doi:10.1016/j.ymssp.2005.09.012
5. Peng Y, Dong M, Zuo MJ (2010) Current status of machine prognostics in condition-based
maintenance: a review. Int J Adv Manuf Technol 50(1–4):297–313. doi:10.1007/s00170-009-
2482-0
6. Sikorska JZ, Hodkiewicz M, Ma L (2011) Prognostic modelling options for remaining useful
life estimation by industry. Mech Syst Signal Process 25(5):1803–1836. doi:10.1016/j.ymssp.
2010.11.018
7. Brown ML, Kros JF (2003) Data mining and the impact of missing data. Ind Manag Data Syst
103(8):611–621. doi:10.1108/02635570310497657
8. Pantanowitz A, Marwala T (2009) Evaluating the impact of missing data imputation. In:
Advanced data mining and applications. Lecture Notes in Computer Science, vol 5678, pp
577–586. doi:10.1007/978
9. McKnight PE, McKnight KM, Souraya Sidani AJF (2007) Missing data: a gentle introduction.
The Guilford Press, New York
204 P. Loukopoulos et al.
10. Acock AC (2005) Working with missing values. J Marriage Fam 67(4):1012–1028. doi:10.
1111/j.1741-3737.2005.00191.x
11. Baraldi AN, Enders CK (2010) An introduction to modern missing data analyses. J Sch
Psychol 48(1):5–37. doi:10.1016/j.jsp.2009.10.001
12. Batista GEAPA, Monard MC (2003) An analysis of four missing data treatment methods for
supervised learning. Appl Artif Intell. doi:10.1080/713827181
13. Donders a RT, van der Heijden GJMG, Stijnen T, Moons KGM (2006) Review: a gentle
introduction to imputation of missing values. J Clin Epidemiol 59(10):1087–1091. doi:10.
1016/j.jclinepi.2006.01.014
14. Enders CK (2001) A primer on maximum likelihood algorithms available for use with missing
data. Struct Equ Model Multidiscip J 8(1):128–141. doi:10.1207/S15328007SEM0801_7
15. Enders CK (2010) Applied missing data analysis. The Guilford Press, New York
16. Graham JW (2009) Missing data analysis: making it work in the real world. Annu Rev Psychol
60:549–576. doi:10.1146/annurev.psych.58.110405.085530
17. Horton NJ, Kleinman KP (2007) Much ado about nothing: a comparison of missing data
methods and software to fit incomplete data regression models. Am Stat. doi:10.1198/
000313007X172556
18. Ilin A, Raiko T (2010) Practical approaches to principal component analysis in the presence of
missing values. J Mach Learn Res 11:1957–2000
19. Junninen H, Niska H, Tuppurainen K, Ruuskanen J, Kolehmainen M (2004) Methods for
imputation of missing values in air quality data sets. Atmos Environ 38(18):2895–2907.
doi:10.1016/j.atmosenv.2004.02.026
20. Li L, Li Y, Li Z (2014) Missing traffic data: comparison of imputation methods. IET Intell
Transp Syst 8(1):51–57. doi:10.1049/iet-its.2013.0052
21. Little RJA, Rubin DB (2002) Statistical analysis with missing data, 2nd edn. Wiley, Hoboken,
NJ. doi:10.1002/9781119013563
22. Myrtveit I, Stensrud E, Olsson UH (2001) Analyzing data sets with missing data: An empirical
evaluation of imputation methods and likelihood-based methods. IEEE Trans Softw Eng 27
(11):999–1013. doi:10.1109/32.965340
23. Pigott TD (2001) A review of methods for missing data. Educ Res Eval 7(4):353–383. doi:10.
1076/edre.7.4.353.8937
24. Schafer JL (2000) Analysis of incomplete multivariate data. Chapman & Hall/CRC, London.
doi:10.1201/9781439821862
25. Schafer JL, Graham JW (2002) Missing data: our view of the state of the art. Psychol Methods
7(2):147–177. doi:10.1037/1082-989X.7.2.147
26. Stacklies W, Redestig H, Scholz M, Walther D, Selbig J (2007) pcaMethods – a bioconductor
package providing PCA methods for incomplete data. Bioinformatics 23(9):1164–1167.
doi:10.1093/bioinformatics/btm069
27. Yozgatligil C, Aslan S, Iyigun C, Batmaz I (2013) Comparison of missing value imputation
methods in time series: the case of Turkish meteorological data. Theor Appl Climatol 112
(1–2):143–167. doi:10.1007/s00704-012-0723-x
28. Oba S, Sato M, Takemasa I, Monden M, Matsubara K, Ishii S (2003) A Bayesian missing value
estimation method for gene expression profile data. Bioinformatics 19(16):2088–2096. doi:10.
1093/bioinformatics/btg287
29. Qu L, Li L, Zhang Y, Hu J (2009) PPCA-based missing data imputation for traffic flow
volume: a systematical approach. IEEE Trans Intell Transp Syst 10(3):512–522. doi:10.1109/
TITS.2009.2026312
30. Rustum R, Adeloye AJ (2007) Replacing outliers and missing values from activated sludge
data using Kohonen self-organizing map. J Environ Eng. http://doi.org/10.1061/(ASCE)0733-
9372(2007)133:9(909)
31. Troyanskaya O, Cantor M, Sherlock G, Brown P, Hastie T, Tibshirani R, Botstein D, Altman
RB (2001) Missing value estimation methods for DNA microarrays. Bioinformatics 17
(6):520–525. doi:10.1093/bioinformatics/17.6.520
Addressing Missing Data for Diagnostic and Prognostic Purposes 205
32. Ljung L (1999) System identification: theory for the user, 2nd edn. Prentice Hall, Englewood
Cliffs
33. Moahmed TA, Gayar NE, Atiya AF (2014) Forward and backward forecasting ensembles for
the estimation of time series missing data. Lect Notes Comput Sci 8774:93–104. doi:10.1007/
978-3-642-12159-3
34. Folguera L, Zupan J, Cicerone D, Magallanes JF (2015) Self-organizing maps for imputation
of missing data in incomplete data matrices. Chemom Intel Lab Syst 143:146–151. doi:10.
1016/j.chemolab.2015.03.002
35. Bishop CM (1999) Variational principal components. In: 9th International conference on
artificial neural networks ICANN 99, No. 470, pp 509–514. doi:10.1049/cp:19991160
36. Tipping ME, Bishop CM (1999) Probabilistic principal component analysis. J R Stat Soc
Series B Stat Methodology 61(3):611–622. doi:10.1111/1467-9868.00196
37. Harel O, Zhou X-H (2007) Multiple imputation: review of theory, implementation and
software. Stat Med 26(16):3057–3077. doi:10.1002/sim.2787
38. Ljung L (2015) System identification toolbox TM user β€TM s guide. The MathWorks, Inc
Imperfect Coverage Analysis
for Cloud-RAID 5
Acronyms
1 Introduction
2 Cloud-RAID 5 System
There are seven levels in the traditional RAID [12] architecture. While the proposed
methodology is theoretically applicable to all the levels, we use the cost-effective
level 5 to illustrate the cloud-RAID architecture and the proposed reliability
evaluation methods.
In RAID 5, data are divided into blocks and are striped across multiple disks that
form an array [13]. Distributed parity is used, where the parity stripes are distrib-
uted across all the disks. In general a RAID architecture can be implemented using
disk drives with different sizes. The total usable storage space, however, is limited
by the disk with the smallest size.
Figure 1 shows the general structure of an example cloud-RAID 5 system, where
the data are split into blocks or stripes with parity stripes being distributed across
five disk drives from different cloud providers. The blocks on the same row of the
array form a virtual disk drive with one of them storing the parity information,
which provides data redundancy and thus fault tolerance. If any disk drive (or cloud
storage provider) fails or is not available, the entire storage system can still work
due to the use of data redundancy. Particularly, the failed or unavailable stripe can
be recovered using the parity stripe and remaining data stripes through, for exam-
ple, the exclusive OR operation.
In the context of reliability analysis, the example cloud-RAID 5 system can be
modeled using a 4-out-of-5 model, where the system functions or is reliable if at
least four disks are functioning correctly.
User
Internet
Fig. 1 Architecture of an example cloud-RAID 5 (adapted from Liu and Xing [8, 9])
210 L. Mandava et al.
In this section we explain the imperfect fault coverage, and binary decision diagram
(BDD) models.
Models that consider effects of imperfect fault coverage are known as imperfect
fault coverage models or simply coverage models [11]. There are three types of
coverage models depending on fault tolerant techniques adopted [14]: element level
coverage (ELC) or single-fault model, fault level coverage (FLC) or multi-fault
model, and performance dependent coverage. This paper mainly concentrates on
reliability analysis of the cloud-RAID 5 system with ELC and FLC, which are
explained below.
With ELC, the fault coverage probability of an element is independent of states
of other system elements. The ELC model is applicable or appropriate when the
selection among the redundant elements is made based on self-diagnostic capability
of individual elements. Effectiveness of the system recovery mechanism relies on
the occurrence of individual element faults. Under ELC, a multi-element system
can tolerate multiple co-existing single element faults. However, for any given
element fault, the success or failure of the recovery mechanism (i.e., the coverage
probability) is not dependent on faults in other elements.
Under FLC, the fault coverage probability is dependent on the number of failed
elements belonging to a particular group. The FLC model is applicable or appro-
priate when the selection among available system elements varies between initial
and subsequent element failures. Effectiveness of the system recovery mechanism
relies on the occurrence of multiple element faults within a certain recovery
window. The FLC model is typically applied in modeling, for example, computer
control systems used in aircrafts applications [15, 16], and multi-processor systems
in a load-sharing environment [17].
BDD is the state-of-the-art data structure for Boolean logical function representa-
tion and manipulation. A BDD is a rooted, directed acyclic graph model based on
Shannon decomposition rule of (1) [11, 18].
g e h ¼ iteðx; G1 ; G0 Þ e iteðy; H 1 ; H 0 Þ
8
< iteðx; G1 e H 1 ; G0 e H 0 Þ indexðxÞ ¼ indexðyÞ
¼ iteðx; G1 e h; G0 e hÞ indexðxÞ < indexðyÞ ð2Þ
:
iteðy; g e H 1 ; g e H 0 Þ indexðxÞ > indexðyÞ
p(x) q(x)
f f
x=0 x=1
212 L. Mandava et al.
4 Combinatorial Methods
We use an ELC model representing the behavior of the system in response to the
occurrence of a component fault, as shown in Fig. 3.
The entry point to the ELC model signifies the occurrence of a component fault,
and the three exits correspond to three possible outcomes [20]. Transient restoration
exit (R) is taken when the fault is transient and can be treated without discarding the
component. Permanent coverage exit (C) is taken when the fault nature is perma-
nent, and the faulty component must be discarded. Single-point failure exit (S) is
taken when the component fault (by itself) causes the entire system to fail. The three
exits form a partition of the event space, thus the three exit probabilities (r, c, and s)
sum to one.
Let qd(t) denote the failure probability of disk d (d ¼ 1, 2, 3, 4, 5) at time t. In the
case of disk d following the exponential distribution with parameter λd, we have
qd(t) ¼ 1exp(λd*t). Based on the ELC model, we have the probability that disk
d does not fail (denoted by n(d )), fails covered (c(d)), and fails uncovered (u(d )) as
Based on the simple and efficient algorithm in [20, 21], we define two events for
the considered cloud-RAID 5 system:
E1: no disks experience a single-point of failure
E2: at least one disk experiences a single-point of failure.
Define E as an event that the system fails. Based on the total probability law, the
unreliability of the cloud-RAID 5 system with ELC can be obtained as.
Then, Pr(E|E1) can be evaluated using the BDD method without considering
the ELC.
Figure 4 illustrates the BDD model for evaluating Pr(E|E1) of the example cloud-
RAID 5 system, which has a well-defined 4-out-of-5 lattice structure. The solid
edge represents that the corresponding disk is working and the dashed edge
represents the disk is failed. Sink nodes ‘0’ and ‘1’ represent the system working
and failed states, respectively.
Based on the generated BDD model, Pr(E|E1) can be evaluated as
2 3 1
3 4 1
4 5 1
0 0
214 L. Mandava et al.
PrðEjE1 Þ ¼ qe q
1 2
e þ qe 1 qe qe þ qe 1 qe 1 qe qe
1 2 3 1 2 3
4
e e
þq1 1 q2 1 q3 1 q4 q5 þ 1 q1 qe
e e e e e
2 q3
þ 1 qe e e e e e
1 q2 1 q3 q4 þ 1 q1 q2 1 q3
e 1 qe qe
4
5
e e
þ 1 q1 1 q2 q3 q4 þ 1 q1 1 q2 q3 1 q4 qe
e e e e e e
5
e e e
þ 1 q1 1 q2 1 q3 q4 q5 e e
ð10Þ
With Pr(E1) and Pr(E|E1) the unreliability of the cloud-RAID 5 system consid-
ering effects of ELC can be given by (7). The reliability of the cloud-RAID 5 system
can thus be obtained as
Under FLC, the fault coverage probability ci cidepends on the number of failed
elements i that belong to a specific group. Thus, calculation of ci depends on a set of
disks that already failed. In general for a k-out-of-n system, the system fails after
(n k + 1) failures (irrespective of fault coverage), thus ci for i (n – k + 1) are not
applicable and can be considered as zero [17]. For the example Cloud-RAID
5 system, n ¼ 5 and k ¼ 4. Thus only c1 is used. By definition, c0 is 1.
According to [17], Eq. (12) gives a method for evaluating ci for systems with
n identical components following the exponential time-to-failure distribution with
constant failure rate λ.
ci ¼ exp½-ðn-iÞ*λ*τ ð12Þ
where i represents the failure number, τ represents the recovery window time.
Applying the BDD method with the consideration of FLC, we have the BDD
model as shown in Fig. 5 where only paths to sink node ‘0’ (representing the
function of the example cloud-RAID 5 system) are shown. Coverage probabilities
c0 ¼ 1 and c1 are added to the corresponding paths of the BDD model. Based on the
BDD model, the example Cloud-RAID 5 system remains working if at least four
disks are working. If more than two disks fail, then the entire system fails.
Whenever a single disk is in a failed state it can be covered with a coverage factor
c1 and the system is still functioning.
It is assumed that all the five disks of the example cloud-RAID 5 system have the
same failure rate and recovery window time so that they have the same coverage
factor c1 based on (12). In the case of disks with non-identical failure rates
(in general, failure time distributions) or recovery window time, Eq. (12) needs to
Imperfect Coverage Analysis for Cloud-RAID 5 215
2 2
3 3
4 4
5 5
c0 c1
This section presents numerical evaluation results for the reliability of the example
cloud-RAID 5 system to demonstrate effects of ELC and FLC.
216 L. Mandava et al.
Fig. 6 Reliability of
Cloud-RAID 5 system
with ELC
The reliability of the example Cloud-RAID 5 system with FLC is given by (13).
Equation (12) is used to compute c1, where λ of all the disks is assumed to be
0.0001/h. Table 3 lists values of c1 calculated for various values of τ (in hours) and
corresponding system reliability RFLC evaluated for t ¼ 1000 h using (13). Figure 7
plots the coverage probabilities as the recovery window time increases. Figure 8
illustrates the system reliability trend as the coverage probability increases.
From Table 3 we can observe that when the value of c1 is 1, the system reliability
is the highest and is actually equal to the reliability of system with perfect fault
coverage. As shown in Table 3 and Fig. 8, as the coverage probability decreases, the
system reliability gets worse.
218 L. Mandava et al.
Fig. 8 Reliability of
Cloud-RAID 5 with FLC
6 Conclusion
Cloud-RAID 5 based on single-bit parity code is one of the effective solutions for
enhancing data reliability in the cloud storage. The existing reliability modeling
methods for the cloud-RAID systems assumed that the system fault detection and
recovery mechanism is perfect, which is often not true in practice. This paper makes
new contributions by relaxing this assumption through BDD-based combinatorial
Imperfect Coverage Analysis for Cloud-RAID 5 219
methods for the reliability analysis of cloud-RAID 5 subject to ELC or FLC. The
methods are applicable to heterogeneous disks with arbitrary types of time-to-
failure distributions. As demonstrated through numerical results, failure to consider
the imperfect fault coverage may lead to inaccurate (overestimated) system reli-
ability results, which can mislead the system design and optimization activities.
References
1. Deng J, Huang S C H, Han Y S, Deng J H (2010) Fault tolerant and reliable computation in
cloud computing. In: Proceedings of IEEE Globecom workshops, Miami, FL, pp 1601–1605
2. Erl T, Puttini R, Mahmood Z (2013) Cloud computing concepts, technology & architecture,
the Prentice Hall service technology series. Prentice Hall, Upper Saddle River, NJ
3. Wang C, Xing L, Wang H, Dai Y, Zhang Z (2014) Performance analysis of media cloud-based
multimedia systems with retrying fault-tolerance technique. IEEE Syst J Spl Iss Recent Adv
Cloud-Based Multimedia Syst 8(1):313–321
4. Robinson G, Narin A, Elleman C (2013) Using Amazon web services for disaster recovery.
Amazon web services
5. Bausch F (2014) Cloud-RAID concept. http://blog.fbausch.de/cloudraid-3-concept/. Accessed
May 2016
6. Jin T, Yu Y, Xing L (2009) Reliability analysis of RAID systems using repairable k-out-of-n
modeling techniques. In: The international conference on the interface between statistics and
engineering, Beijing, China
7. Fitch D, Xu H (2013) A RAID-based secure and fault-tolerant model for cloud information
storage. Int J Softw Eng Knowl Eng 23(5):627–654
8. Liu Q, Xing L (2015a) Hierarchical reliability analysis of multi-state Cloud-RAID storage
system. In: Proceedings of international conference on quality, reliability, risk, maintenance,
and safety engineering, Beijing, China
9. Liu Q, Xing L (2015b) Reliability modeling of cloud-RAID-6 storage system. Int J Future
Comput Commun 4(6):415–420
10. Zhang R, Lin C, Meng K, Zhu L (2013) A modeling reliability analysis technique for cloud
storage system. In: Proceedings of 15th IEEE international conference on communication
technology, Guilin, China, pp 32–36
11. Myers A (2010) Complex system reliability, 2nd edn. Springer series in reliability engineering
12. Jin T, Xing L, Yu Y (2011) A hierarchical Markov reliability model for data storage systems
with media self-recovery. Int J Reliab Qual Saf Eng 18(1):25–41
13. Patterson D A, Chen P, Gibson G, Katz R H (1989) Introduction to Redundant Arrays of
Inexpensive Disks (RAID). In: Proceedings of thirty-fourth IEEE computer society interna-
tional conference: intellectual leverage, Digest of papers, San Francisco, CA, USA, pp
112–117
14. Amari S V, Myers A and Rauzy A (2007) An efficient algorithm to analyze new imperfect fault
coverage models. In: Proceedings of annual reliability and maintainability symposium
15. Myers A (2007) k-out-of-n: G system reliability with imperfect fault coverage. IEEE Trans
Reliab 56:464–473
16. Myers A, Rauzy A (2008) Assessment of redundant systems with imperfect coverage by means
of binary decision diagrams. Reliab Eng Syst Saf 93(7):1025–1035
17. Amari SV, Myers A, Rauzy A, Trivedi K (2008) Imperfect coverage models: status and trends.
In: Misra KB (ed) Handbook of performability engineering. Springer, Berlin
18. Xing L and Amari S V (2015) Binary decision diagrams and extensions for system reliability
analysis, Wiley-Scrivener, MA, isbn:978-1-118-54937-7
220 L. Mandava et al.
19. Xing L, Wang H, Wang C, Wang Y (2012) BDD-based two-party trust sensitivity analysis for
social networks. Int J Secur Netw 7(4):242–251
20. Amari SV, Dugan JB, Misra RB (1999) A separable method for incorporating imperfect fault-
coverage into combinatorial models. IEEE Trans Reliab 48:267–274
21. Xing L, Dugan JB (2002) Analysis of generalized phased mission system reliability, perfor-
mance and sensitivity. IEEE Trans Reliab 51(2):199–211
Research on Data Analysis of Material System
Based on Agent Simulation
Abstract There are two sides of system reliability parameters, such as basic reliability
and mission reliability, for reliability requirement in the standard, with fine relevance and
coordination. It reflects in the determination of the parameters and index and also in the
prediction and allocation for basic reliability and mission reliability. In the paper, it sets
up a fine mission reliability simulation model of materiel system, and analyzes from data
collection and process, to lay the good foundation for analyzing dynamically mission
reliability and making a perfect mission reliability simulation concept of materiel system.
1 Introduction
There are two sides of system reliability parameters, such as basic reliability and
mission reliability, for reliability requirement in the standard, with fine relevance and
coordination. It reflects in the determination of the parameters and index and also in the
prediction and allocation for basic reliability and mission reliability. So it emphasizes
the research on the analysis of mission reliability of materiel system to better the analysis
of reliability simulation and settle the complexity of analysis of reliability simulation [1].
Adaptive Agent can gather relative reliability data of itself with independent
module of data gathering and processing, and computers statistically and shows
in the chart. The metadata main includes:
• The failure times of a component: The failure times in the mission time for the
working unit, as UFM.
Control
Environment Environment Dependence
Simulation Agent
Conmunication
Reliability
Effect
System System
System ……
Agent Agent
Simulation
(m)
Component …… ……
Component
Agent 2
Agent 2
Fig. 1 The model of system structure and relationship of materiel system simulation
• Single accidental shutdown of a component: The time is when the failure unit is
from into the state of failure to leave the state of failure in the time of mission.
• Standby time of a component: The time is when the working unit is in the state of
standby in the time of mission.
• Single-time-to-repair of a component: The time is when the working unit is in
the state of maintenance shown in Fig. 1.
On the basis of metadata, Adaptive Agent processes data statically and outputs in
some forms. So there is main output as follows:
• Failure rate of a component, as λM: It is the basic parameter on the product
reliability. It is the ratio of the total of failure time to the total of mission time in
the simulation course of mission reliability for the working unit.
PM
UF
UTMi
i¼1
λM ¼ ð1Þ
MT
Research on Data Analysis of Material System Based on Agent Simulation 223
Thereinto:
λM—Failure rate of a component in mission profile.
UTMi—unit downtime of a component.
UFM—Number of unit failure.
MT—Mission time.
The data of data collecting module of system simulation Agent is mainly debugging
and analytical. Analytical data includes mainly the following metadata:
• Single standby time of a system: The time is when the system in the state of
standby in the time of mission.
• Single-time-to-repair of a system: The time is when the system is from the start
state of maintenance to repair faults.
• The number of maintenance: There are the times for a system in the state of
maintenance.
• Single-time-to-repair of a critical failure of a system: The time is when the
system is from the start state of maintenance to repair a critical failure.
• The number of critical failures of a component: There are the times for which a
component takes occur critical failures.
• The number of critical failures of a system, as SFM: There are the times for
which a system takes occur critical failures, as the summation of critical failures
of every components.
On the basis of metadata, System Simulation Agent processes data statically and
outputs in some forms. The main output is as follows:
• Mean Time Between Critical Failures of a system as MTBCF: It is the parameter
relative to mission reliability. It is the ratio of the total of mission time to the
total of failure in the series of mission profiles.
In the simulation course of mission reliability for the working unit.
MT
MTBCF ¼ ð2Þ
SFM
Thereinto:
MTBCF—Mean Time Between Critical Failures of a system.
MT—Mission Time.
SFM—Number of critical failure of a system.
Mission time to mean recovery function of a system, as MTTRF: In the simulation
course of the mission profile, it is the mean time for excluding critical failures. It
is a kind of maintenance parameter relative to mission success. In the given
mission profile and maintenance condition, it is calculated as follows:
224 Y. Shen et al.
PM
SF
STTRMi
i¼1
MTTRF ¼ ð3Þ
SFM
Thereinto:
MTTRF—Mission time to mean recovery function of a system.
STTRM—The time of repairing a critical failure.
SFM—Number of critical failure of a system.
Operational Availability of a system, as Ao: It is a kind of availability parameter
relative to up time and down time in the simulation course of a system. It is
calculated as follows:
P
N P
U
WTBFi þ STk
i¼1 k¼1
Ao ¼ ð4Þ
P
N P
U P
M
WTBFi þ STk þ TTRj
i¼1 k¼1 j¼1
Thereinto:
Ao—Operational Availability of a system.
WTBF—Working time for no failure.
N—The number of samples of working time for no failure.
STk—Standby time of a system.
U—The number of samples of standby time.
TTR—The time of repairing a failure of system.
M—The number of samples of repairing time.
t1 t2 t3 t4 t5 t6
main main main
Failure tena Work Failure tena Work Alert Work Failure tena Work Alert Work Alert Work Failure
nce nce nce
(2) The times of mission success: There are the times of mission capable success-
fully. If a mission is successful, the submission is successful.
Environment Agent processes statistically the collected Metadata. The main
contains:
(1) Mean Time Between Failure: Mean Time Between Failure, Abbreviated as
MWTBF, is the mean work time between failure [2], as follows:
P
N
Li
i¼1
MWTBF ¼ ð6Þ
N
Thereinto:
MWTBF is Mean Time Between Failure.
Li is an effective sample of work time Between Failure.
N is the sample for the mean work time Between Failure.
(2) Mean Mission Capable Time is the mean time to execute a mission as is the
mean of mission time, as follows:
P
N
MT i
i¼1
MMT ¼ ð7Þ
N
226 Y. Shen et al.
Thereinto:
MMT is Mean Mission Capable Time.
MT is mission time.
N is the times of mission capable.
(3) Dependability: It is the rate of mission success [3], as follows:
M
RMS ¼ 100% ð8Þ
N
Thereinto:
Rms is dependability.
M are the times of mission success.
N is the total times of mission capable.
5 Conclusion
It devises three types of Agent, as Component Agent, System Agent and Environ-
ment Agent to express reliability relations dynamically of materiel system in the
mission with the change of mission profiles. Component Agent shows the basic
working unit in the reliability model and simulates the reliability situation of a
component. System Agent expresses a system object. Environment Agent simulates
the change of mission and working condition. Because of limited length, it does not
specify the simulation courses. But in the study, with an example of a communi-
cation product in the reference, it is effective for the simulation analysis results in
comparison with the means and results of the reference. So in the paper, it analyzes
data collection and process, defines the key elements for the analyses of mission
reliability based on Agent simulation, and lays the good foundation for setting up a
fine mission reliability simulation model of materiel system, analyzing dynamically
mission reliability and making a perfect mission reliability simulation concept of
materiel system.
References
1. Shen Y (2012) Research on reliability analysis methods for materiel system based on adaptive
agent simulation. PhD Thesis, Academy of Armored Force Engineering
2. Chang XZ (2002) Supportability engineering, vol 59. Weapon Industry Press, Beijing, pp 65–69
3. GJB 451A-2005 (2005) Reliability, maintainability and supportability terms
A Relationship Construction Method between
Lifecycle Cost and Indexes of RMSST Based on
BP Neural Network
Abstract In the paper, it divides LCC that affects the materiel effectiveness of the
performance objectives of RMSST into the cost of development and the cost of
support of operation and maintenance, and evaluates the cost of development, or the
cost of support of operation and maintenance on BP Neural Network. There is the
main limitation and insufficient of selecting BP Neural Network in addition to the
method itself. But it provides important supporting to study the relationship
between the performance objectives of RMSST and life cycle cost based on the
theory of BP Neural Network.
1 Introduction
BP Neural Network is the core of feed forward neural, with broad adaptability and
effectiveness [1, 2]. As we know, Life cycle cost is abbreviated to LCC, and
RMSST is in an abbreviated form of reliability, maintainability, supportability,
security and testability. For cost evaluation on stages of life cycle, it is hard to
express the relationship between cost and the indexes of RMSST in the general
linear relationship and model [3]. And if in a model, it does not set up the model to
the phase, in low precision and too much deviation. BP Neural Network sets up the
relationship between input and output on the study of the samples, in case of
ambiguity model, inadequacy information and unclear system, which takes advan-
tages on the modelling, prediction, decision. So in the paper, it divides LCC that
affects the materiel effectiveness of the performance objectives of RMSST into the
cost of development and the cost of support of operation and maintenance, and
evaluates the cost of development, or the cost of support of operation and mainte-
nance on BP Neural Network shown in Fig. 1.
Network Initialization
Sample bias
(Error)
For the problem of the modelling of the LCC and the indexes of RMSST, it is the
nonlinear mapping between the critical indexes of RMSST that affects the effec-
tiveness of materiel and the cost of development, or operation and maintenance
support. Because the three-layer of BP Network is arbitrary precision, it sets up the
input layer, hidden layer and output layer. When there are many nodes in a hidden
layer, it can add a hidden layer.
2.1 Input-Layer
1. Input Variable
In general, there are the following basic rules when selecting input variables:
• It can make a big difference for output, be detected or extracted.
• The variables are mutually exclusive or in quite small correlation.
So it takes the critical indexes of RMSST that affects the effectiveness of materiel,
such as the cost of development or the one of operation and maintenance
support, as the input variables, with mutual exclusiveness, no iteration and on
overlap.
2. Transfer Function
BP Neural Network is with the implicit layer of sigmoid and linear Output
Layer, with very good mapping capability. If the output layer is tan–sigmoid, the
range of output is from 1 to +1. While if the transfer function of output layer is
purelin, the output is any value. So the implicit layer is the function of sigmoid,
the output layer is linear function and broadens value field.
So for the transfer function, the input layer is the function of tansig(), the
middle layer is the function of logsig(), and the output layer is the linear function
of purelin() or the function of logsig().
3. Pre-processing of the input
It presents mainly scale variation, also named as normalization or standard-
ization, which limit the input in the interval of [0,1] or [1,1] by the Change of
Address.
When the dimension of every component is different, it is for the input
transferred separately in the value range. When in the same physical significance
and the dimension, it defines the maximum as xmax and the minimum as xmin in
the total data domain, and transfers unifiedly. For example, if the input is
transferred into the range of [0,1], by the follows:
xi xmin
xi ¼ ð1Þ
xmax xmin
230 Y. Shen et al.
Thereinto:
xi—Input Data;
xmin—the minimum of data change.
xmax—the maximum of data change.
xmid—the median of data change range.
If processed with nondimensionalization for primary data, it is by the follows:
xij
x∗
ij ¼ ð2Þ
sij
sffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
P 2
Xij Xj
sij ¼ ð3Þ
n1
2.2 Output-Layer
By and large, the output represents the functional goals of the system, as the desired
output, which is defined easier, for example the performance indexes, the categories
of classification problems or the functional value of nonlinear function. The output
of the network sketches out the cost of development, or operation and maintenance
support. So the transfer function of the output layer is the function of purelin() or
logsig().
Since the input is multidimensional, the model is multi-input and single-output.
2.3 Hidden-Layer
It makes a difference for the number of nerve cells in the hidden layer. For less, the
net is in poor acquisition capability, which can not summarize and embody the
sample rule for the training set. For more, it remembers non-regularity as noise,
which not only lowers generalization but also increases the training time. In
general, the common method of ensuring the best number of hidden-layer nodes
is cut-and-trial. It first sets fewer hidden-layer nodes, and then increases. It trains in
the same sample sets, and determines the number of hidden-layer nodes at the time
of minimum error of the net. For the initial value, it determines by the empirical
formula:
pffiffiffiffi
m¼ nl ð4Þ
A Relationship Construction Method between Lifecycle Cost and Indexes of. . . 231
Thereinto:
m—the number of hidden-layer nodes;
n—the number of input-layer nodes;
l—the number of output-layer nodes.
Since the net is nonlinear, it is for initial weight if the study is in the minimum in
the local, in the convergence and in much influence with the training speed. It takes
the initial weight as the random number in the range from 1 to 1.
For the middle and small size of the BP net, it takes best the algorithm as
Levenberg–Marquardt, which is in the best rate of convergence and lower size of
memory. But if it runs short of storage space, it selects other types of rapid
algorithm. For large-scale net, it selects best conjugate gradient algorithm of Scaled
or elastic gradient algorithm.
So for the point of the BP net, it trains in the algorithm of Levenberg–Marquardt.
3 Example Analysis
The cost of operation and maintenance is related to the service life of the materiel.
The service life is longer, and the cost of maintenance support is more. But it is
uncertain that the analysis of the cost of operation and maintenance. As the above
principle, it sets a model of BP neural network with five indexes of RMSST as the
independent variables and the cost of operation and maintenance as the dependent
variable. The original data as inputs is shown in Table 1 and corresponding pre-
processing data is given in Table 2.
Table 1 The table of original data (The cost of operation and maintenance Versus the related
performance of RMSST for some type of tank)
The cost of
operation and
maintenance
support in some
year (ten thousand
Material X1 X2 X3 X4 X5 yuan)
A 0.60 8.00 60.00 4.00 0.900 2129.00
B 0.50 6.00 40.00 4.50 0.700 1000.00
C 0.45 5.00 30.00 4.80 0.600 800.00
D 0.40 4.50 26.00 4.90 0.500 700.00
E 0.45 4.30 25.00 5.00 0.450 520.00
F 0.30 4.00 20.00 5.00 0.400 300.00
232 Y. Shen et al.
(a) (b)
10 -2 1
Train Data
0.9
Best Fit
Goal Y=T
0.8
-3
10
Mean Squared Error (mse)
Output~=1*Target+0.0015
0.7
0.6
10 -4 0.5
0.4
0.3
10 -5
0.2
0.1
10 -6 0
0 0.5 1 1.5 2 2.5 3 3.5 4 0 0.2 0.4 0.6 0.8 1
4 Epochs Target
Fig. 2 The training and test results: (a) Training error as a function of training epoch and (b)
Training output versus the target value
The test results are provided in Fig. 2. From Fig. 2, we can know that the
relationship between the lifecycle cost and indexes of RMSST is generally linear,
and the computational accuracy of the proposed method is acceptable.
A Relationship Construction Method between Lifecycle Cost and Indexes of. . . 233
4 Summary
There is the some presentation of the study in the task by the above analysis. It sets
up the relationship model between the performance objectives of RMSST and
development cost. It shows the veracity of BP Neural Network compared with
regression analysis. There is the main limitation and insufficient of selecting BP
Neural Network in addition to the method itself.
• It introduces the initial BP Neural Network model based on the performance
objectives of RMSST to evaluate the cost of development. To the concrete issue,
it fits the model according to the relative data, when it adjusts probably the
relative function and method, to improve accuracy.
• The generalization of BP Neural Network is associated with prediction of the
testing data. It need set and train the net based on the data. In view of the
complexity and specialty on materiel construction, and with the new peculiarity
of the newly developed materiel, it is different between current data and the
relationship of some performance objectives of RMSST and development cost.
So it changes for the performance objectives of RMSST that affect materiel
effectiveness.
In view of the complexity and specialty of materiel construction and with the
new characteristics of new developed materiel, it is different between the relation-
ship of the performance objectives of RMSST and development cost and current
data, so the objectives of RMSST affecting materiel effectiveness change, as the
input-layer of network probably changes. If the trained network predicts, it is biased
by the benchmark. It need more study and extendedly analyze on the basis of
modelling theory if with quite strong generalization and veracity. But it provides
important supporting to study the relationship between the performance objectives
of RMSST affecting materiel effectiveness and life cycle cost based on the theory
of BP Neural Network, and settles effectively the problems of setting up the cost
appraisal models by stages in the whole life cycle and improving the accuracy of the
models.
References
1. Shi Y, Han LQ, Lian XQ (2009) Design methods and example analysis of netual network. The
Press of Beijing University of Posts and Telecommunications, Beijing
2. Zhu K, Wang ZL (2010) Mastering MATLAB netual network. Publishing House of Electronics
Industry, Beijing
3. Shenying HC et al (2015) Relative RMSST trade-off analysis study
Asset Performance Measurement: A Focus
on Mining Concentrators
1 Introduction
In the last 40 years, the demand for platinum group metals (PGMs) has grown
rapidly [1] but in recent times, the price of the commodity has manifested
high volatility, somewhat in response to stricter standards for diesel-powered
vehicles, the general slowdown in the global economy, and the consequential
sales of stockpiles by car manufacturers [2]. With decreasing reserves of rich
ore bodies coupled with high volatility in price of the commodity, platinum
sector firms have to find ways of optimising the performance of engineered
assets throughout the mining value chain. Most firms are challenged to
produce more with existing natural resources and engineered assets [3], and
A. Snyman (*)
Lonmin Plc, Marikana, South Africa
e-mail: [email protected]
J. Amadi-Echendu
Department of Engineering and Technology Management, University of Pretoria, Pretoria,
South Africa
e-mail: [email protected]
Quality is the third and perhaps the most difficult component to measure since
concentrators have no distinct product which can be inspected in real time. A
plausible basis for determining the quality of the output is to use the recovery
after the run-of-mine has passed through the concentrator. However, the assay
results for recovery are only available three to five days after production. This
Total Time
Equipment
Time Available for Runtime
Downtime
Supply Shortage
External Events
Maintenance
Consequential
Plant Running,
no downtime
Breakdowns
Standby
Delays
1
Performance is also referred to as “Efficiency” in literature. Performance of the asset should not
be confused with the nameplate performance of a specific piece of equipment.
238 A. Snyman and J. Amadi-Echendu
To establish the ranges for the quality metrics, the average of the top 5% of
recovery figures was taken and this figure represents 100% quality.
3 Case Study
Technology has moved long since the days of manual data capturing and it
should be avoided for data of this importance [11–15]. The first data set
extracted was an indication of whether or not the concentrator was running
and, as suggested by Andersson and Bellgran [16], a simplistic solution was
opted for. In light of this suggestion, the mill at a concentrator was the best asset
to monitor runtime, but the electrical load on a mill was also taken into account
since mill running status can be misleading. It is safe to assume that a mill is
running contributively when it is consuming a significant amount of electricity,
typically 1.5 MW. The concentrator was subsequently deemed to be running
when the mill was both on and it consumed more than a specified amount of
electricity. This measure can easily be configured on a plant historian and
integrated to a system.
Components of the OEE time allocation picture in Fig. 1 are based on operator
selected reason recorded in the plant historian, e.g. when the operator selected a
reason which is categorised as an “affects lost time” reason, the downtime is
interpreted as lost time. All non-productive times are documented along with the
associated reasons for delay.
The actual mill feed rate used in the performance calculation (Eq. 3) was sourced
from the supervisory control and data acquisition (SCADA) system and averaged
over the timeframe for which the metric was calculated. The average of top 6 hourly
mill feed rates looked at history trends for the specific concentrator as well as the
specific ore type milled. The top 6 feed rates provided a measure of the nameplate
performance of the concentrator.
OEE is a difficult concept to apply in a mining concentrator environment due to
the absence of a quality parameter on a real time basis. Metallurgists normally
Asset Performance Measurement: A Focus on Mining Concentrators 239
4.1 Availability
No two concentrators are alike and this would have an impact on the results of
metrics being measured. Figure 2 shows the boxplots of the availability metric for
240 A. Snyman and J. Amadi-Echendu
100
95
Availability (%)
90
85
80
C1 C2 C3 C4 C5 C6 C7
Concentrator ID
120
110
100
Performance (%)
90
80
70
60
50
40
C1 C2 C3 C4 C5 C6 C7
Concentrator ID
It is noted that some of the concentrators have a fourth quarter performing above
100%. Although this is not impossible, it is not recommended as equipment should
not be pushed beyond the limits.
Firstly, the acceptable range for each of the metrics contributing to the quality
metric needed to be established. The Recovery figure for each concentrator was
recorded for the last two and a half years and the quality metrics’ data for
corresponding timespan was matched to it. After the erroneous data points were
brushed out, 500 data points per concentrator remained. The 25 values with the
Recovery figure in the top 5% of the so-called clean data was separated and used to
establish the ranges for the quality metrics.
242 A. Snyman and J. Amadi-Echendu
100.0
97.5
95.0
Quality (%)
92.5
90.0
87.5
85.0
C1 C2 C3 C4 C5 C6 C7
Concentrator ID
When the abovementioned metrics are in range, it can be assumed that the
recovery will be high and subsequently, the quality of the product produced will
be 100%. Upon calculation, the boxplot in Fig. 4 confirms the validity of the quality
metric as well as the heterogeneity in concentrators.
4.4 OEE
The OEE measure calculated shows the unique differences between each concen-
trator. It is noticed that most concentrators operate in a narrow band of OEE when
compared to itself, but the OEE range has wide averages lying between 45% and
80% when looking at different concentrators. Figure 5 clearly shows a spike in
August 2014s data which is attributed to the performance of each concentrator post
the strike (Fig. 6).
Asset Performance Measurement: A Focus on Mining Concentrators 243
130
120
110
100
90
OEE (%)
80
70
60
50
40
C1 C2 C3 C4 C5 C6 C7
Concentrator ID
120
110
100
90
%
80
70
60
50
40
Oct 2013
Oct 2014
Jun 2015
Aug 2014
Nov 2013
Dec 2013
Sep 2014
Nov 2014
Dec 2014
Jan 2015
Feb 2015
Mar 2015
Apr 2015
May 2015
C1 C2 C3 C4 C5 C6 C7
5 Summary
The articles Amadi-Echendu [6] and Amadi-Echendu et al. [5] make that point that
performance of an engineered asset should be approached from a value doctrine.
This article has briefly described a case study on the challenges encountered in
measuring the parameters for measuring asset performance and calculating OEE in
244 A. Snyman and J. Amadi-Echendu
practice. The challenges surround the proper definition of the parameters and
collection of pertinent data. Muchiri and Pintelon [15], Ljungberg [14], Dal et al.
[11], Kullh and Josefine [13], and Jonsson and Lesshammar [12] all reiterate that
data should be collected electronically and analysed. Our case study reveals that, in
practice, the cost doctrine still remains the primary driver for the application of the
OEE concept. Hence, the measurement of asset utilization is not a priority. Since an
asset is a thing of value, the effectiveness of an asset must depend on how it is used,
therefore, we argue that the measurement and assessment of the performance of an
engineered asset like a platinum mining concentrator should be approached from a
value doctrine so that the utilization is appropriately defined and measured.
References
1 Introduction
Assets are the lifeblood of most organizations. They may include digital assets,
human assets, and financial assets. Most companies also have physical assets. These
physical engineering assets (e.g. machinery, plant and equipment, etc.) can be used
to turn raw material into finished goods, supply electricity and energy, provide
transportation services, or control huge utility operations. Many organizations rely
heavily on these engineering assets to maintain and monitor daily operations.
During the lifecycle of these engineering assets, an enormous amount of data is
produced. The data is captured, processed and used in many computer information
systems such as Supervisory Control and Data Acquisition (SCADA) systems,
Facility Maintenance and Management Systems (FMMS), and Geographic Infor-
mation Systems (GIS).
F. Syafar (*)
Universitas Negeri Makassar, Makassar, Indonesia
e-mail: [email protected]
A. Koronios (*) • J. Gao
University of South Australia, Adelaide, Australia
e-mail: [email protected]; [email protected]
Sokianos et al. [8] emphasise that in order to manage the sophisticated AM process
and to provide its data requirements, particular technology and systems are
required. The system that captures, maintains, and manages all the needed asset
information throughout the entire asset lifecycle is critical in providing
effective AM.
In contrast, mobile technologies and solutions are very popular in consumer
applications, and the exploitation of mobile technologies will keep on expanding. In
heavy industry and maintenance, the uses of mobile solutions have not yet become
very popular. One reason is the lack of competence and knowledge in adopting
mobile solutions successfully for professional use. Many companies have poor
experience of adopting mobile solutions in maintenance due to previously inoper-
ative telecommunication connections, lack of suitable devices, or just because the
organisation had insufficient preparation for the adoption and implementation
process. Another reason is that the benefits of mobile solutions are unseen or
unknown, for example, in the maintenance domain [9]. Mobile technologies
248 F. Syafar et al.
nowadays are mature enough to face the challenge and requirements of professional
use in the engineering industry.
The use and implementation of mobile services has been studied globally and
extensively from a context-driven organisational problem-solving view [10]. When
considering the use of mobile solutions in industry, and especially in maintenance,
the available studies and research focuses mainly on e-maintenance [11–13]. The
term e-maintenance still refers to quite a large concept where mobile solutions can
be just one part. Some e-maintenance specific case studies focus on mobile device
architectures where the mobile device can, for example, help the maintenance
engineer perform maintenance tasks [11]. Mobile solutions can bring maintenance
management closer to daily practice in the field, and lead to more efficient main-
tenance operations.
Some research has been conducted on the role of mobile technology in the
workplace, but only few applied to asset maintenance works. Moreover, several
mobile maintenance systems have been invested in by EAM organisations to
enhancing their AM and maintenance systems. But these technologies/systems do
not adequately support the maintenance collaboration requirements associated with
different maintenance stakeholders.
A multiple case-study approach was adopted for the case-study methodology in this
research. It is aim to identify the role of mobile technologies in asset maintenance
activities with specifics focuses on
a. The current use of mobile technologies in asset maintenance
b. Collaborative asset maintenance requirements
c. Issues and problems associated with the current mobile technologies
The reasons for choosing a multiple case study approach over a single case
approach was its capacity to handle the complexity of the phenomenon under study
[14], and the fact that it augmented external validity, helping guard against observer
bias [15]. It is recommended to be of assistance in capturing the complexity of
social settings, and facilitating the comparison of activities across a variety of
settings and situations [16]. The multiple case-study approach uses replication
logic to achieve methodological rigour [14, 17], and triangulate evidence, data
sources and research methods [18].
Eight Australian and Indonesian engineering asset organisations were selected
for the case study in this research. All were chosen from large sized organisations
taking into consideration the complexity of maintenance process, such as having
more functions, covering more operation and maintenance perspectives, involving
more and variety of maintenance stakeholders, and more importantly, having strong
motivation to improve their maintenance productivity. They truly reflect the engi-
neering asset organisations that need, or have been implementing, collaborative
Mobile Technologies in Asset Maintenance 249
5 Research Findings
6 Conclusion
information for informed decision-making on the fly. On site they can monitor
workload, fill in expenses and work done, and continuously report job progress so
an engineering organisation’s entire workforce can be optimised on the right job at
the right time, and meet its service level agreement. However, in order to fully take
advantages of mobile technologies, it is an ongoing journey for asset management
organisations to undertake.
References
1. Pintelon L, Muchiri PN (2009) Safety and maintenance. In: Handbook of maintenance man-
agement and engineering, pp 613–648. isbn 1848824718
2. Dekker R (1996) Applications of maintenance optimization models: a review and analysis.
Reliab Eng Syst Saf 51(3):229–240
3. Gouws L, Trevelyan J (2006) Research on influences on maintenance management effective-
ness. In: Paper presented at the 2006 world conference on engineering asset management.
Surfer’s Paradise, Queensland, 11–14 July 2006
4. Soderholm P, Holmgren M, Klefsjo B (2007) A process view of maintenance and its stake-
holders. J Qual Maint Eng 13(1):19–32
5. CEN, European Committee for Standardization (2001) Guide on preparation of maintenance
contracts. European Standard, Brussels: CEN ENV-13269:2001
6. Koochaki J (2009) Collaborative learning in condition based maintenance. Lect Notes Eng
Comput Sci 2176(1):738–742
7. Jardine AKS, Lin D, Banjevic D (2006) A review on machinery diagnostics and prognostics
implementing condition-based maintenance. Mech Syst Signal Process 20(7):1483–1510
8. Sokianos N, Drke H, Toutatoui C (1998) Lexikon productions management. Landsberg,
Germany
9. Backman J, Helaakoski H (2011) Mobile technology to support maintenance efficiency-mobile
maintenance in heavy industry. In: Proceedings of the 9th IEEE international conference on
industrial informatics, pp 328–333
10. Burley L, Scheepers H (2002) Emerging trends in mobile technology development: from
healthcare professional to system developer. Int J Healthc Technol Manag 5(3):179–193
11. Campos J et al (2009) Development in the application of ICT in condition monitoring and
maintenance. Comput Ind 60(1):1–20
12. Koc M, Ni J, Lee J, Bandyopadhyay P (2003) Introduction of E-manufacturing. In: Refereed
proceedings of the 31st North American manufacturing research conference (NAMRC),
Hamilton, Canada
13. Muller A, Marquez AC, Iung B (2008) On the concept of E-maintenance: review and
current research. Reliab Eng Syst Saf 93(8):1165–1187
14. Yin RK (2003) Case study research: design and methods, 3rd edn. Sage, Thousand Oaks, CA
15. Leonard-Barton D (1988) A dual methodology for case studies: synergistic use of longitudinal
single site with replicated multiple sites. Organ Sci 1(3):248–266
16. Adams ME, Day GS, Dougherty D (1998) Enhancing new product development performance:
an organisational learning perspective. J Prod Innov Manag 15(5):403–422
17. Donnellan E (1995) Changing perspectives on research methodology in marketing. Ir Mark Rev
8:81–90
18. Eisenhardt KM (1989) Building theories from case studies. Acad Manag Rev 14(4):532–550
19. Neuman WL (2006) Social research methods: qualitative and quantitative approaches.
Pearson Education, Boston
Mobile Technologies in Asset Maintenance 253
20. Snitkin S. (2003) Collaborative asset lifecycle management vision and strategies.
Research report, ARC Advisory Group, Boston, USA
21. Syafar F, Gao J (2013) Mobile collaboration technology in engineering asset maintenance:
a delphy study. In: Paper presented at the 2013 I.E. 17th international conference on CSCWD,
Whistler, BC, Canada, 27–29 June 2013
22. Syafar F, Gao J, Du T (2013) Applying the international delphi technique in a study of
mobile collaborative maintenance requirements. In: Refereed proceedings of the 17th Pacific
Asia conference on information systems (PACIS), Jeju Island, Korea
Essential Elements in Providing Engineering
Asset Management Related Training
and Education Courses
Peter W. Tse
1 Introduction
Korea (2012), Hong Kong (2013), South Africa (2014), Finland (2015) etc. The
objective of WCEAM is to bring together academics, practitioners and scientists
from all around the world to promote research, development and application in the
field of EAM and to strengthen the link between academic researchers and indus-
trial practitioners in the EAM field [4].
For the core courses, as aforementioned, a student must select 18 CUs or 6 courses
from the following courses that can match with most of the required 39 subjects.
The italic words are the course names. The rest of the sentence contains the matched
subjects.
1. Asset and Maintenance Management: shutdowns and outage strategy and
management, asset creation and acquisition, maintenance delivery
2. Quality and Reliability Engineering: contingency planning and resilience anal-
ysis, audit and assurance
3. Risk and Failure Analysis: fault and incident response, reliability engineering,
risk assessment and management
4. Managerial Decision-making Systems: capital investment decision-making,
operations and maintenance decision-making, management review
5. Policy Making and Management: AM policy, AM strategy and objectives,
demand analysis, strategic planning and AM planning, management of change,
management review
6. Financial Engineering for Engineering Managers: asset creation and acquisi-
tion, resourcing strategy, resource management, asset costing and valuation,
stakeholder engagement
7. Life Cycle Management: lifecycle value realization, asset operations, asset
decommissioning and disposal, sustainable development, management review
8. Condition Monitoring, Diagnosis and Prognosis: asset operations, asset per-
formance and health monitoring, AM system monitoring
9. Statistical and Information Analysis: asset information strategy, asset informa-
tion standards, asset information systems, data and information management
10. Project Management: resourcing strategy, configuration management, organi-
zational structure, organizational culture, competence management
Besides the requirement of taking six core courses, the student must also select
12 CUs or four courses from the following elective courses listed in the EAM
program.
1. Global Business Enterprise and Planning: organizational structure, organiza-
tional culture, competence management
2. Global Human Resource Management: resourcing strategy, resource
management
3. Emerging Issues in Multinational Strategic Management: organizational struc-
ture, organizational culture, competence management
4. Management Science: AM leadership
5. Quality Improvement, Systems and Methodologies: systems engineering
6. Marketing Strategy for Engineers,
7. Supply Chain Management: procurement and supply chain management,
Essential Elements in Providing Engineering Asset Management Related. . . 261
8. Business Law and Contract for Engineers: technical standards and legislation
9. Management of Technological Innovation,
After taking the core courses and the electives listed above, the students should
possesses sufficient knowledge and skills to handle and practice the required
39 subjects. The graduated student can classify himself as a qualified specialist
in EAM.
6 Conclusion
The recent release of the ISO standards, the ISO 55000-2 and the BSi PAS 55 on
Asset Management do attract a number of organizations to offer certificate training
programs in physical asset management and a handful of universities to establish
EAM based curriculum. However, as of today, there is no formal regulation to
monitor and evaluate the quality of these training programs and curriculum. Based
on the standards’ requirements, this paper presents a list of essential subjects and
courses in the development of EAM-based curriculum and the surveillance of
quality in offering related training programs to public. The ultimate goal is to
establish suitable proper guidelines and code-of-practice so that any institution or
university can use the recommended curriculum to benchmarking her intended
curriculum to ensure it is comply with the requirements of the standards. Hence,
the graduated students can be recognized by world-wide organizations and
accredited by well-known professional organizations that serve EAM.
262 P.W. Tse
Acknowledgments The work described in this paper was fully supported by two grants from the
Research Grants Council of the Hong Kong Special Administrative Region, China (Project
No. CityU 11201315).
References
1 Introduction
Wind energy plays a very important role as an alternative energy supply nowadays,
due to its properties of renewability and generality on earth. In the year of 2011, the
worldwide installed wind power capacity was reported to reach 237 Gigawatt (GW). It
is expected that the total installed wind power capacity will be 1000 GW by 2030, as
reported by the World Energy Association. For most of them, the utilization of wind
energy is achieved in the form of transformation to the electricity using wind turbines.
In order to make full use of local wind energy resources, multiple wind turbines are
placed in cluster which is called the wind farm or wind park. However, the
non-isolated wind turbines bring about the wake interactions, namely, wake interfer-
ence or wake effect, which greatly reduce the total power production of a wind farm.
By optimizing the wind farm layout, the power losses can be regained to a large extent.
Among all the previous wind farm layout optimization studies reported, most of
them have considered the scenario with constant wind turbine hub height [1, 2]. How-
ever, for a wind farm mostly using the identical type of wind turbine, the hub height
of turbines can be selectable. Meanwhile, it is reported that the use of different wind
turbine hub heights has the potential to reduce the power losses caused by the wake
interaction and hence contribute to the wind farm layout optimization. For the
existing literature that report the optimization with different hub heights [3], none
of them have claimed to apply the unrestricted coordinate method (use wind turbine
coordinate to determine the position in wind farm) to study the layout optimization,
which is believed to be superior than the counterpart method. Therefore, this paper
attempts to discuss the wind farm layout optimization using the unrestricted coordi-
nated while considering different wind turbine hub heights. The effect of applying
different wind turbine hub heights on the power production for both single wind
turbine and wind farm is discussed in detail through simple two turbine model and
wind farm model. In the meantime, different optimization methods including the
single layout optimization, hub height optimization and simultaneous optimization
are applied to evaluate their effectiveness under different wind conditions.
The rest of the paper is organized as follows. Section 2 discusses the methods
applied for the wind farm layout optimization studies. It includes the optimization
algorithm applied for the studies, the calculation of the objective function, the
representation of the solution for the objective function using different optimization
methods and finally process of the wind farm layout optimization studies using
different methods. Section 3 discusses the results and Sect. 4 draws the conclusion.
2 Methods
The methods for the three different types of optimizations are introduced in this
section. They are presented in the aspects of optimization solution for the methods,
optimization algorithm, objective function calculation and finally the optimization
process for the methods
Optimizing the Unrestricted Wind Turbine Placements with Different Turbine. . . 265
For all the three different optimization methods as described above, one feature that
they have in common for the solution X is that they all applied the simple real
coding method with different number of variables. Therefore, simple Single Objec-
tive Genetic Algorithm (SOGA) is employed in this paper to study the optimization
of wind farm with different hub heights. GA is a search heuristic that mimics the
process of natural selection. It begins with encoded solutions to the optimization
problem. The main principle of GA is the maintenance of these encoded solutions
which are evolved with the generations to be guided towards the optimum solutions
step by step. A simple GA works as follows [4]: a random initial population (a set of
encoded solutions) is created. The fitness of individual (the single encoded solution)
is evaluated based on the optimization objective function.
1. The raw fitness values are transformed into the range of values that are suitable
for the selection process through the fitness scaling procedure.
2. The individuals with the best fitness values are guaranteed to survive to the next
generation, while other individuals are used to select parents to produce new
population individuals for the next generation.
3. Other new population individuals are generated through the crossover and
mutation operators.
4. The current population is replaced with the new generation.
5. Steps 2–6 are repeated until the stopping criteria is met.
Before calculating the wind farm power, the single wind turbine power PWT and the
wind speed approaching the rotor of every single wind turbine should be deter-
mined first. In this paper, the wind turbine model applied in the reference [5] is
employed, and means of determining the wind speed for the turbines affected by the
multiple wakes can be referred in the reference [1, 2].
For the discrete wind condition applied in this paper, based on the individual
wind turbine power model and the approaching wind velocity for each wind
turbine, i-th wind turbine power Pi can be obtained. The total power output Ptot
with NWT number of wind turbines and finite number of wind directions Nd can be
calculated as:
" #
X
Nd X
N WT
Ptot ¼ pj Pi ð1Þ
j¼1 i¼1 j-th direction
where pj is the probability of occurrence of j-th wind direction. The wind farm
power output is calculated as the accumulation of all wind turbine power output:
266 L. Wang et al.
X
N WT
Ptot ¼ Pi ð2Þ
i¼1
Based on the calculation of wind farm power output and the wind farm cost
models, the Cost of Energy (CoE), which is the objective function for the wind farm
optimization study in this paper, can be represented by:
where cost is the wind farm cost given in the reference [6] and Ptot is the wind farm
power output calculated above.
The optimization solution X indicates the individual of encoded solution, and each
encoded element among them indicates the variable that needs to be optimized
through the algorithm.
a. Layout Optimization
For the simple wind farm layout optimization with identical wind turbine hub
height, the optimization solution X can be represented as follows:
X ¼ x1 xi xN y1 yi yN ð4Þ
|fflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflffl{zfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflffl}
2 N decimal digits
Where x and y store the X and Y coordinates for different wind turbines. N is the
number of wind turbines to be optimized which is predetermined by the user. So
X applies the real variable value codification with altogether 2*N number of
variables.
b. Hub Height Optimization
For the simple wind turbine hub height optimization, it is conducted based on the
result of the optimized wind farm layout through the simple wind farm layout
optimization method. The optimization solution X for the optimization method can
be represented as follows:
X ¼ H1 Hi HN ð5Þ
|fflfflfflfflfflfflfflfflfflfflfflfflfflfflffl{zfflfflfflfflfflfflfflfflfflfflfflfflfflfflffl}
N decimal digits
where H stores the wind turbine hub height value for different wind turbines and
X has N number of variables in total.
Optimizing the Unrestricted Wind Turbine Placements with Different Turbine. . . 267
c. Simultaneous Optimization
For the simultaneous wind farm layout and wind turbine hub height optimization,
the optimization solution X can be represented as follows:
X ¼ x1 xi xN y1 yi yN H 1 H i H N ð6Þ
|fflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflffl{zfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflffl}
3 N decimal digits
where it is obvious that the solution X for the method can be regarded as the
integration of the two solutions for the above two optimization methods.
Figure 1 depicts the general process of the three methods applied in this paper for
the wind farm layout optimization studies. Initially, the wind farm design param-
eters are randomly generated for the different optimization approaches. For the
simple wind farm layout optimization method, the random layout parameter (wind
turbine position) of the initial population is generated with fixed wind turbine hub
height. For the simple hub height optimization method, the random wind turbine
hub height value of the initial population is generated while the wind turbine
positions are fixed which are imported from the layout optimization results. Unlike
these two optimization methods, both wind turbine positions and wind turbine hub
height are variable and optimized during the process for the simultaneous optimi-
zation. Hence both parameters are initialized randomly as the initial population for
this method. After the initialization, all the related wind turbine parameters used for
the calculation of wind farm power output are ready using different optimization
approaches. Then individuals of the initial population are evaluated according to
Fig. 1 Process of the wind farm layout optimization study using different optimization methods
268 L. Wang et al.
Before applying different wind turbine hub heights while designing the wind farm,
the simple two wind turbine model is employed to study the influence of the factors
including surface roughness, wind turbine spacing and wind direction on the wake-
affected wind turbine power production with different hub heights. After this, the
wind farm layout optimization study using different optimization methods are
carried out and the effectiveness of the methods are compared.
For the two-turbine model, we fix the first wind turbine hub height and study the
variation of the second wind turbine power affected by the wake effect (normalized
to the free stream wind turbine power) with different hub heights. First, the
influence of the surface roughness length (Z0) on the results is studied with wind
direction aligned with the wind turbines and fixed spacing of 15D (D is rotor
diameter). Figure 2a depicts the study model while four different surface roughness
values are employed from the smooth surface of open sea (Z0 ¼ 0.0002) to the
tough surface of high crop obstacle ground (Z0 ¼ 0.3). As can be seen from Fig. 2b,
the tougher surface is, the more power wake-affected wind turbine produces in
general. When the surface roughness length value is big (0.3), there is no need to
Fig. 2 Variation of
normalized wake-affected
wind turbine power
production with different
wind turbine hub height for
four characteristic surface
roughness lengths
Optimizing the Unrestricted Wind Turbine Placements with Different Turbine. . . 269
Fig. 3 Variation of
normalized wake-affected
wind turbine power
production with different
wind turbine hub height for
different spacing between
wind turbines
apply different wind turbine hub heights for the back turbine in this case since it is
in full wake of the front turbine anyway. When the surface roughness length value
is small (Z0 is less than 0.001), the variation of Z0 won’t have much impact on the
power production, and increase of power production by applying different hub
height is more prominent. The bigger difference back turbine hub height has from
front turbine hub height, the larger power production increase it yields compared
with constant hub height.
Figure 3 indicates the results of back turbine power production with different
hub heights for different spacing (S) between the two turbines from 5D to 20D (D is
the rotor diameter). Obviously, the more distance between turbines, the less wake
power losses it has. Also it is found that the percentage of power increase due to the
enlarged distance becomes less as it continues. Still when applying different wind
turbine hub heights, there is basically the same power increase regardless of the
spacing it applied. For the wake-affected back turbine, its power productions with
different hub heights are symmetrical with that of the front turbine hub height value
(60 m), at which point the back turbine has the most wake power losses.
Next we study the effect of wind scenarios on the power production of wake-
affected wind turbine by applying different hub heights. When varying the wind
speed with fixed wind direction still aligned with the two wind turbines, it is found
that the wake effects are the same for the back turbine with constant normalized
power production, and the results are not quantitatively shown here. Figure 4
reports the variation of normalized power production of wind turbine with different
hub heights when the wind blowing angle varies from 0 to 10 . The model is
illustrated in Fig. 4a and the results are shown in Fig. 4b. As can be seen when the
wind direction is very much aligned with the two wind turbines (up to 4 ), the
power production of the back turbine is the same regardless of the hub heights
applied. The power production increases as the wind direction continues to
increase, and the power increase due to the use of different wind turbine hub
270 L. Wang et al.
Fig. 4 Variation of
normalized wake-affected
wind turbine power
production with different
wind turbine hub height
under different wind
blowing angles
heights is also increasingly pronounced along with the wind direction (up to 9 ).
When the wind direction exceeds 10 , the back turbine is not located in the wake of
the front turbine anymore, which reaches its maximum power production with free
stream wind.
Three characteristic wind farm conditions, which have been popularly employed
for the wind farm layout optimization study, are applied in this paper to compare the
effectiveness of different optimization methods for the wind farm with different
hub heights.
a. Constant Wind Speed and Wind Direction
Figure 5 reports the wind farm layout optimization results under constant wind
speed of 12 m/s and constant wind direction that is aligned with the wind farm
length direction (defined as 0 ). Figure 5a shows the optimized fitness of the
objective function, which is the normalized value for cost of energy production
(CoE) using the three methods. The standard deviation of the fitness results with
repeated calculation are incorporated in the figure as well. It can be seen that the
simultaneous optimization method achieves the best optimization results (the
smallest fitness value) through repeated calculation, while it has large deviation
and hence relies on the repetition to obtain the best optimization result. When using
simple layout optimization method, it also has large deviation. Based on its best
optimization result through repeated calculation, the hub height optimization
attempts to further optimize the wind turbine hub height but has little improvement.
Optimizing the Unrestricted Wind Turbine Placements with Different Turbine. . . 271
Fig. 5 Results of constant wind speed and constant wind direction aligned with wind farm length:
(a) normalized fitness of the objective function using different optimization methods (standard
deviation depicted) and (b) optimal wind farm layout with different turbine hub heights (turbines
are represented by the cuboid with heights indicated)
It is also found that the hub height optimization method has no deviation through
repeated calculation which means that the results are unchanged and the perfor-
mance is stable. Figure 5b shows the optimal wind farm layout using different wind
turbine hub height after the optimization comparison. It should be noted that the
selectable range of wind turbine hub height is from 60 to 70 m in this paper for the
wind farm optimization study with all three different wind conditions. The distri-
bution of the wind turbine is obvious for the simple constant wind condition. Nearly
half number of turbines is located near to the one side of the wind farm boundary
perpendicular to the wind direction in windward direction, and they all have the
lower bound of the hub height value which is 60 m. Most of the rest of wind turbines
are distributed near to the other normal side of the wind farm boundary in leeward
direction. As can be seen, most of them have higher hub height than that of the wind
ward turbines to escape from the wake of upstream wind turbines.
b. Constant Wind Speed and Variable Wind Directions
Figure 6 reports the optimization results for the constant wind speed of 12 m/s and
variable wind directions. The wind is coming from 36 wind directions evenly
distributed with 10 interval, that is 0 , 10 , . . ., and 350 . It can be from Fig. 6a
that for this wind condition the simultaneous optimization achieves way much
better results than the counterpart method. Meanwhile, the deviation of repeated
calculation for the method is relatively small compared with the results of above
constant wind condition, which implies the performance of the method is more
stable. In comparison, the simple layout optimization method has large deviation
and obtains worst optimization results, while the results of hub height optimization
are also inferior to the simultaneous optimization method and the deviation of the
hub height optimization with repeated calculation is still zero. By observing the
optimal wind farm layout shown in Fig. 6b, it is obvious that most of wind turbines
are distributed along the four sides of wind farm boundaries. Seven of them have
272 L. Wang et al.
Fig. 6 Results of constant wind speed and variable wind directions: (a) normalized fitness of the
objective function using different optimization methods and (b) optimal wind farm layout with
different wind turbine hub heights
Fig. 7 Histogram of wind condition for variable wind speeds and wind directions
the upper bound of hub height value (70 m), and the rest of wind turbine has hub
height of 60 m.
c. Variable Wind Speeds and Variable Wind Directions
Before discussing the results under the variable wind speeds and variable wind
directions, the wind condition is introduced here as shown in Fig. 7. The wind
condition also consists of 36 wind directions with 10 interval. Every component of
the wind directions comprises three different wind speeds and each wind speed in
each wind direction is assigned with a value to represent the probability of occur-
rence. The same wind condition is also applied in the reference [7].
Optimizing the Unrestricted Wind Turbine Placements with Different Turbine. . . 273
Fig. 8 Results of variable wind speeds and variable wind directions: (a) normalized fitness of the
objective function using different optimization methods and (b) optimal wind farm layout with
different wind turbine hub heights
Results of the wind farm layout optimization study with different hub heights for
this wind condition are shown in Fig. 8. Like the above condition of constant wind
speed and variable wind directions, the simultaneous optimization achieves the best
fitness results with the small deviation of repeated calculations. The results using
the layout optimization method are the worst with large deviation, based on which
there is a small improvement for the results of hub height optimization without
deviation of repeated calculation for it. According to the optimal wind farm layout
for this wind condition shown in Fig. 8b, most of the wind turbines are spread along
the four sides of wind farm boundaries to enlarge the distance between wind
turbines with blowing wind from all 360 . Five of them employ the largest wind
turbine hub height with 70 m, and for the rest of wind turbines the hub heights of
them are near to 60 m.
4 Conclusions
The wind farm layout optimization, which applies the unrestricted coordinate
method to determine the wind turbine position, is carried out in this paper for the
first time by considering different wind turbine hub heights with same rotor
diameter (same wind turbine type). Before introducing the wind farm design
results, the simple two wind turbine model is studied in order to discuss the
influence of related factors on the wake-affected turbine power output when
applying different hub heights. It is found out by altering the back turbine hub
height (either increasing or decreasing based on the front turbine hub height value),
the power output of the back turbine can be regained. The bigger difference of hub
height between the two turbines, the more power output increase it has. However,
depending on the distance between the two turbines, the variation of the hub heights
274 L. Wang et al.
becomes ineffective when the surface roughness length exceeds certain value as the
back turbine is always located in the full wake of front turbine. Based on the wake
model applied in this paper, with all other fixed factors except for the blowing wind
speed the normalized power output for the back turbine is unchanged regardless of
the hub height employed. Nonetheless, it is closely related to the blowing wind
direction. Under certain value of wind angle, the wake-affected turbine power is
fixed as it is always in the full wake of upstream turbine regardless of the hub height
employed. As wind direction increases, the effect of altering the turbine hub height
on the increase of power output becomes pronounced since it enables the back
turbine rotor to jump out of the wake zone or at least decrease the wake-affected
area. When the wind direction exceeds certain value, the back turbine is not affected
by the wake of front turbine even with constant hub height, and hence the use of
different hub heights becomes needless.
Three characteristic wind conditions, which have been widely employed in the
wind farm layout optimization studies, are applied in this paper to test the effec-
tiveness of different methods for the different hub height optimization studies of
wind farm. For the simplest wind condition of constant wind speed and constant
wind direction, the layout optimization is effective enough to obtain the good
results by repeated calculation. Based on the results, the improvement of applying
hub height optimization is not evident. In comparison, the simultaneous optimiza-
tion has large deviation with repeated calculation and the improvement of the best
optimization results obtained by the simultaneous method is relatively small. For
the other two wind conditions, the effectiveness of applying simultaneous optimi-
zation method is obvious compared with the counterpart method and the perfor-
mance is more stable with relatively small deviation of repeated calculation. This is
because for these two wind conditions incorporating all wind directions of 360 the
single layout optimization is inefficient with constant wind turbine hub height.
Based on the inferior fixed wind turbine positions, the hub height optimization has
little contribution to the increase of the final wind farm results. In contrast, the
simultaneous optimization method facilitates the design to both optimize the wind
turbine position and hub height at a time and hence much better results can be
achieved.
References
1. Wang L, Tan AC, Gu Y, Yuan J (2015) A new constraint handling method for wind farm layout
optimization with lands owned by different owners. Renew Energy 83:151–161
2. Wang L, Tan AC, Gu Y (2015) Comparative study on optimizing the wind farm layout using
different design methods and cost models. J Wind Eng Ind Aerodyn 146:1–10
Optimizing the Unrestricted Wind Turbine Placements with Different Turbine. . . 275
3. Chen Y, Li H, Jin K, Song Q (2013) Wind farm layout optimization using genetic algorithm
with different hub height wind turbines. Energy Convers Manag 70(0):56–65
4. Mitchell M (1998) An introduction to genetic algorithms. MIT Press, Cambridge, MA
5. Grady S, Hussaini M, Abdullah MM (2005) Placement of wind turbines using genetic algo-
rithms. Renew Energy 30(2):259–270
6. Mosetti G, Poloni C, Diviacco B (1994) Optimization of wind turbine positioning in large
windfarms by means of a genetic algorithm. J Wind Eng Ind Aerodyn 51(1):105–116
7. Chen L, MacDonald E (2014) A system-level cost-of-energy wind farm layout optimization
with landowner modeling. Energy Convers Manag 77:484–494
Predicting Maintenance Requirements
for School Assets in Queensland
1 Introduction
School facilities are crucial assets that are designed to provide safe and comfortable
places for learning. It is generally acknowledged in the relevant literature that the
physical condition can significantly affect teaching, learning, and the health and
safety of students and teachers [1–4], making maintenance a critical part of the
mission of schools. Yet, setting the appropriate level and allocation of maintenance
funds remains a challenge. In Queensland, this is evident from the significant and
persistent maintenance backlog, which was estimated at $232 million since the
efforts to clear the 2011–2012 school maintenance backlog of $298 million [5].
Funding allocation models for buildings are typically based on the value of the
building asset. For instance, in Queensland, the Department of Housing and Public
Works [6] recommends a minimum allocation of 1% of the building Asset
2 Data Description
1
We actually use the moving average of the past 4 years of maintenance expenditures due to the
infrequent inspection intervals and the challenges with aligning the precise dates of maintenance
with the inspection times.
280 R. Wang et al.
The condition of a particular element group was quantified by the indicative cost, or
the estimated cost to repair, and a degradation rate is thus a change indicative cost
per unit time. The indicative costs for a selected EGC for a particular building (BID)
can be computed by summing up the indicative costs for all elements (E) that
belong to the EGC
X
I BID, EGC ðtÞ ¼ I BID, E ðtÞ ð1Þ
E2EGC
where IBID , E(t) is the indicative cost of element E at time t and the expression
E 2 EGC (in an abuse of notation) indicates that an element belongs to a particular
EGC. It is implicitly assumed that the indicative cost on un-inspected elements is
zero. That is, a report only appears when the element is in need of repair.
The BID level indicative costs may be used to assess the impact of the building-
level candidate factors: building age, building GFA, and heritage listing. However,
most of the analysis of the indicative cost patters was conducted at the school level
(CIS) by summing the indicative costs for each BID for the EGC of interest
X
I CIS, EGC ðtÞ ¼ I BID, EGC ðtÞ ð2Þ
BID2CIS
The reason for conducting the analysis at the school level is that the majority of
the candidate influential factors are specified at the CIS level. By examining the
indicative cost of the school, the effect of sparse of building-level condition
assessments is less pronounced, provided that the element group is inspected on
some buildings for a school in a particular year.
Figure 2 shows the time history of EFIN indicative cost per m2 (i.e. GFA), which
we use as an indication of school size.2 The results clearly indicate that there are
significant regional differences in the various regions of Queensland: Darling
Downs South West, North Queensland, and Central Queensland have significantly
higher EFIN costs per m2 than the remaining regions. This is important to note for
any prediction, since increasing the GFA in one of these regions would have a
larger effect on total EFIN maintenance costs than elsewhere.
Figure 3 shows per m2 indicative costs for IFIN. From these plots it is observed
that the total maintenance work to be done for IFIN has stabilized and is now
(slightly) decreasing for most regions. It is also noted that Darling Downs South
West and Central Queensland are the regions that have the worst IFIN indicative
costs per m2.
2
One could also use the number of students to normalize the costs. However, motivated by our
later statistical analysis, we only show GFA.
Predicting Maintenance Requirements for School Assets in Queensland 281
Finally, the regional trends in BLDG condition were examined. Figure 4 shows
per m2 indicative costs that Central Queensland, and North Queensland, and Far
North as have the highest indicative costs per m2. It is also noted that most of the
regional indicative costs have decreased significantly, but have started to increase
again.
Clearly, there are regional differences even after normalisation by GFA. This
variation motivates the inclusion of the region as a potential influential factor.
Examining the map of the regions in Queensland in Fig. 1, we can see a common
factor of the high-cost regions: they all have significant inland regions, motivating
the inclusion of the distance from the cost as a potential influential factor as well.
4 Model Development
In this section a model for the degradation of the EGCs is developed. The condition
of a particular EGC was quantified by the indicative cost, so degradation is simply
the change in total indicative cost of an EGC per unit time.
Predicting Maintenance Requirements for School Assets in Queensland 283
where t‘ is the time of the last inspection. We will hereafter refer to ΔICIS, EGC (t) as
the school indicative cost rate and analogously define ΔICIS , BID, the building
indicative cost rate as an obvious modification to Eq. (1).
Two statistical tools were employed to assess the significance of the potential
factors: multivariate regression will be used to assess the strength of the effect of
each factor while and analysis of variance (ANOVA) will be used to assess the
strength of the evidence that the factor is important. The difference between these
“strengths” is subtle but important: a high strength of effect indicates that the factor
has a large influence on the indicative cost rate while a high strength of evidence
means that this effect is real and not just a statistical fluke. Accordingly, our
working model for examining the significance of each factor at the school level
was formed as follows:
ΔI CIS, EGC ðtÞ ¼ α1 GFACIS þ α2 UCIS ðtÞ þ α3 EXPCIS ðtÞ þ α4 ECIS ðtÞ þ τr
þ α5 CLCIS þ b þ εðtÞ ð4Þ
ΔI BID, EGC ðtÞ ¼ α1 GFABID þ τr þ α2 CLBID þ α3 AgeBID ðtÞ þ τH BID þ b þ εðtÞ ð5Þ
where ε(t) is the prediction error and the remaining variables are defined in Sect. 2.
To assess the strength of evidence, we will examine the p-values resulting from the
ANOVA. We will adopt the conventional (albeit somewhat arbitrary) p < 0.05 as
statistically significant factors, i.e. we wish to have less than a 5% chance that the
factor is falsely stated as significant.
The school level model of Eq. (4) will be our primary tool for exploring the
strength of the candidate factors. However, the age of the school isn’t meaningful
since each building can be of significantly different age. Heritage listing is also a
building level property. Thus, we also employ a building level model to assess the
influence of heritage listing and age.
The procedure for assessing a factor’s significance was as follows:
1. Begin by considering one factor at a time. Eliminate all factors that are not
statistically significant on their own.
2. For all the statistically significant factors from step 1, fit a model with the first
level factors (i.e. “main effects” in ANOVA terminology).
3. If all p < 0.05 stop. Report factors. If not, exclude the factor with the highest p-
value and repeat steps 2–3.
In the following sections, the results of statistical analysis for each EGC at the
school and building level will be detailed. For the following results, the regions are
encoded as in Table 1.
284 R. Wang et al.
For EFIN, the identified influential variables were displayed in Table 2. The results
demonstrate GFACIS positively influences the changes in the indicative cost while
EXPCIS(t) negatively influences it. This is of course intuitive: a larger floor area is
more difficult to maintain and maintenance spending should improve the condition,
decreasing the indicative cost.
The region, r, was also identified as significant influential factor with approxi-
mately zero p-value. Regional coefficients τr in red of Darling Downs South West,
North Queensland and Central Queensland are significantly higher than other
regions, confirming our ad hoc analysis in Sect. 3. Figure 5 shows a 1D projection
of the prediction along the GFA factor to give a sense for the quality of the fit.
Table 3 shows the results for IFIN. We once again find that GFACIS, EXPCIS(t),
are significant, however r is just outside our significance cut-off, and we briefly
depart from convention to show its coefficients. We see that the high-cost regions
are once again represented (1, 2, and 6), but we also note that the regional
coefficient of the South East (7) region is high. The South East region has the
third highest GFA, which is directly in line its third-highest indicative cost (Fig. 3).
However, the fast rise at the beginning of the series leads to the large estimated τ7
(Table 3). This may be due in part to the “start-up” effect noted in Sect. 3, but it was
decided that the data series was too short to leave out these early points for this
exploratory analysis.
The results for BLDG can be seen in Table 4. Following our procedure, the
influential variables would have been GFACIS and, to a lesser extent, ECIS
( p-value ¼ 0.015). However, ECIS was excluded from the influential variable list
for two reasons:
1. ECIS is highly correlated with GFACIS (correlation coefficient of 0.88), meaning
that they contain very similar information.
2. The coefficient of ECIS was negative, which makes no physical sense since more
students would decrease the indicative cost.
Predicting Maintenance Requirements for School Assets in Queensland 285
Point 1 is supported when the second level factors of ANOVA are considered
(i.e. “interaction” terms). We find that the interaction term is significant, but the
main effect for ECIS is not. This means that ECIS influences the result only
through GFA. We thus neglect it as a significant factor on its own, but a more
sophisticated prediction model may need to consider this nonlinear effect for
optimal accuracy.
We also note that maintenance expenditure was not significant for BLDG. This
is likely due to the fact that in the data available, there is no straightforward way to
extract only maintenance expenditures that are related to BLDG alone. Thus, the
286 R. Wang et al.
entire maintenance expenditures must be used, much of which did not pertain to the
external finishes.
Based on the analysis for the three element groups EFIN, IFIN and BLDG on
school level, GFA and region (r) have been identified as strongest influencing the
degradation on school level. Maintenance expenditure does have an effect, but only
when expenditure data related directly to the EGC can be extracted, as was done for
EFIN and IFIN.
A building-level model was proposed in Eq. (5) and the identified influential
variables for EFIN and IFIN are displayed in Tables 5 and 6, respectively. For
EFIN, the identified influential variables at the building-level are AGEBID and
region, while for IFIN the influential variable is region alone. Interestingly, GFABID
is not a significant factor. This may be due to the lack of maintenance information,
which adds uncertainty to the analysis and makes it more difficult to find smaller
correlations. It might also be that at the school level, GFA is a proxy for the number
Predicting Maintenance Requirements for School Assets in Queensland 287
of buildings to be maintained, while this is not so at the building level (there’s only
one).
Examining the EFIN regional coefficients, we see that the three largest regional
coefficients are indeed the same as at the school level, further supporting the idea
that these regions are indeed significantly different in terms of EFIN degradation.
On the other hand, for IFIN we see that the regional effects with the high τr values
correspond to the high un-normalized indicative cost regions. This is likely due to
the fact that the lack of maintenance expenditure information has left region as the
only significant factor at the building level, forcing it to explain all of the variation
in the indicative cost rates alone.
The results for BLDG at the building level can be seen in 7. We clearly see that
all considered variables are significant, with a small effect from the heritage listing.
Interpreting the regional effects with regard to Fig. 5 also suggests that Central and
Far North Queensland are confirmed as being more severe regions for BLDG
degradation. However, the high degradation rates in North Queensland appear to
be explained in part by other factors included in the model.
288 R. Wang et al.
Based on the statistical analysis on both school and building levels, influential
factors can be concluded as follows:
• Region, GFA, Age are confirmed to be influential factors correlated to indicative
cost rates
• Maintenance expenditure effect is a significant factor when we can extract the
maintenance that pertains to a particular EGC.
• Heritage listing has a statistically significant effect on BLDG indicative cost, but
the effect is small relative to GFA and AGE.
The models established here confirm the feasibility of constructing a model to
predict the indicative costs over time. Furthermore, for the models that include
expenditure, we can have a coarse estimate of the expenditure required to stabilize
the indicative cost by setting ΔICIS, EGC(t) ¼ 0 and solving for EXPCIS(t) in Eq. (4).
The significant influential factors on school maintenance cost have been identi-
fied through statistical analysis of Multivariate Regression and ANOVA. From the
available data region, GFA, and Age are verified as influential factors in the
degradation school facilities. Maintenance expenditure is clearly significant as
well, but only when more targeted maintenance expenditure information can be
obtained using the descriptions or maintenance programs.
Based on the model presented in this paper, the effects of future school assets can
be predicted. For instance, it can be seen that expanding school assets in the inland
regions (e.g. Darling Downs South West) will have a larger impact on external
finish maintenance costs than expansion elsewhere. Additionally, ageing schools
are likely to lead to larger maintenance costs; a factor which will play an important
role in budget planning and asset renewal decisions.
Clearly, the proposed (linear) models can provide a prediction of the mainte-
nance cost for these new assets. However these simple statistical models here were
Predicting Maintenance Requirements for School Assets in Queensland 289
indented as tools for exploring the statistical significance of various factors and not
focused on the accuracy of the prediction. We have also ignored the interaction
effects of various factors, which a more sophisticated regression could use to
enhance the prediction accuracy. For a future funding allocation model, it is
recommended for inclusion in a funding allocation model.
Acknowledgments This project was funded by the Queensland Department of Education and
Training (DET). The authors wish to thank Ariane Panochini, Greg Duck, Malvin White, and
Nadeia Romanowski for the in-depth discussions and insight which greatly aided the research
presented in this paper.
References
1. Lawrence BK (2003) Save a penny, lose a school: the real cost of deferred maintenance. Policy
Brief Series on Rural Education
2. Lyons JB (2001) Do school facilities really impact a child’s education. Council of Educational
Facility Planners International, Scottsdale, Arizona (ERIC reproduction service no. ED 458791)
3. Mahli M, Che-Ani A, Tawil MA-RN, Yahaya H (2012) School age and building defects:
analysis using condition survey protocol (CSP) 1 matrix. World Acad Sci Eng Technol 6
(7):1830–1832
4. Schneider M (2002) Do school facilities affect academic outcomes? Academic Achievement 25
5. Queensland Audit Office (2015) Maintenance of public schools (Report 11: 2014–15), edited by
Queensland Audit Office
6. Queensland Department of Housing and Public Works (2012) Policy for the maintenance of
Queensland government buildings, maintenance management framework
7. Bello MA, Loftness V (2010) Addressing inadequate investment in school facility maintenance.
School of Architecture, Paper 50
8. Department of Education and Trainin, Queensland Government (2016) Regional map of Queens-
land. http://education.qld.gov.au/hr/recruitment/teaching/locations.html. Accessed 8 Feb 2016
Research on Armored Equipment RMS
Indexes Optimization Method Based on System
Effectiveness and Life Cycle Cost
1 Introduction
indexes tradeoff analysis based on system effectiveness. Huixin Xiao etc. provided
a gist to integrate tradeoff and optimization of unrepairable equipment by putting
forward a method of RMS synthetical effectiveness model and indexes design
analysis [2]. Directing at unrepairable and repairable faults of equipment during
executing missions, Yong Yu etc. put forward a system effectiveness evaluation
model based on RMS characters and got an expression of system effectiveness
[3]. Jinzhuo Li etc. studied RMS indexes design problem of military aircraft with
simulation. They analyzed an instance of the military aircraft RMS indexes design
with Monte Carlo simulation method and got a RMS indexes solution set which
could afford the support effectiveness requirements [4]. Dianfa Ping etc. built
an index evaluation system of airborne EW system with combination of ADC
model. They chose appropriate partition granularity system based on the objective
structure of airborne EW system, to analyze the system reliability, availability,
capabilities and building an evaluation model [5]. Wenhua Peng etc. established a
mapping relationship between RMS and system effectiveness with WSEIAC
model, and got an intuitional curve of RMS-effectiveness by computer simulation,
which provided a gist for RMS indexes tradeoff in naval gun weapon system
[6]. Kun Han etc. established a system effectiveness model of armored vehicles,
which consists of operational readiness, battle dependability and capability.
Aiming at system effectiveness, they proposed two RMST tradeoff analysis
method, which provided an efficient means to reasonably determine the armored
vehicles RMST quantitative requirements and improve system effectiveness
[7]. Haidong Du etc. proposed synthetic expression of system effectiveness and
related constraints based on equipment RMS indexes, and they also built a tradeoff
dynamic simulation model of system effectiveness. Via simulation, they screened
and optimized index schemes so that they could provide a reference for RMS
design and verification of armored equipment [8]. Considering the relationship
between equipment RMS indexes and life cycle cost, Geng etc. presented the
tradeoff for RMS indexes based on the availability calculation models and LCC
models, on the condition that the availability value is not below the given value,
the decision-making is optimal if the life cycle cost is the minimum value
[9]. American expert Charles E. Ebeling proposed a life cycle cost general
model [10] and modified model etc. [11], which provided gist for the calculation
of life cycle cost.
This paper establishes a system effectiveness model and life cycle cost model
based on RMS indexes, and proposed a RMS indexes optimization model which
aims to maximize system effectiveness of armored equipment, on condition of the
certain life cycle cost. The optimal RMS indexes are gotten by using parame-ter
time-varying PSO algorithm, which provides reference for practical equipment
development.
Research on Armored Equipment RMS Indexes Optimization Method Based on. . . 293
System effectiveness refers to the ability of fulfilling given quantitative traits and
service requirements in given condition, which compositely reacts system avail-
ability, mission dependability and inherent capability. Relationship between system
effectiveness and the three elements could be reflected from model, which is the
most widely applied model.
T MTBF
A ¼ AO ¼ ð1Þ
T MTBF þ T MTTR þ T MPT þ T MLDT
where TMTBF is mean time between failures, belonging to reliability indexes; TMTTR
is mean time to repair, belonging to maintainability indexes, which reflects the
complexity of corrective maintenance; TMPT is mean time to preventive mainte-
nance, belonging to maintainability indexes, which reflects the complexity of
preventative maintenance; TMLDT is mean logistics delay time, belonging to sup-
portability indexes, which reflects the complexity of planned support resource
adequacy and applicability, as well as support system effectiveness. Apparently,
availability A is totally decided by RMS indexes, preventative maintenance regime,
administration, operation, support and various kinds of elements.
Mission availability refers to the ability of usage and fulfilling given function at any
random moment in specified mission section. Mission availability describes
whether equipment could continuously keep working or not, which measures the
level of required function that panzer could reach during the mission. It is affected
by mission reliability, mission maintainability, security and survivability etc. In the
mission interval, equipment come into failures at a certain probability, and the
faulted equipment go back to the mission at a certain renovating probability.
The condition conversion of equipment between good condition and failure obeys
the Markov process. Mission dependability in moment t could be described as
formula (2):
294 Z. Wang et al.
DðtÞ ¼ eλt ð1 λÞ þ 1 eλt Kμ
ð2Þ
¼ Kμ þ ð1 λ KμÞeλt
T MTBF
E ¼ ADC ¼ ∗
T þ T MTTR þ T MPT þ T MLDT
K
MTBF
T MRF T MTBCF T MRF KT MTBCF T
ð4Þ
þ 1e T MTBCF
∗C
T MRF T MRF T
Research on Armored Equipment RMS Indexes Optimization Method Based on. . . 295
RMS is important influencing factors for equipment, as well as factors that affect
equipment life cycle cost. According to the influence that RMS act on life cycle
cost, we modify the general model set up by American expert Charles E. Ebeling
and build a life cycle cost mathematical model based on RMS indexes [10] which is
described as formula (5).
T MTBF0 β t0
LCC ¼ Fz þ F þ A0 C0 þ A0 Cf Sa ð5Þ
T MTBF T MTBF
where LCC is life cycle cost; FZ is the cost of equipment and labor to produce a
reliable system; β is a constant; T MTBF0 is the statistic of mean time between failures
when equipment is used in practical mission; F is the fixed use cost; C0 is annual
cost of equipment; t0 is annual use time of equipment; Cf is the fixed cost of every
failure; Sa is net residual value of equipment.
With those formulas above, RMS indexes and other parameters, we could
predict the life cycle cost of equipment.
Each particle updated the velocity based on the last velocity and the best position of
history and the whole particle swarm:
V kþ1
j ¼ ωV j
k
þ c r
1 1 P j
k
X j
k
þ c r
2 2 P k
g X k
j ð10Þ
Xkþ1
j ¼ Xjk þ V kþ1
j ð11Þ
c1 and c2 are time-varying learning factors, shown in formulas (13) and (14)
k
c1 ¼ c1f c1g þ c1g ð13Þ
kmax
k
c2 ¼ c2f c2g þ c2g ð14Þ
kmax
The procedure would save ever best solution and stop when iteration frequency
reaches kmax.
298 Z. Wang et al.
Step1: set parameters in algorithm, including ωmax, ωmin, c1f, c1g, c2f, c2g and kmax.
Step2: set k ¼ 1 and generate initial particle swarm S ¼ {X1, X2, , Xm} ran-
∗ ∗
domly, set E∗ i , X and E as particle best solution, global best solution and optimal
target value respectively.
Step3: update particle velocity and location with formulas (10) and (11).
Step4: calculate evolution function value and renew best solution E∗ i of every
particle. If Ei > E∗ and the constraints are satisfied, update global best solution X∗
and optimal target value E∗.
Step5: if k < kmax, k ¼ k þ 1, otherwise, stop calculation and output global best
solution X∗ and optimal target value E∗.
Inertia weight: ωmax ¼ 0.9, ωmin ¼ 0.1, learning factor: c1f ¼ 0.5, c1g ¼ 2.5, c2f ¼ 2.5,
c2g ¼ 0.5, particles size: m ¼ 300, iterations: kmax ¼ 200.
Research on Armored Equipment RMS Indexes Optimization Method Based on. . . 299
0.785
0.780
0.775
0.770
0.765
0.760
0.755
0 20 40 60 80 100 120 140 160 180 200
Iterations
Fig. 1 Experimental result
Programming with Parameter Time-Varying PSO for 100 times in Matlab, the
possibility to get best solution is 95%. Averagely, we get the best solution in
145th iteration and the result is quite accurate. Figure 1 shows a calculation
procedure of getting best solution.
As shown in Fig. 1, algorithm converges in 146th iteration, RMS indexes is
X∗ ¼ (113.3, 5.0, 0.9, 250.0), optimal effectiveness is E∗ ¼ 0.78. Therefore, time-
varying PSO is an effective method in solving RMS indexes optimization problem.
300 Z. Wang et al.
5 Conclusions
This paper establishes a system effectiveness model and life cycle cost model based
on RMS indexes, and proposes a RMS index optimization model which is aimed at
maximize armored equipment system effectiveness under a certain life cycle cost.
Particle swarm optimization (PSO) with parameter time-varying is used to get the
optimal RMS indexes. View from the model and algorithm structure, PSO with
parameter time-varying is same efficient as other models such as system effective-
ness optimization model based on RMS index description, which could also provide
reference for armored equipment RMS demonstration and design.
References
1. Gan MZ, Kang JS (2005) Military equipment maintenance engineering. National Defence
Industry Press, Beijing
2. Xiao HX, Wang JB (2008) Study on system effectiveness evaluation model of weapons and
equipment based on RMS characteristics. Fire Control Command Control 33(5):130–132
3. Yu Y et al (2007) Efficiency evaluation model of weapons and equipments in use phase based
on R&M&S. Armament Autom 26(3):6–7
4. Li JZ, Wang NC (2010) RMS research based on three dmiension model of support system with
simulation method for military aircraft. J Beijing Univ Aeronaut Astronaut 36(12):1485–1489
5. Ping DF, Liu ZY (2013) Effectiveness evaluation of airborne ECM system based on the ADC
model. Aerosp Electron Warf 1:34–37
6. Peng WH, Wang MH (2007) Study of influence on naval gun system efficiency by RMS. Ship
Electron Eng 27(4):192–194
7. Han K, He CM (2014) Trade-off analysis of reliability/maintainability/supportability/testabil-
ity of armored vehicle based on system effectiveness. Acta Armamentarii 35(2):268–272
8. Du HD, Wu W (2013) Simulation research for RMS indexes of armored equipment based on
system effectiveness analysis. J Syst Simul 25(8):1947–1950
9. Geng JB, Jin JS, Zhang J (2010) Tradeoff analysis for RMS of equipment based operation
availability and life cycle cost. Int J Plant Eng Manag 15(1):18–21
10. Ebeling CE (2010) Reliability and maintainability engineering. Tsinghua University Press,
Beijing
11. Qiao SB et al (2012) Warship’s RMS based on LCC of balance technology analysis. Equip
Manuf Technol 1:90–93
12. EberHart RC, Kennedy J (1995) A new optimizer using particle swarm theory. In: Proceedings
on 6th international symposium on micromachine and human science, pp 39–43
Dealing with Uncertainty in Risk Based
Optimization
Ype Wijnia
1 Introduction
Y. Wijnia (*)
Asset Resolutions B.V., P.O. Box 30113, 8003CC Zwolle, The Netherlands
e-mail: [email protected]
According to Aven [3], there is no generally agreed upon definition of the concept
of risk. In everyday language risk may be used to indicate threat, probability,
consequence or even the object at risk [4], though in literature many more defini-
tions can be found [3]. If the many definitions of the concept have to be structured,
the most differentiating aspect is whether risk is used to indicate the thing that could
result in unexpected outcomes (risk as an entity) or that risk is the measure for the
Dealing with Uncertainty in Risk Based Optimization 303
Potential Likelihood
Finance Safety Reliability Unlikely Remote Probable Annually Monthly Weekly
Severity class <0,003 <0,03 <0,3 <3 <30 >=30
Extreme > 10 M€ Several > 20 M cml
fatalities M H VH U U U
Fig. 1 Risk matrix for energy distribution, derived from NTA8120 [14]. Risk levels are indicated
by Negligible, Low, Medium, High, Very High, Unacceptable
between 1 and 10 million (if geometric mean is used about 3M). This allows
virtually any asset management decision on risk to be monetized.
Even though the cost benefit analysis of a single intervention can be regarded as an
optimization (it is still about the best option), normally optimization requires
multiple alternatives to be evaluated. This can be a choice between technologies
(like several types of switchgear), the optimal size within a single technology (like
the cable diameter for expansions) or just a timing option for a single decision. Most
decisions on existing assets fall into the timing category. The typical example
(as referred to in the introduction) is the optimization of the maintenance interval,
but upgrades, expansions and replacements also have an element of timing.
3 Case Introduction
In the electricity distribution grid cables make up for more than 50% of the total
asset value. Decisions on the replacement and upgrades of those cables thus are of
vital importance. Figure 2 gives a typical situation for decision making. A medium
voltage substation is connected to a high voltage substation by means of two cables.
Dealing with Uncertainty in Risk Based Optimization 305
Peak loads in the distribution grid tend to grow continuously, with some 3% per
year. The peak load of the MV substation above has reached the redundant power
(i.e. the power that can be provided with one of the cables out of operation). This
means that in case of a single cable failure the other cable could be in an overload
situation, potentially resulting in an overload induced failure. What is the right
moment to upgrade the circuit with a third cable?
Classically, this decision would be resolved in by comparing the potential peak
load with the nominal rating of the cable. This would result in upgrading right now.
Two considerations result in postponement of the decision. First of all, the peak
load occurs only 1 h/year. Figure 3 shows a typical load duration curve.
Only if the first cable would fail precisely during this peak load, an overload
would result. But that probability is currently 1 in 8760, and will only slowly
increase over time. The use of peak demand is therefore risk averse. Unfortunately,
the load pattern of cables is normally not measured, only the peak load.
The second consideration is that peak load is compared with the nominal rating,
which is the load the cable can endure for 30 years continuously. At least part of the
306 Y. Wijnia
The classical comparison of peak load with the nominal rating of the cable can be
regarded as a risk free approach. No matter at what moment the failure occurs, the
remaining cable will be able to supply the full load. However, this is not the
maximum load level to be risk free. Even at the dynamic power rating, there will
be no risk of thermally induced failures, simply because the cable temperature will
not rise above that of the nominal rating due to the load cycle. But beyond this
dynamic rating there is a risk of thermal overload.
Unfortunately, the mechanism of thermally induced overload is not well under-
stood. On the one side, there is the understanding that chemical degradation is
strongly related to temperature, but that a few hours of overload will not eat away
too much lifetime. Cables may run at a double load (twice the nominal rating) for
several hours without considerable life expectancy effects. If this is the case (the
best case approach, with no overload induced risk), the decision should only be
based on the energy losses. The left part of Fig. 4 compares the cost of the energy
losses with the annual equivalent investment costs. According to this diagram, the
investment should be postponed until (roughly) the full capacity was used in an
everyday situation, but beyond that the losses would pay for a new cable.
On the other hand, it is also known that every cable has hotspots (points which
heat up more than the normal cable), like the joints. Furthermore, the resistance of
metals is generally positively correlated with the temperature, meaning that hot
cables heat up faster. Therefore the overload risk may possess runaway
Dealing with Uncertainty in Risk Based Optimization 307
Fig. 4 Optimization of upgrade based on upgrade costs and energy loss (left) or upgrade costs and
overload risk (right)
characteristics, with a trouble free operation and then sudden failures. The worst
case approach is that these failures occur at even 1 h above the safe limit. The right
part of Fig. 4 shows the overload risk compared to the investment costs. According
to this diagram (comparing risk and annual equivalent costs of the investment)
some risk can be taken, but not much. At about 10% above the safe dynamic limit of
12 MVA the new cable pays out.
The difference between these extreme approaches is significant, about 6 MVA in
load level. Given the growth rate of 3% per year, this is almost 20 years difference.
Figure 5 holds all optimizations discussed so far. Additionally a best guess (average
between worst and best case) has been added.
Especially given the large differences in Total Cost of Ownership (TCO), it
seems very difficult to make the right decision. However, this changes if the
optimizations are reviewed in terms of relative performance (i.e. the cost relative
to the optimal value in the scenario). This is shown in Fig. 6. In the enclosed
optimum (where the relative curves of best and worst case intersect) the perfor-
mance is equally bad with regard to both worst and best case, thought this choice is
308 Y. Wijnia
in either case only 10% more expensive. As reality is somewhere in between the
extremes, the premium towards reality is less. For example, the premium (or regret)
towards the best guess is only 2%. This means that even though the precise failure
mechanism is not fully understood and large uncertainties in failure probabilities
exist, it is possible to make a very good and robust decision.
The other major uncertainty in the decision is the actual value of the failure. Most
obvious is the value of the interruption to the customers. The direct value of an
interruption is estimated at some 0.50 € per CML, though that value can rise
significantly if the customers regard the network operator to blame. Another
uncertain factor is the precise duration of the repair. The first failed cable can be
expected to be repaired in 8 h, but if the second cable then fails due to overload, the
first cable cannot be used to restore power completely. Additional measures like
emergency power supplies are needed. It is probably more like 24 h before every
customer is back on. The third uncertain factor is the damage to the overloaded
cable. Thermally induced failures cannot be repaired, as the insulation will be
degraded too much. Therefore, at least parts of the cables will have to be replaced,
with a significant probability that the whole cable needs replacement. Total
expected cost estimate would be in the neighborhood of the cost of replacing the
cable, but with significant uncertainty.
Dealing with Uncertainty in Risk Based Optimization 309
Interestingly, optimizations do not depend that much on the precise value of the
consequences. This is demonstrated in Fig. 7. Even if only the relatively certain
costs (upgrade costs plus energy loss) are regarded in the best guess assumptions, it
results in an outcome some 25% more expensive that the true optimum. With half of
the risk captured (either repair costs or outage, which are about equal in this
example) the difference is less than 5%. Optimization based on 50% of the risk
then is more than 95% accurate. For normal practice that would be good enough.
References
1. Woodhouse J (2014) Asset management is growing up. Tutorial at the 9th WCEAM, Pretoria
2. Wijnia YC (2015) Towards quantification of optimality in asset management (to be publsihed).
WCEAM2015. Tampere, Finland
3. Aven T (2012) The risk concept—historical and recent development trends. Reliab Eng Syst
Saf 99:33–44
4. Slovic P, Weber EU (2002) Perceptions of risk posed by extreme events. In: Risk management
strategies in an uncertain world, New York
5. ISO (2009) ISO 31000: risk management- principles and guidelines
6. ISO (2014a) ISO 55000 asset management-overview, principles and terminology, Geneva
7. ISO (2014b) ISO 55001 asset management-management systems-requirements, Geneva
8. ISO (2014c) ISO 55002. Asset management-management systems-guidelines for the applica-
tion of ISO 55001, Geneva
9. Aven T (2011) On the new ISO guide on risk management terminology. Reliab Eng Syst Saf
96:719–726
10. Leitch M (2010) ISO 31000:2009—the new international standard on risk management. Risk
Anal 30:887–892
11. Klinke A, Renn O (2002) A new approach to risk evaluation and management: risk based,
precaution based and discourse based strategies. Risk Anal 22:1071–1094
12. Wijnia Y (2012) Asset risk management: issues in the design and use of the risk matrix. In:
Mathew J, Ma L, Tan A, Weijnen M, Lee J (eds) Engineering asset management and
infrastructure sustainability. Springer, London
13. Cox LA Jr (2008) What’s wrong with risk matrices? Risk Anal 28:16
14. NEN (2009) NTA 8120:2009 asset management—Eisen aan een veiligheids-, kwaliteits- en
capaciteitsmanagementsysteem voor het elektriciteits- en gasnetbeheer
15. Wijnia YC, Herder PM (2005) Options for real options: dealing with uncertainty in investment
decisions for electricity networks. International conference on systems, man and cybernetics,
Hawaii
16. IEC (1985) IEC 60853: calculation of the cyclic and emergency current rating of cables. Part 1:
cyclic rating factors for cables up to and including 18/30 (36) kV
17. Wijnia YC, Peters JCFM (2008) Integrating sustainability into risk based asset management.
International conference on infrastructure systems: building networks for a brighter future,
Rotterdam
The Design and Realization of Virtual
Maintenance and Training System of Certain
Type of Tank
Abstract Starting from the practical equipment maintenance and training, the
designing thought, scheme and key technology of the virtual maintenance and
training system of certain type of tank are elaborated in this paper. This article
also introduced the method to realize the virtual maintenance and training system.
Furthermore, an expectation on the developmental direction of virtual maintenance
and training is also given in the paper.
1 Introduction
Along with the constant deepening of the preparation of military struggle, new types
of armament have been equipped to troops in succession. The new equipment has the
prominent features of small number, complex structure, high scientific and techno-
logical content, and expensive materials of new types of equipment. The traditional
real equipment maintenance and training are highly priced and low efficiency. In
addition, tank components and parts are often damaged because of man-made mis-
takes and the high cost-benefit ratio. All these problems are particularly standing out;
therefore, it is difficult to meet the requirements for new equipment maintenance and
training. Fortunately, with the development of computer science and technology, it
is mature to adopt the thought of virtualization and the technology of simulation to
construct a training system and platform as an important auxiliary means for new
equipment maintenance, support and training [1, 2].
L. Xu (*)
Department of Technical Support Engineering of Academy of Armored Force Engineering,
Beijing 100072, P. R. China
e-mail: [email protected]
S. Wang • Y. Li • L. Ma
Department of Basic Courses of Armored Force Engineering, Beijing 100072, P. R. China
e-mail: [email protected]
2 System Design
secondary secondary
End user System management development user development user
Translation
maintenance Component modules Tools and materials bank
learning bank of certain type of of certain type of tank
tank
maintenance Expansion dynamic
examination link bank
server
Fig. 1 Virtual maintenance and training system for certain type of tank
The virtual maintenance and training system for certain type of tank can be run on
single computer and on the LAN, using SolidWorks as cartographic software to
establish 3-D models for tank parts and components, maintaining tools and mate-
rials. Using Cortona3D Rapid Learning platform as interactive engine for mainte-
nance actions, and using HTML files for independent release.
3 Key Technology
Looking at the past modelling experience and process, we find that the most virtual
technologies adopted the software like 3D MAX, MAYA. The type of modelling
system is based on polygonal modelling system, for example, adopting Loft, Edit,
Poly, Mesh, and HSDS. Comparatively, it is appropriate to consider this type of
modelling method under static art modelling is presented, or the required accuracy
is not high.
However, due to the high requirements of the arm or equipment maintenance and
training system for the size, accuracy, complexity, and consistency with real object
structure, it is very hard to achieve the desired effect if the polygon-based sys-tem is
adopted to construct primitive 3-D modelling data. Therefore, the comprehensively
selected 3-D digital modelling methods are considered which is described as
follows:
1. Adopting entity modelling software such as SolidWorks, ProE, etc, to design the
primitive digital models of the equipment assemblies, parts and components.
314 L. Xu et al.
2. Lightening the weight of the primitive digital models and save them as poly-gon-
based 3-D digital models (*.wrl) that can be directly used by the system.
3. Taking the files of *.wrl format as standard 3-D digital model guiding format for
the virtual maintenance and training system.
The 3-D visual scene control of the virtual maintenance and training system
involves hidden-surface elimination, illumination model and surface algorithm,
cut-ting algorithm, topographic data display and topography matching optimiza-
tion. This article highlights the exploration of the coordinate transformation within
three- dimensional space. Coordinate transformation can be divided into subject
position transformation, and coordinate transformation from one system to another
system.
As for the subject position transformation within a coordinate, it involves three
kinds of transformations such as translation, zoom and rotation. Sometimes it also
includes reflection and transvection transformation. Taking a translation as
ex-ample, in the 3-D homogeneous coordinates, the arbitrary point P (x, y, z) can
be transformed into point P(x0 , y0 , z0 ) using the transformation formula Eqs. (1) and
(2):
2 0 3 2 3 2 3
x 1 0 0 tx x
6 y0 7 6 0 1 0 ty 7 6y7
6 07¼6 7•6 7 ð1Þ
4z 5 40 0 1 tz 5 4 z 5
1 0 0 0 1 1
0
P ¼ T•P ð2Þ
Or, in the 3-D space the translation of an object can be realized through the
object to be translated. As for the object represented by a group of polygons, the
vertexes of each surface can be translated and then the transformed position can be
expressed as.
2 0 3 2 3 2 3
x Sx 0 0 0 x
6y 7 6 0
0
Sy 07 6 7
6 07¼6 0 7•6y7 ð3Þ
4z 5 4 0 0 Sz 05 4 z 5
1 0 0 0 1 1
0
P ¼ S•P ð4Þ
The premise and basis for improving the training quality with virtual maintenance
and training system is to establish a set of scientific and reasonable maintenance
procedures. In the course of virtual maintenance training, the subjects (equipment
parts and components, tools and materials) in the training scene are constantly
changed, so it is needed to establish a special model to describe the maintenance
procedures and their constraints. This article adopts Finite State Ma-chine (FSM) to
describe the maintenance procedures and uses starting condition controlling func-
tions to realize the procedure constraints.
1 condition that meets the starting of procedure i
Si ¼ ð5Þ
0 condition that doesn’t meets the starting of procedure i
1 condition that procedure i is done
Fi ¼ ð6Þ
0 condition that procedure i isn’t done yet
Fi ¼ ϕðS1 ; S2 ; . . . ; Sn Þ ð7Þ
4 System Realizations
According to the functions required by the virtual maintenance and training system
for certain type of tank, the virtual maintenance and training system framework is
composed of visual scene controlling platform, data bank system, virtual mainte-
nance inactive engine and system management platform. The visual scene control-
ling platform mainly realizes setting simulation; the data bank system is used to
store equipment parts and components, tools and materials; the virtual maintenance
inactive engine is used to realize man-machine interaction for the maintenance
actions; the system management platform is used to control the system running for
demonstration, learning and examination.
The realization process of system involves SolidWorks 3-d modelling; using
Cortona3D interactive engine to disintegrate, set up and edit the maintenance
procedures and interactions; using Rapid Learning platform to edit the functions
like demonstration, learning, and examination; using Virtual Training Viewer to
make independent release with HTML files (see Figs. 2, 3, and 4).
316 L. Xu et al.
5 Conclusions
As a brand new training mode, the virtual maintenance and training system have
aroused the attention of many experts and scholars in the field of equipment
support. The virtual maintenance and training system can provide advanced exper-
imental environment and simulating means for the arm or equipment maintenance
and training. It can also greatly facilitate the training effect, promote talent culti-
vation quality and improve the level of scientific re-search. The virtual maintenance
and training system has become a main developmental trend of the maintenance
and training technologies, hence it should be given extensive attention on it.
References
1. Xing HG, Wang GH, Zhou SH (2014) Virtual maintenance system for certain type of tank gun. J
Sichuan Ordnance 35(7):1–4
2. Zhao C, Li XX, Xia K et al (2014) Control method of concurrent operation for collaborative
virtual maintenance. Comput Syst Appl 23(10):198–201
Measures to Ensure the Quality of Space
Mechanisms
Abstract Space mechanisms (SMs) refer to the devices attached to the space-
craft to perform defined functions by mechanical movements. It is very impor-
tant to assure the quality of SM to meet the requirements of spacecraft, because
most quality defects of SMs can result in severe performance degradation or
failure of the spacecraft. Based on 20 years’ experience of SMs research and
development, a set of critical measures that have been found to be extremely
important in quality assurance in the most demanding environments are identi-
fied. The measures can be adopted in different phases of the development
process, such as design phase, manufacture phase, ground validation phase,
storage and transportation phase, final assembly and integration phase, etc. It
has been proved that these measures can ensure the quality of SM effectively,
and can also contribute to the research and development project for new SM.
1 Introduction
Space mechanisms (SMs) refer to the devices which are attached to the spacecraft to
perform defined function by mechanical movements [1, 2]. Their main functions
contain the hold-down and release of two components of a spacecraft or a component
and a spacecraft, the docking and separation of two spacecrafts, the deployment and
locking or shape preserving of component, landing shock alleviation of spacecraft or
astronaut, on-orbit vibration attenuation of high precision payload, tracking and
pointing of the payload, orientation adjusting of whole spacecraft, etc. The spacecraft
is usually equipped with many kinds of SMs with different functions. It’s very
important to assure the quality of SMs, because most quality defects of SMs can
result in a fatal performance degradation or failure of the spacecraft [3, 4]. SMs will
generally suffer severe environment conditions of launch and on-orbit, which can
result in unignorable faults. Besides, it’s extremely difficult, risky and costly to repair
or maintain SMs on-orbit. The high reliability of SM is therefore quite important. For
example, in order to provide steadily energy for the communication satellite during
the whole life cycle, the driving mechanism which drives the solar wing pointing to
the sun is ordinarily required to work normally at least 15 years in the space [5–7].
To insure SMs work reliably on-orbit, some researchers have discussed the
measures of design [8–10]. However, measures of only design stage cannot assure
the quality of SM. To insure SMs’ reliability, measures of the whole development
process should be considered. It has been proved in engineering that appropriate
control measures in all of SM development phases of design, manufacture, ground
validation, storage and transportation, final assembly and integration should be
taken into consideration to assure the SMs’ reliable work on-obit.
2 Design Measures
For important documents such as the precept design report need to be multi-level
checked and signed before they become effective. Ordinarily, the multi-level
checks contain the signatures of designer, corrector, reviewer and approver. Some-
times it needs the countersignature of the technicians of other departments. A
different signer has different and definite responsibility. With the development of
the signature level, the signer will often ask some specialists to work together to
dealing with the problems and to developing ideal solutions. Only when the related
solutions are fully discussed, the signer will put his signature. The adoption of
multi-level check and signature integrates the collective wisdom into the scheme,
which can insure the scheme’s feasibility, scientificity, comprehensiveness and
pertinence. With the strategy of multi-level checks and signatures, the limitation
and randomness of the scheme can also be avoided effectively. The quality of SMs
can be fundamentally assured by this measure.
The important precept designs can only take effect when they have passed multi-
level reviews. In general, it contains the reviews of engineering group, local
department, user, and so on. Different levels of reviews will invite different
specialists to discuss problems and evaluate the solutions. Only when the lower-
level review has been gotten through can the higher-level review be implemented.
In each review, the replies to all of the experts’ questions should be summarized. If
the advice is accepted, the implementation measures should be pinpointed. How-
ever, if the advice is denied, the reasons for rejection should be clarified in detail.
All the replies will be fed back to the relevant experts. Only when all the experts
agree with the answers and give their signatures, can the review be passed. The
implementation of this regulation involves the knowledge and skills of many
experts into the scheme, so the feasibility and scientificity of the scheme will be
further ensured. The quality of SMs is further assured too.
In order to decrease the mass or the volume of the SM, the high strength materials
are usually adopted, or the special heat treatments are usually taken to improve the
load bearing capacity of the part. To deal with heavy loads, especially the tension
load, the strength and the toughness need to be both considered at the same time. If
the strength is over emphasized and the toughness is neglected, the brittle fracture
can be induced. It’s important that both the strength and the toughness of the high
load bearing part are required explicitly. As a result, the load bearing capacity can
be ensured and the brittle fracture can be avoided. In addition, stress concentration
should be also eliminated on high load conditions.
For some SMs, the surrounding environment temperature on-orbit may change
greatly. Deferent parts for one joint (e.g., friction pair) should be manufactured
with the same material, in order to avoid the failure of joint movement caused by
the different coefficients of expansion. However, once the temperature changes too
fast or too much, the uneven expansion caused by the temperature gradient cannot
be avoided by the above mean. In this case, additional thermal control measures
should be taken. So the temperature gradient can be limited within a safe range, and
the movement fault of the joint can be avoided.
All the joints of SM should be lubricated. The appropriate lubricating methods can
be selected by the joint’s running speed, on-orbit life, load conditions, etc. When
solid lubrication film is selected, the effect of the film thickness on the properties of
joint movement should be considered thoroughly. The lubrication film can be
applied to the surface of one part of the two mating parts, if the other part is
difficult to lubricate. For example, when the slender shaft-hole fitting is coped with,
lubrication film can be applied only to the surface of the slender shaft.
The spring or pyrotechnic device is usually used instead of electric motor to drive
the SM which only works once on-orbit. Because the segments subject to potential
failure in spring and pyrotechnic devices are much fewer than electric motor, so
they are more reliable than electric motor. And the quality and performance of the
Measures to Ensure the Quality of Space Mechanisms 323
3 Manufacture Measures
The management of the technology scheme is the same as the control of design
scheme. In the implement process of the technology scheme the three-level checks
are always emphasized, which is check by oneself, check through each other and
check by professional. Detailed records for every check and for the handover of
different procedures are all made and preserved for a long time. These records will
provide convenience for the traceability of the probable problems that appear in the
implement process. If the out-of-tolerance condition happens, the quality control
department will have a meeting with relevant people, such as designer, technologist
and processing worker to discuss the reason and the solution of the problem. Two
goals will be achieved through the technology management. The first one is to
ensure that the performance of each part meets its requirements. And then the
quality of the whole mechanism can be assured. The other one is that the quality of
the whole mechanism can be ensured by taking some necessary measures, even if
the performance of some parts doesn’t meet the requirements. For example, if the
diameter of a shaft is smaller than required, the diameter of the relevant mating hole
can be smaller and then the mating performance can be kept the same.
There is always no sign before the hydrogen embrittlement happens. Thus its
influence on SM is often fatal. For the materials with strength of more than
1400 MPa, the electroplating corrosion protection is forbidden in order to avoid
the hydrogen-induced brittle of the parts. And anticorrosion coating of physical or
electroless nickel deposits will be adopted instead. In addition, the acid cleaning for
the part should be also avoided in the whole process to eliminate the probability of
the part to contact with hydrogen.
The heat treatment of the titanium alloy should be implemented in a vacuum
tank to avoid the alloy contact with hydrogen in the air at the high temperature.
Hydrogen-induced brittle of the part which is made of titanium alloy can be avoided
effectively by this method [2].
324 J.-Z. Yang et al.
A list of all parts and materials is always made before assembly, which contains
standard parts, purchased parts, manufactured parts, auxiliary materials, equipment
and tools, etc. And then they will be cleaned and checked one by one according to
the request. The unqualified samples should be picked out and a record will be made
in detail. The sealing rings and sealing surface should be checked more carefully
with magnifier to ensure that there is no scratch reducing the seal performance.
Electric power and control signal are necessary for the mechanism which will
operate on-orbit for a long time. So the cable is usually an important component
for SM. In order to avoid the damage of cable and plug in the launch period, it’s
necessary to bind and fix the cable properly. In order to eliminate the loosening of
plug or the additional force acting on plug as well as the probable resulting
resistance of the mechanism, the fixed point and diameter of bunched cable should
be determined appropriately. In order to prevent the cable from getting hard due to
Measures to Ensure the Quality of Space Mechanisms 325
the low temperature environment, thermal control measures should also be taken to
avoid addition resistance to the joint movement caused by the hard cable.
Comprehensiveness and safety of ground validation are both very important for
quality assurance of SM. It contains the effectiveness and reasonability of valida-
tion scheme, the accuracy of measured parameters, pre-validation and the compre-
hensiveness of data analysis.
Effectiveness of validation scheme refers to effective simulation of on-orbit
environmental conditions and validation of the mechanism performance under the
conditions. The reasonability of validation scheme refers to low cost, mature
equipment, low risk and easy implementation. Sometimes it’s very difficult to
simulate all the space conditions. In this situation, the influences of environmental
condition parameters on the performance of the mechanism need to be fully
analyzed, and only those with distinct effects will be simulated.
Accuracy of parameter measurement refers to that the method is reliability, the
equipment is mature and all the critical measurement points or data have a backup.
In order to measure the parameters accurately, the contractor will devise the
implementation program in detail, and the program will be treated according to
the design scheme. So the validation target will be achieved smoothly.
To ensure the safety of important validation, one or two pre-test will be
conducted to valid the effectiveness of the program and the conditions of the
equipment. Improvement measures will be taken according to the exposed prob-
lems. Thus the formal experiment can be conducted smoothly and effectively, and
the safety of personnel and samples can also be assured.
The results of validation are often different from the expectation. The difference
can always be explained by analyzing the measurement data. And the validity of the
data and the success of the experiment can be judged correctly. In addition, some
performances of SM can be determined only by the summary and analysis of the
measurement data. For example, the reliability of SM can only be assessed based on
the measurement data of the reliability test. The reliability of landing gear of
Chang’E-3 Probe has been assessed by this way [11].
Many parts of SM will be stored for a long time before the final assembly. In order
to assure the stability of the quality, all the storage demands must be met including
conditions of cleanliness, temperature and humidity, etc. For some special parts,
such as slender rod and thin wall cylinder, need to be supported by special
equipment, so that they will not be deformed by the action of the weight.
326 J.-Z. Yang et al.
6 Conclusions
References
1. Chen LM (2005) Spacecraft structures and mechanisms. China Science and Technology Press,
Beijing
2. Yu DY, Yang JZ (2011) Spacecraft mechanism technology. China Science and Technology
Press, Beijing
3. Cong Q (2012) The advance of integration technology for space mechanisms. Spacecr Environ
Eng 4:384–387
Measures to Ensure the Quality of Space Mechanisms 327
4. Ma XR, Yu DY, Sun J (2006) The researching evolvement of spacecraft deployment and
driving mechanism. J Astronaut 6:1123–1131
5. Chen J (2007) Development of foreign communication satellite technology. Aerosp China
2:38–46
6. Liu H (2012) Reliability analysis of foreign commercial communication satellite platform.
Space Int 11:14–19
7. Shao RZ, Fan BY (1996) Reliability study of long-life communication satellites. Chin Space
Sci Technol 4:24–33
8. Peng CR (2011) System design for spacecraft. China Science and Technology Press, Beijing
9. Yang JZ, Zeng FM, Man JF (2014) Design and verification of the landing impact attenuation
system for chang’E-3 lander. Sci Sin Technol 5:440–449
10. Yuan JJ (2004) Design and analysis of satellite structure. Astronautic Publishing House,
Beijing
11. Wu Q, Yang JZ, Fu HM (2014) Deployment reliability test and assessment for landing gear of
Chang’E-3 probe. J Donghua Univ 6:782–784
12. Yang JZ (2015) Landing gear of spacecraft. Astronautic Publishing House, Beijing
13. Yang JZ, Wu Q, Man JF (2015) Reliability design and validation for the landing gear of
Chang’E-3 lander. In: Proceedings of international conference on quality, reliability, risk,
maintenance, and safety engineering (QR2MSE), Beijing, China
A Decision-Making Model of Condition-Based
Maintenance About Functionally Significant
Instrument
Xiang Zan, Shi-xin Zhang, Yang Zhang, Heng Gao, and Chao-shuai Han
1 Introduction
With the development of high and new technology and fuzz of fault law and failure
mode, limits for validity of time-based maintenance is reducing. CBM (Condition-
based Maintenance) is valued gradually, which can insure reliability, improve
operational availability, reduce cost of maintenance support.
Many decision-making models and methods of CBM are developed by scholar
in home and abroad. Based on markov chain theory, markov decision-making
model is established to describe law of condition changed. Levy process model
can describe deterioration process on the condition of continuum damage. Delay-
time model is used to divide change of condition into two stages. For example,
Salvinder and cooperators established assessment model based markov chain to
assess condition change of vehicle crank-link mechanism, and based the result of
assessment the result decision-making can be acquired [1]. Sha and cooperators
researched step-stress accelerated life testing through proportional hazards model,
and combine proportional hazards model with Weibull model to establish decision-
making model. At last, showing the model worked well and effectively through an
example [2]. WANG Wen and cooperators analyzed the effect of drop height on the
solder joints’ life span through proportional hazards model [3].
2 Related Theory
Most faults occur in gradual process. The Process can show in Fig. 1. In Fig. 1,
O shows fault-beginning point, which time fault begins to happen. P is Potential
Failure point, which abnormal condition can be inspected. F is Functional Failure
point, which fault finally happens. The time between P and F is P-F interval [4].
The fundamental principle of CBM is that if P-F interval is large enough, fault
sign can be discovered to prevent fault from happening.
Equipment
Condition F Functional Failure point
P-F interval
Equipment Life
PHM is present by Cox in 1972 [6]. Then Jandine and others put PHM in CBM, in
which condition parameters, service load, fault and so on are looked as follow
factors of equipment life. Product property theory can appear in real risk of
equipment by PHM [7]. The property of PHM is the hazard of different individuals
is proportion [8].
The form of PHM shows below.
In the expression, λ(t,X) is fault probability, λ0(t) is the only related to fault
probability, X is condition number in t, β is regression variable efficiency, which is
used for establishing the relation between condition number and fault probability.
Due to Weibull model can match most life distribution of mechanical products,
WPHM (Weibull Proportional Hazards Model) is used for establishing reliability
model of equipment.
Integration of PHM and two-parameter Weibull model, WPHM can be acquired,
which can be used for describing the relation between equipment condition and
fault probability. The expression can show below.
δ t δ1
λðt; XÞ ¼ expðβXÞ ð2Þ
α α
3 Decision-Making Model
Y
q Y
n q Y
Y k
L¼ λðT i Þ Rð T i Þ i
p Xj1 ; Xji ð3Þ
i¼1 i¼1 i¼1 j¼1
In the expression, R(Ti) shows reliability function of i, λ(Ti) shows fault proba-
bility function of i. p(Xj1i,Xji) shows probability from Xj1i to Xji.
Based on Eq. (2), reliability function with condition parameters can be
acquired.
0 1 0 1
Zt Z t δ1
δ s
RðtjXÞ ¼ exp@ λðt; XÞdsA ¼ exp@ expðβXÞ dsA ð4Þ
α α
0 0
Put Eqs. (2) and (4) into Eq. (3), the result shows below.
A Decision-Making Model of Condition-Based Maintenance About Functionally. . . 333
0 0 11
q δ1
Y Y Z t δ1
δ s n
@exp@ δ s
Lðδ; α; βÞ ¼ expðβXÞ expðβXÞ dsAA
i¼1
α α i¼1
α α
0
Y k
q Y
i
p Xj1 ; Xji
i¼1 j¼1
ð5Þ
ln ðLÞ
X q X q Xn Z δ1
t
δ s δ1 δ s
¼ q ln þ ln þ βX expðβXÞ ds
α i¼1
α i¼1 i¼1
α α ð6Þ
0
n X
X m
þ i
ln p Xj1 ; Xji
i¼1 j¼1
In assumption
Ztj δ1
δ s
U¼ expðβXÞ ds ð7Þ
α α
0
Because assuming the law of state transition is right indicial function, Eq. (7) can
show below.
m Z δ1
tj
X δ s
U¼ expðβXÞ ds
α α
j¼1
tj1
ð8Þ
Xm
tj δ tj-1 δ
¼ exp βX tj
j¼1
α α
ln ðLÞ
X X
δ
q
s δ1
q Xm X m tj δ tj-1 δ
¼ q ln þ ln þ βi X exp βi Xi, j
α i¼1
α i¼1 i¼1 j¼1
α α
XXn m
þ i
ln p Xj1 ; Xji
i¼1 j¼1
ð9Þ
4. Iteration
Based on meaning of Weibull model parameters, combined with experience
of practice, every parameter can be determinate below.
8
< δ 2 103 ; 15
α 2 103 ; 5000 ð10Þ
:
βi 2 ½10; 10
Based on failure risk to decision-making of CBM, the target is to acquire the best
inspection interval through controlling failure risk. In general, through setting a
threshold or acceptable range of failure risk, making dynamic decision about
inspection interval of CBM.
Putting Eq. (2) into Eq. (11), expression can be acquired.
0 1
Z
δ s δ1
tþΔT
Because assuming the law of state transition is right indicial function and putting
^ , ^δ and β^ i , Eq. (12) can show below.
α
h i
exp β^ X ^ ^
r ¼ 1 exp ðt þ ΔT Þδ tδ ð13Þ
^ ^δ
α
From Eq. (13), the equation of inspection interval can show below.
A Decision-Making Model of Condition-Based Maintenance About Functionally. . . 335
" #^1
ln ð1 r Þ
^δ ^
δ
ΔT ¼ t
α ^δ t ð14Þ
exp β^ X
4 Case Study
In the same condition, engine oil data of two same type vehicles (called vehicle A
and vehicle B) is detected at regular time. Main metallic wear data are chose as
samples. The data of vehicle B is used for parameter estimation. The data of vehicle
A is used for test data. After vehicle B data normalized, result shows in Table 1.
Through the method of principal component analysis in Document [11], the
result shows the profit estimation of the first principal component is 92.43%. The
first principal component almost includes all information of samples. At last,
condition parameter of vehicle B can be acquired as in Table 2.
15 t 14
λðt; XÞ ¼ expð0:63XÞ ð16Þ
628:75 628:75
1. Data processing
Data of vehicle A is preconditioned, normalized, principal component ana-
lyzed; condition parameter of vehicle A can be acquired as in Table 3.
2. Determination on inspection interval
Based on Eqs. (14) and (15), the expression of inspection interval can be
acquired below.
151
ln ð1 r Þ
ΔT ¼ t 15
628:7515 t ð17Þ
expð0:63XÞ
A Decision-Making Model of Condition-Based Maintenance About Functionally. . . 337
140
120
100
Inspection Interval
80
60
40
20
0
420 440 460 480 500 520 540 560 580 600 620
Life (h)
The more serious failure risk, the more dangerously equipment runs. The lower
serious failure risk, the more expenses equipment maintenance is. Based on expe-
rience, failure risk between 0.02 and 0.05 can be accepted.
After engine running a series time, condition descending and inspection interval
need decrease to acquire condition of engine. The dynamic decision-making result
of vehicle A engine after 430 motor-hours shows in Fig. 2.
Figure 2 shows decision-making range of inspection interval through controlling
failure risk. In general, with life of equipment rising and degradation degree of
condition rising, inspection interval need decrease to acquire change of equipment
condition. It can confirm safety and task be controlled in acceptable range. Through
338 X. Zan et al.
controlling failure risk, upper bound and lower bound of inspection interval at
different stages can be acquired. Decision maker can make reasonable decision
after considering different factors.
5 Conclusion
References
Xiang Zan, Shi-xin Zhang, Heng Gao, Yang Zhang, and Chao-shuai Han
1 Introduction
which are forward propagation of signal and back propagation of error. In the study
process, if forward output signal is mismatching with real result, weight and
threshold will be revised through back propagation. Weight is continually adjusted
by forward and back propagation to train the best neural network [7].
Basic BP neural network constitutes of input layer, output layer and hidden
layer, and there may be many hidden layers [8]. There is a network including
M layers, P nodes and N samples (xk,dk) (k ¼ 1,2,. . .,N ). If Ijkm shows input total of
j-th note in m-th layer, output is Ojkm and Wij shows weight between i-th note in m-
1-th layer and j-th note in m-th layer, result can be acquired below.
X
nl
I jkl ¼ W ij Ol1
jk ð1Þ
i¼1
Because the number of notes in every hidden layer is uncertain, assume the
number is nl.
Ojkl ¼ f I jkl ð2Þ
1 X N
E¼ Ek ð4Þ
2S k¼1
∂E
W ij ðt þ 1Þ ¼ W ij ðtÞ η ð5Þ
∂W ij
vij wjp
x1
y1
x2 ̤̤̤
̤̤̤
̤̤ yi
xi
v ij w jp
x1 y1
)HDWXUH 6WDWH
VSDFH x2 'LDJQRVLV
̤̤ VSDFH
6 UHVXOW
&
̤̤ yi
xi ̤̤
2. Initialization populations
The binary initial populations is produced random, whose initial length is
confirmed by the length of vij,wij,aj,bp (i,j,p ¼ 1,2,3,. . .).
3. Crossover
Single cutting-point crossover is used. Selecting strategy of directly propor-
tion is applied to parent-chromosomes. Middle parts of elder chromosomes are
exchanged. Inherit-chromosomes can be acquired.
Mutation is changing location value of some codes selected chromosomes
4. Fitness function
Because the target of Genetic Algorithm is reducing error of BP neural
network output, error between output values of samples and real values is used
as fitness function.
5. Termination condition
When the time of iteration can reach to setting times, operation is over.
5 Case Analysis
1. Case data
Fault Diagnosis of equipment transmission is the case. Based on text [10], six
parameters are chosen as feature information in Fig. 4.
Research on the Fault Diagnosis Method of Equipment Functionally Significant. . . 345
start
Calculate on FitnessValues of
Determinate on Initial Structure of
Individuals in Population
BP Neural Network
Choose Parent-chromosomes
Based on Selecting Strategy of
Initialize Weights and Thresholds
Code to Acquire Initial Population Directly Proportion
End
Pulsation Parameter F3
Margin Parameter F4
Kurtosis Parameter F5
Skewness Parameter F6
acquire ideal output showing Normal-001, Wear-010 and Crack-001. After coded,
Training Sample Data shows in Table 2.
Other detecting data is used as Testing Sample Data. After disposed as above,
Testing Sample Data shows in Table 3.
2. Design of Neural Network
Three layers of BP neural network are designed based on the question, which
owns six nodes in input layer, three nodes in output layer. Based on experience,
there is a proximity relation between the number of hidden layer n2 and the
number of input layer n1 showing below [11].
Research on the Fault Diagnosis Method of Equipment Functionally Significant. . . 347
n2 ¼ 2 n1 þ 1 ð6Þ
10 -2
Mean Squared Error (mse)
10 -4
10 -6
Through diagnosis result comparison, error of the testing results from optimized BP
neural network reduces from 1.0339 to 0.5933. And error of the training samples
from optimized BP neural network reduces from 1.0339 to 0.5933. Optimization
effect is obvious. Through optimized BP neural network to fault diagnosis of
Functionally Significant Instrument, the result fits the real result and veracity can
be confirmed.
Research on the Fault Diagnosis Method of Equipment Functionally Significant. . . 349
6 Conclusion
References
1. Onk S, Maldonado FJ et al (2012) Predictive fault diagnosis system for intelligent and robust
health monitoring. J Aerosp Comput Inf Commun 9(4):125–143
2. Cheng Y, Hou Q, Jiang B (2012) Design and simulation of fault diagnosis based on NUIO/LMI
for satellite attitude control systems. J Syst Eng Electron 23(4):581–587
3. Xu Q-H, Geng S, Shi J (2012) Aero-engine fault diagnosis applying new fast support vector
algorithm. J Aerosp Power 27(7):1605–1611
4. Zhang Y-H, Wang S-H, Han X-H (2013) Research status and prospect of condition-based
maintenance decision-making. J Acad Armored Force Eng 2:6–13
5. GJB1378-92 (1992) Requirements and procedure of developing preventive maintenance
program for material
6. Nazri Mohd N, Abdullah K, Mohammad ZR (2013) A new back-propagation neural network
optimized with cuckoo search algorithm. Lect Notes Comput Sci 7971(1):427–437
7. Zhu D, Shi H (2006) Artificial neural network and application. Press of Science, Beijing
8. Tian L, Luo Y, Wang Y (2013) Prediction model of TIG welding seam size based on BP neural
network optimized by genetic algorithm. J Shanghai Jiaotong Univ 47(11):1691–1701
9. Li J, Chen H, Zhong Z et al (2014) Method for electromagnetic detection satellites scheduling
based on genetic algorithm with alterable penalty coefficient. J Syst Eng Electron 25
(5):822–832
10. Li S (2007) Research into the fault diagnosis expert system based on neural network. Northeast
Normal University
11. Feng S, Wang H, Yu L (2011) 30 cases study of MATLAB metaheuristic. Press of Beijing
University of Aeronautics and Astronautics, Beijing
Gearbox Fault Diagnosis Based on Fast
Empirical Mode Decomposition
and Correlated Kurtosis
Abstract Gear and bearing are widely used in a variety of rotating machinery.
They are considered as the most critical components of gearbox system. Vibration-
based feature extraction is an effective approach to fault diagnosis of gear and
bearing units. Fault diagnosis plays an important role in assuring equipment
availability and reducing operational costs. This paper proposes a new fault diag-
nosis method that synthesizes time synchronous technique, fast empirical mode
decomposition and correlated kurtosis. Our study shows that an improved fault
feature extraction method can be obtained from sampled vibration signals. Energy
of gear wheel rotational frequency and its harmonics are designated as the degra-
dation indicators. The effectiveness of proposed method is verified and validated
using the vibration data from gearbox test rigs and commercial wind turbines.
1 Introduction
EMD algorithm is able to decompose a signal into finite IMFs which are extracted
through an iterative sifting process. Generally, IMFs satisfy two conditions: First, in
the whole time domain, the number of extrema and the number of zero-crossings
are either equal or differ by only one. Second, the mean value of the envelopes
defined by local maxima and local minima is zero. Given a signal y0(t), t2[1, n], the
upper and the lower envelops can be acquired by cubic splines. The average of two
envelops is obtained by subtracting from the original signal. First IMF is obtained
through repeating the sifting process several times. Usually, the first IMF contains
relatively higher frequencies compared to the residual signal. Next, this residual
354 X. Zhang et al.
signal can be continually decomposed into second IMF and new residual signal. If
this process is repeated, a series of IMFs and residual signal r(t) are obtained. The
whole decomposition process can be represented using the following equation:
X
M
yo ðtÞ ¼ cm ðt Þ þ r ðt Þ ð1Þ
m¼1
Compared with laboratory experiments, load profile and rotating speed of rotating
machinery in industrial applications generally are non-stationary. Take the wind
turbine for example. A three percent of speed variation can cause 10 Hz difference
of the bearing fault frequency [16]. If the raw vibration signal is used for EMD
processing, the spectral analysis of selected IMF cannot accurately reveal the true
fault information. This may lead to a miss-detection of faults. Therefore, time
synchronous technique should be used for preprocessing the vibration signal col-
lected from rotating machinery.
Presence of sideband frequencies around gear mesh frequency is the indication
of amplitude modulation. For fault diagnosis, sideband frequencies represent the
abnormality in the gearbox elements due to some faults or irregularity. The space
between the sideband frequencies generally represents the rotating frequency of
gear wheel. Therefore, one object of gear fault diagnosis is to find whether there
exists an intense, equally spaced sideband frequency in frequency spectrum.
Another objective could be the energy of gear mesh frequency and their associated
harmonics. For bearing fault diagnosis, fault frequencies and their harmonics
indicate the bearing fault presence. The object is to find the presence of these
fault frequencies across the entire frequency spectrum. The detail diagnosis frame-
work can be divided into four steps:
Gearbox Fault Diagnosis Based on Fast Empirical Mode Decomposition and. . . 355
P N Q M 2
t¼1 yðt mτÞ
m¼0
Correlated kurtosis of M shift ¼ CK M ðτÞ ¼ P Mþ1 ð2Þ
N 2
t¼1 y ð tÞ
where y(t) is the input signal, τ is the interesting period of the fault, and N is the
number of samples of y(t). If τ ¼ 0 and M ¼ 1, it gives the traditional kurtosis and
can be used to detect the specific periodic impulse signals. For example, if the
desired fault frequency is 50 Hz and the sampling frequency is 10,000 Hz, the value
of τ could be 200 samples.
CK can detect the presence of some specific fault types. This characteristic can
be used to find the best IMF that contains specific fault information and is very
beneficial for degradation tracking when multi-fault exist.
Step 4: The selected IMF can be used for post processing. For gear fault
diagnosis, gear wheel rotational frequency and the associated harmonics in enve-
lope spectrum are important indicators to determine which gear wheel has fault. In
addition, gear wheel rotational frequency energy and their associated harmonics
can be used to track the fault degradation trend. For bearing fault diagnosis, specific
fault frequencies of envelope spectrum are the indicators to identify faulty compo-
nent. The fault diagnosis framework based on Fast EMD and CK is illustrated in
Fig. 1.
4 Experimental Validation
In this section, the proposed method is used to analyze the experimental vibration
signal of a gearbox in a laboratory setting. The setup of the gearbox test rig is
illustrated in Figs. 2 and 3. The test rig contains a 4 kW three phase asynchronous
speed-adjustable motor for driving the gearbox, two-stage gearbox and a magnetic
powder brake. The adjusted load is provided by the powder brake connected to the
gearbox output shaft.
Vibration signals are measured by accelerometer secured by magnetic bases.
The accelerometer is mounted on the top of low shaft speed bearing pedestal of the
tested gearbox and the data are sampled at the frequency of 20 kHz. The tachometer
Fig. 3 (a) Gearbox schematic. (b) Partial gear wheel. (c) Tooth material removal
Gearbox Fault Diagnosis Based on Fast Empirical Mode Decomposition and. . . 357
resolution is 60 impulses per revolution. The teeth number and rotational speed of
high speed shaft are 25 and 13.33 Hz, respectively. Gear wheel on the high speed
shaft meshes with gear wheel on intermediate speed shaft with 50 teeth. Gear wheel
on the low speed shaft has 81 teeth and meshes with gear wheel on intermediate
speed shaft with 19 teeth, which gives 1:8.53 gearbox transmission ratio and
1.56 Hz speed of low speed shaft. Gear wheel on low speed shaft is used for testing.
The presented results refer to nominal low speed shaft torque of 5.54 Nm and the
tests were carried out under constant load.
Distributed faults, as uniform wear around whole gear wheel generally produce
high level side bands and its associated harmonics in the envelope spectrum. The
goal of CK is to detect the periodic impulse with a period of about 25,479 samples
(new sampling frequency after time synchronous processing) divided by gear wheel
rotational frequency. However, the length of time synchronous averaging (TSA)
generally is very short. It can only guarantee that there is one period in TSA data.
Therefore, there is no need to perform the averaging operation. It only needs to
resample the signal according to tachometer signal for assuring equal angle-interval
sampling. That ensures the signal length is long enough to calculate the CK value.
After time synchronous processing, the signal can be decomposed into ten IMFs
using Fast EMD. For illustration purpose, the vibration data of gear wear state 2 is
being analyzed. Figure 4 is the results of ten IMFs, and Fig. 5 is the corresponding
frequency spectrum of FFT.
In this paper, all the computations including the plot of figures are performed in
Matlab environment. Matlab is a very useful tool for scientific calculations in
research. Certainly, other software applications can also be used to implement
these functions also.
5 Discussions
From Fig. 5, we can see that IMF 2 contains more harmonics of low gear wheel
rotational frequency and higher energy. The values of three indicators of these
IMFs are depicted in Figs. 6, 7, and 8. They correspond to correlation coefficient
(CC), kurtosis and correlated kurtosis (CK), respectively. The approximately best
IMF can be determined from these indicators. Root mean square values (RMS) of
original vibration signals prior to EMD processing, and optimal IMF selected by
CK value can be acquired for five wear states of gear. In addition, energy of low
358 X. Zhang et al.
8 0.2
6
4 0.1
IMF6
IMF1
2
0
0
-2 -0.1
-4
-6 -0.2
0 2 4 6 8 10 12 0 2 4 6 8 10 12
Revolutions Revolutions
4 0.15
0.1
2
0.05
IMF2
IMF7
0 0
-0.05
-2
-0.1
-4 -0.15
0 2 4 6 8 10 12 0 2 4 6 8 10 12
Revolutions Revolutions
3 0.05
2
1
IMF3
IMF8
0 0
-1
-2
-3 -0.05
0 2 4 6 8 10 12 0 2 4 6 8 10 12
Revolutions Revolutions
1 0.04
0.5 0.02
IMF4
IMF9
0 0
-0.5 -0.02
-1 -0.04
0 2 4 6 8 10 12 0 2 4 6 8 10 12
Revolutions Revolutions
0.4 0.1
0.2 0.05
IMF10
IMF5
0 0
-0.2 -0.05
-0.4 -0.1
0 2 4 6 8 10 12 0 2 4 6 8 10 12
Revolutions Revolutions
0.06 x 10-3
8
0.05
6
0.04 1×LSS rate
IMF1
IMF6
0.03 4
0.02
2
0.01
0 0
0 5 10 15 20 0 5 10 15 20
Frequency(Hz) Frequency(Hz)
0.07 x 10-3
7
0.06 6
0.05 1×LSS rate 5
IMF2
IMF7
0.04 4
0.03 3
0.02 2
0.01 1
0 0
0 5 10 15 20 0 5 10 15 20
Frequency(Hz) Frequency(Hz)
-3
0.04 x 10
6
1×LSS rate
5
0.03
4
IMF3
IMF8
0.02 3
2
0.01
1
0 0
0 5 10 15 20 0 5 10 15 20
Frequency(Hz) Frequency(Hz)
0.025 x 10-3
1×LSS rate
4
0.02
3
IMF4
IMF9
0.015
0.01 2
0.005 1
0 0
0 5 10 15 20 0 5 10 15 20
Frequency(Hz) Frequency(Hz)
0.012 x 10-3
1×LSS rate 5
0.01
4
0.008
IMF10
IMF5
0.006 3
0.004 2
0.002 1
0 0
0 5 10 15 20 0 5 10 15 20
Frequency(Hz) Frequency(Hz)
Fig. 6 CC values
of ten IMFs
Fig. 8 CK values
of ten IMFs
gear wheel rotational frequency and its second and third harmonics in the envelope
spectrum can also be obtained. Figure 9 represents the normalized RMS and energy
values. It shows that gear wheel rotational frequency energy is the best indicator to
track the gear degradation, because they have systematic increasing trend compared
with RMS values of other two types.
Gearbox Fault Diagnosis Based on Fast Empirical Mode Decomposition and. . . 361
6 Conclusions
This paper proposes a hybrid gear and bearing fault diagnosis method based on three
indicators, namely, time synchronous technique, fast EMD and correlated kurtosis.
To demonstrate the effectiveness in real applications, the method is tested on the gear
wear fault identification of fixed-axis gearbox; and on the bearing faults in drive train
gearbox of wind turbines. Fast EMD algorithm can decompose a long signal into
several IMFs in real time manner, enabling us to obtain high resolution frequency
spectrum. Correlated kurtosis can determine the best IMF used for post processing.
The length of time synchronous averaging is very short, which guarantees a single
period of TSA data. Therefore, the averaging operation is not required; instead it only
resamples the signal by the tachometer through equal angle-interval waveform. This
ensures that the signal is long enough for calculating the CK value. After time
synchronous processing, the signal can be decomposed into ten IMFs using Fast
EMD. Finally, gear and bearing fault data of test rigs and field wind turbines are used
to verify the effectiveness of proposed method. The results show the rotational
frequency energy with first and second harmonic frequency outperforms the root
mean square for degradation tracking. In the future the proposed method will be
extended to industrial applications where machines or systems are subject to random
shocks or vibration signals of certain components are available.
References
1. Huang NE, Shen Z, Long SR, Wu MC, Shih HH, Zheng Q, Yen NC, Tung CC, Liu HH (1998)
The empirical mode decomposition and the Hilbert spectrum for non-linear and non-stationary
time series analysis. Proc R Soc London Ser A 454:903–995
2. Mandic D (2011) Filter bank property of multivariate empirical mode decomposition. IEEE
Trans Signal Process 59:2421–2426
362 X. Zhang et al.
Abstract Bulge deformation is one of the key factors that affect the slab quality in
continuous casting involving heat transfer, phase transition and thermo-mechanical
coupling problem. A systematical investigation into the bulge deformation rules of
the narrow side was provided in the present study by using ABAQUS; And a three-
dimensional thermo-mechanical coupling model has been established with taking
into account the dynamic contact between slab and rolls. The result reveals that the
cumulative bulge deformation without the constraint from the roller supporting is
greater than the recovery count which compensates deformation in the wide side.
The influences of various casting process parameters on bulge deformation of the
slab have been also investigated in this paper. The consequence has been obtained
that bulge deformation in the narrow side increases with the increase of casting
speed and roll pitch; And the method of fixes-gap and variable-diameter was
suggested to reduce the bulge deformation for manufacturing.
1 Introduction
Bulge deformation of the slab affects not only the quality of the billet, but also the
stability of the production. However, continuous casting process involves heat
transfer, phase transition and thermo-mechanical coupling problem. Therefore, the
analysis and measurement of bulge deformation is a very difficult issue to be solved.
In recent years, analytical methods and finite element methods have been generally
applied to investigate bulge deformation of the slab. In the early stages of the bulging
research, analytical methods were initially used due to their convenience. In these
methods, casting slabs were usually simplified to continuous beam models or plate
models according to the theory of bending beam or plate theory [1, 2]. However the
calculation accuracy of analytical methods is relatively low because there are some
differences between the actual conditions and the oversimplified models. Therefore
the finite element methods have been widely applied to study the bulging problems.
Two-dimensional models were firstly adopted by the majority of previous researches
[3, 4]. However, bulge deformation of the narrow side in these models cannot be
investigated. Therefore, some three-dimensional models of bulge deformation have
been developed by some researchers for improving the calculation accuracy.
K. Okamura et al. have applied a three-dimensional elasto-plastic and creep model
to investigate the effect of the narrow face shell on restraining the bulging deflection
[5]. K. Feng et al. have established a three-dimensional coupling model and obtained
the detailed external and internal distributions of stress and strain for slab [6]. Unfor-
tunately, these models were based on the static contacts between the slab and the
rolls. Q. Qin et al. have established a three-dimensional thermo-mechanical coupling
model which was based on the dynamic contact between the slab and rolls [7]. But
this model was merely used to discuss the differences between the 3D and 2D models
in calculating the bulge deformation. The deformation distributions of the narrow
side have not been analyzed in detail.
The objective of this paper is to provide a systematical investigation into the
bulge deformation rules of the narrow side by using ABAQUS, and has established
a three-dimensional thermo-mechanical coupling model with considering the
dynamic contact between slab and rolls. Moreover, the influences of casting
parameters on bulge deformation in the narrow side have also been discussed to
design for manufacturing.
Bulge analysis involves the highly non-linear behaviors including multiple objects,
complex thermo-mechanical coupling problems. Some simplified treatments were
adopted in this model and shown as follows:
• A quarter bulging model was built due to the symmetry of heat transfer and
model structure.
• Both the driving rolls and the driven rolls were assumed to be rigid bodies.
• The pressure was assumed to apply outward at the surface of the quarter bulging
model.
Continuous casting is a high-temperature process, and material properties of the
slab are sensitive to the temperature. Young’s modulus and Poisson’s ratio of Q235
steel were listed in Table 1, and the coefficient of thermal expansion of Q235 steel
is shown in Fig. 1. Moreover, creep behavior of the slab was taken into consider-
ation besides mechanics behavior in this paper. And the time hardening creep
model was adopted to describe the viscoelastic behavior of the slab as Eq. (1) [7]:
Bulge Deformation in the Narrow Side of the Slab During Continuous Casting 365
10
Coefficient of thermal expansion (10-5 °C-1)
-2
-4
0 200 400 600 800 1000 1200 1400 1600
Temperature (°C)
ε_ ¼ CexpðQ=TÞσ n tm ð1Þ
In order to fit the actual situation, boundary conditions and contact definition
were taken into consideration as follows and the three-dimensional dynamic finite
element model involving the slab and the rolls:
• The symmetrical displacement constraint was applied to the symmetrical surface
of the slab. And both ends of the slab were controlled with translate and rotate
constraint.
• The pressure caused by the liquid core was assumed to apply as the uniform
force increasing linearly with distance below the liquid steel meniscus.
366 Q. Qin et al.
• The penalty friction model was applied to the dynamic contact between the slab
and the rolls. The friction coefficient of the driving rolls was considered as 0.3
and for the driven rolls was 0.001 according to the conversion of friction
coefficient of the ball bearing.
• The driving rolls were applied with the angular velocity and the driven rolls were
applied with constraints.
Generally, bulge deformation of the wide side has been widely discussed, but that
of the narrow side has been rarely investigated. Based on the establishment of the
dynamic bulging model, two typical nodes were chosen to analyze the character-
istics of the bulge deformation in the narrow side. As shown in Fig. 2, nodes A1 is
directly over the rolls while nodes B1 is at the center of two successive rolls. With
the results of the bulging simulation, the horizontal displacement curves along with
time of the nodes show the same trend of decline (Fig. 3). Under the pressure caused
by the liquid core, bulge deformation in the narrow side develops more freely
without the constraint from the roller supporting.
The deformation increase in the narrow side is greater than the count of the
recovery which compensates bulge deformation of the wide side. As a result, bulge
deformation in the narrow side is cumulative. Moreover, the horizontal displace-
ment situations of two nodes are on the contrary because they differ half of a roll
pitch. In the first several casting cycles, bulge deformation of the narrow side is in
the unstable state, and then becomes stable with the increase of distance along the
casting direction. The difference between values of maximum and minimum
deformation approaches certain in the steady deformation zone.
Figure 4 shows the difference between values of maximum and minimum
deformation to accurately calculate deformation accumulation of the narrow side.
The deformation cycle and the residual deformation have been marked as T and
0.2
Node A1
Node B1
Horizontal Displacement(mm)
0.0
-0.2
-0.4
-0.6
-0.8
0 8 16 24 32 40 48 56 64 72 80
Time (s)
Fig. 3 Horizontal displacement curves along with time of two nodes
-0.1
-0.2
Horizontal Displacement (mm)
-0.3
∆U
-0.4
-0.5
T
-0.6
-0.7
16 24 32 40 48 56 64 72 80
Time (s)
Fig. 4 Bulge deformation curve along with time in the narrow side
368 Q. Qin et al.
-0.2
-0.3
Horizontal Displacement (mm)
-0.4
-0.5
-0.6
-0.7
-0.8
0.00 0.04 0.08 0.12
Distance from central surface along the thickness (m)
Fig. 5 Bulge deformation of the narrow side along the thickness direction
ΔU, respectively. The absolute value of the slope of the boundary was found to be
jKj ¼ 4.39 103, and then bulge accumulation of the narrow side was calculated
as listed in Table 2.
Bulge deformation of the narrow side along the thickness direction was not
uniform as shown in Fig. 5. Resultly, bulge deformation of the outer node in narrow
side is 0.661 mm when the bulge deformation of the inner node is 0.306 mm, which
accounts for 46.3% of that of the outer side. Figure 6 shows the displacement curves
of the narrow side that are various with the increase of distance along the thickness
direction. However, the relative positions corresponding to the displacement ampli-
tudes of the inner and outer nodes in narrow side are on the contrary, where bulge
deformation of the inner side is minimum when that of the outer side is maximum.
Generally, bulge deformation of the narrow side develops more freely than the wide
sides and is not uniform along the thickness direction.
Bulge Deformation in the Narrow Side of the Slab During Continuous Casting 369
0.2
The central node of the narrow side
The edge node of the narrow side
0.0
Horizontal Displacement (mm)
-0.2
-0.4
-0.6
-0.8
-1.0
0 8 16 24 32 40 48 56 64 72 80
Time (s)
Fig. 6 Relative Positions corresponding to the displacement amplitudes in the narrow side
As mentioned in the introduction, many influence parameters, such as roll pitch, the
ferrostatic pressure, temperature fields of the slab, the thickness of the solidified
shell and casting speed, are involved in the production conditions of continuous
casting. However, casting processing parameters are often coupled with each other.
For instance, the ferrostatic pressure, temperature fields and thickness of the slab
have been found to be affected by casting speed. Therefore, their effects on bulge
deformation could be taken into overall account for the influence of casting speed.
Consequently, the influences of casting speed and roll pitch on bulge deformation
were respectively investigated for manufacturing in this study.
The influence of casting speed was investigated under various casting speeds and
constant roll pitch. The bulge deformation was obtained by using the dynamic
bulging models. Figure 7 shows the bulge deformation of the slab under various
casting speeds. Generally, bulge deformation of the narrow side increases about
0.23 mm with the increase each 0.1 m/min of casting speed. That is mainly due to an
increase in temperature fields of the slab induced by the increase of casting speed,
which consequently leads to the increase of the bulge deformation.
370 Q. Qin et al.
0.60
The method of fixed-diameter and variable-gap
The method of fixed-gap and variable-diameter
0.55
The average deformation (mm)
0.50
0.45
0.40
0.35
0.30
325 350 375 400
Roll pitch (mm)
Similarly, the influence of roll pitch was conducted by using the dynamic
bulging models under various roll pitches and constant casting speed. In this
study, two methods were adopted to change the size of roll pitch. One method is
fixed-gap variable-diameter, where roll gap is kept a constant value of L0 ¼ 70 mm
and roll diameter is variable. Constrastly, another method is fixed-diameter and
variable-gap that roll diameter is kept a constant value of Φ0 ¼ 230 mm and roll gap
is variable. Bulge deformation curves of the narrow side under various roll pitches
are shown in Fig. 8. By using two methods, bulge deformation of the narrow side
respectively increases about 0.037 mm and 0.043 mm when roll pitch increases
each 25 mm. The increase of roll pitch directly induces an increase in creep time of
the slab, which consequently results in the increase of the bulge deformation.
Additionally, bulge deformation simulated with the method of fixes-gap and
variable-diameter is smaller than that with the method of fixed-diameter and
variable-gap. That is a result of the increase of the contact area between the slab
and rolls caused by the larger roll diameter of fixed-gap and variable-diameter
method, which then indirectly increases the stiffness of the slab.
In generally, bulge deformation of the narrow side increases with growth of
casting speed and roll pitch. In addition, the method of fixes-gap and variable-
diameter is better than the method of fixed-diameter and variable-gap to reduce the
bulge deformation.
Bulge Deformation in the Narrow Side of the Slab During Continuous Casting 371
3.0
2.5
The average deformation (mm)
2.0
1.5
1.0
0.5
0.0
1.6 1.8 2.0
Casting speed (m/min)
Fig. 8 Bulge deformations under various casting speeds
5 Summary
References
1. Jiquan S, Yipin S, Xingzhon Z (1996) Analysis of bulging deformation and stress in continuous
cast slabs. J Iron Steel Res Int 8:11–15
2. Yoshii A, Kihar S (1986) Analysis of bulging in continuously cast slabs by bending theory of
continuous beam. Trans Iron Steel Inst Jpn 26:891–894
3. Janik M, Dyja H, Berski S, Banaszek G (2004) Two-dimensional thermomechanical analysis of
continuous casting process. J Mater Process Technol 153:578–582
4. Koric S, Thomas BG (2007) Thermo-mechanical models of steel solidification based on two
elastic visco-plastic constitutive laws. J Mater Process Technol 197:408–418
5. Okamura K, Kawashima H (1989) Three-dimensional elastoplastic and creep analysis of
bulging in continuously cast slabs. ISIJ Int 29:666–672
6. Feng K, Chen D, Xu C et al (2004) Effect of main thermo-physical parameters of steel Q235 on
accuracy of casting transport model. Spec Steel 25:28–31
7. Qin Q, Shang S, Wu DP, Zang Y (2014) Comparative analysis of bulge deformation between
2D and 3D finite element models. Mech Eng 154:14–17