These de Doctorat: Universit e de Tunis Institut Sup Erieur de Gestion
These de Doctorat: Universit e de Tunis Institut Sup Erieur de Gestion
These de Doctorat: Universit e de Tunis Institut Sup Erieur de Gestion
THESE DE DOCTORAT
en vue de l’obtention du titre de docteur en
NOM DE LA SPECIALITE
Option
Nom de l’Option
TITRE
Sous TITRE
NOM PRENOM
Laboratoire/Unité de recherche:
Acknowledgments
Such descriptions give rise to speculation: are judgment-laden adjectives like “in-
tense” and “hardheaded” simply a way to describe traits that today might be cate-
gorized under juvenile behavioral disorder? A December, 2001, Wired magazine
article titled “The Geek Syndrome” paints the portrait of several scientifically
gifted children diagnosed with high-functioning autism or Asperger Syndrome.
...
Dan Chess, a fellow classmate in the Columbia Science Honors Program, re-
calls Richard Stallman seeming a bit weird even among the students who shared
a similar lust for math and science. “We were all geeks and nerds, but he was
unusually poorly adjusted,” recalls Chess, now a mathematics professor at Hunter
College. “He was also smart as shit. I’ve known a lot of smart people, but I think
he was the smartest person I’ve ever known.”
...
At first, Stallman viewed these notices with alarm. Rare was the software
program that didn’t borrow source code from past programs, and yet, with a single
stroke of the president’s pen, Congress had given programmers and companies
the power to assert individual authorship over communally built programs. It also
injected a dose of formality into what had otherwise been an informal system.
i
Contents
Introduction 1
1 First Chapter 5
1.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
1.3 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
2 Second Chapter 10
2.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
2.3 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
ii
CONTENTS iii
Part II : Contributions 14
3 Chapter 3 15
3.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
3.3 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
4 Chapter 4 18
4.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
4.3 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
5 Chapter 5 20
5.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20
5.3 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21
6 Chapter 6 22
6.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
6.3 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23
7 Chapter 7 24
7.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24
7.2 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25
7.3 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25
Conclusion 26
iv CONTENTS
Appendix 27
A Appendix 1 28
A.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28
A.2 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28
References 29
List of Figures
v
List of Tables
vi
List of Algorithms
vii
Introduction
At the heart of the human immune system is its ability to discriminate between
the body’s own cells, the self elements, and the foreign cells, non-self elements,
by a process called “self-nonself discrimination”. This fundamental process is
performed by several AIS algorithms. However, it was noted that these algo-
rithms cannot produce the same high performance as the human immune system.
Hence, in (Matzinger, 2001), a relatively newer immunological discovery using
more rigorous and up-to-date immunology, known as the “Danger Theory” (DT),
has been proposed. The DT states that the recognition of a pathogen is based on
environmental context rather than the simple self-nonself principle.
...
1
2 Introduction
...
• Contribution 1.
• ...
• Contribution n.
Thesis Outline
The first part, Theoretical Aspects, is composed of two main Chapters which
are the following:
• Chapter 1: . . .
• Chapter 2: . . .
The second part of this Thesis presents our contributions. We can decompose
this part into three main blocks. The first block is composed of two Chapters
dealing with new developed danger theory approaches based on robust data pre-
processing modules:
• Chapter 3: . . .
• Chapter 4: . . .
The second block is composed of two Chapters and deals with the develop-
ment of new DCAs handling the problem of the algorithm sensitivity to the input
class data order. We hypothesize that the causes of such sensitivity may be either
tight to the DCA context assessment phase, first hypothesis, or to the DCA detec-
tion phase, second hypothesis. On this light, we propose to investigate the first
hypothesis in Chapter 5 and the second hypothesis in Chapter 6.
• Chapter 5: . . .
• Chapter 6: . . .
Introduction 3
Lastly, the third block is seen as the fruit of all the previously studied parts. This
block deals with the development of a more general DCA taking all the studied
facts into account. The final fruit is a more robust danger theory classifier based
on dendritic cells within an imprecise framework, and is presented in Chapter 7.
• Chapter 7: . . .
Finally, our dissertation ends with a conclusion and an Appendix. The conclu-
sion summarizes all the work presented in this report and proposes further works
to be done with the dendritic cell algorithm.
1.1 Introduction
In this Chapter, first, we start by presenting the artificial immune system field.
Next, we detail the basics of the danger theory. Finally in Section 1.4, we describe
the dendritic cell algorithm including its biological principals.
As previously stated, the most prominent players of the DT are the dendritic cells
(DCs). An inspiration from the DCs behavior led to the development of an im-
mune algorithm termed the dendritic cell algorithm (DCA) (Greensmith & Aick-
elin, 2005). Before describing the functioning of the algorithm, we give a general
overview of the biological principles used by the DCA.
5
6 Chapter 1 : First Chapter
• Mature DCs: For an iDC to become a “mature DC” (mDC), the iDC must
be exposed to a greater quantity of either PAMPs or DSs than SSs. Sufficient
exposure to PAMPs and DSs causes the DC to cease antigen collection and
migrate from the tissue to the lymph node. Most importantly, mDCs produce
an inflammatory cytokine called “interleukin-12” which stimulates T-cell ac-
tivation in order to be reactive to antigen presentation. Additionally, mDCs
produce costimulatory molecules (CSMs) which are known to facilitate the
antigen presenting process.
Swapping from one DC state to another is dependent upon the receipt of dif-
ferent signals throughout the initial state (iDC). This is shown in Figure 1.1. The
migration to the mature state or to the semi-mature state depends on the concen-
tration of the input signals received from the environment. Immunologically, if the
concentration of SSs is greater than the other categories of signals then DCs mi-
Section 1.2 – The Dendritic Cell Algorithm 7
grate to the semi-mature state. However, if the concentration of PAMPs and DSs
is greater than the concentration of SSs then DCs migrate to the mature context.
For the ease of analysis, the algorithm is divided into the following four main
phases:
...
8 Chapter 1 : First Chapter
...
...
...
1.3 Conclusion
In this Chapter, we have presented the basic notions of the danger theory and
basically the DCA. We have, also, discussed the major DCA works where we have
clarified the strengths and the weaknesses of the algorithm. In the next Chapter,
Section 1.3 – Conclusion 9
we will present the basics of rough set and fuzzy rough set theories as main tools
to be used for the data pre-processing phase of our new DCA proposed versions.
Chapter 2
Second Chapter
2.1 Introduction
Data reduction is a main point of interest across a wide variety of fields. In fact,
focusing on this step is crucial as it often presents a source of significant data
loss. Many techniques were proposed in literature to achieve the task of data
reduction. However, most of them tend to destroy the underlying semantics of the
features after reduction or require additional information about the given data set
for thresholding. Thus, it seems necessary to think about a technique that can on
the one hand reduce data dimensionality using information contained within the
data set and on the other hand capable of preserving the meaning of the features.
Rough Set Theory (RST) and Fuzzy-Rough Set Theory (FRST) can be used as
such tools to discover data dependencies and to reduce the number of attributes
contained in a data set using the data alone, requiring no additional information
(Jensen, 2005).
This Chapter focuses on introducing RST and FRST for feature selection. In
the next Section, a brief refresh on some tools for data reduction is proposed. In
Section 2.3, the basic concepts of RST are highlighted and the fundamentals of
FRST are presented next in Section 2.4.
10
Section 2.2 – Rough Set Based Approach for Feature Selection 11
In this Section, rudiments of the RST will be outlined and basic concepts of the
theory will be illustrated by a simple tutorial example.
Data are represented as a table where each row represents an object and where
each column represents an attribute that can be measured for each object. Such
table is called an “Information System” (IS) or an “Information Table” (IT). To
fit this definition to our research field, an information table can be seen as a rep-
resentation of antigens which are defined via a set of attributes. Formally, an in-
formation system can be defined as a pair IS = (U, A) where U = {x1 , x2 , . . . , xn }
is a non-empty, finite set of objects called the “universe” and A = {a1 , a2 , . . . , ak }
is a non-empty, finite set of “condition” attributes. In supervised learning, a spe-
cial case of the defined information table is considered, called a “Decision Table”
(DT) or a “Decision System” (DS). A DT is an information system of the form
IS = (U, A ∪ {d}), where d < A is a distinguished attribute called “decision”. The
value set of d, called θ = {d1 , d2 , . . . , d s }.
Example 2.1 Using the terminology of RST, the data set presented in Table 2.1
can be considered as an information system IS = (U, A) consisting of 4 condi-
tional features (a, b, c, d) and 8 objects. To illustrate an example of a decision
system, Table 2.2 is used. It consists of 4 conditional features (a, b, c, d), 1 deci-
sion feature (e) and 8 objects.
The indiscernibility relation is the mathematical basis of rough set theory. Every
object of the universe is described by certain amount of information expressed
by means of some attributes used for object description. Objects characterized
by the same information are indiscernible in view of the available information
about them. For every set of attributes P ⊂ A, an indiscernibility relation, denoted
by IND(P) or U/P, is defined in the following way: two objects, xi and x j , are
12 Chapter 2 : Second Chapter
x∈U a b c d x∈U a b c d e
x1 1 0 2 2 x1 1 0 2 2 0
x2 0 1 1 1 x2 0 1 1 1 2
x3 2 0 0 1 x3 2 0 0 1 1
x4 1 1 0 2 x4 1 1 0 2 2
x5 1 0 2 0 x5 1 0 2 0 1
x6 2 2 0 1 x6 2 2 0 1 1
x7 2 1 1 1 x7 2 1 1 1 2
x8 0 1 1 0 x8 0 1 1 0 1
Example 2.2 In order to illustrate how a decision table from Table 2.2 defines
an indiscernibility relation, we consider the following three non-empty subsets of
Section 2.3 – Conclusion 13
the conditional attributes: {a}, {b, c} and {a, b, c}. The relation IND may define
three partitions of the universe.
If we take into consideration the set a, the objects x1 , x4 and x5 belong to the
same equivalence class; they are indiscernible.
2.3 Conclusion
In this Chapter, both rough sets and fuzzy-rough sets for feature selection were
introduced. The application of these theories form the main contribution of this
Thesis. New hybridized models will be presented in the second part of this disser-
tation where the mentioned theories are combined with the DCA. This is to cover
the DCA limitations which are linked to its data pre-processing step.
Part II
Contributions
3.1 Introduction
In this Chapter, we aim to investigate the DCA data pre-processing phase includ-
ing the step of feature selection/reduction and the step of signal categorization.
We aim to detect and to handle the shortcomings of these two steps. To do so,
we suggest to apply Rough Set Theory (RST) as a feature selection and signal
categorization technique in the DCA data pre-processing phase. This will result
on different DCA new automated approaches that will be compared to each other
while pointing out their characteristics.
15
16 Chapter 3 : Chapter 3
• H2: The use of a rough set heuristic can keep the DCA main characteristic
which is its lightweight in terms of processing time. This will be achieved by
finding a trade-off between generating satisfactory classification results and
preserving the lightweight of the algorithm.
The three developed rough DCAs, including RST-DCA, RC-DCA and QR-
DCA, function under four levels of abstraction as shown in Figure 3.1. We will
mainly discuss the algorithms data pre-processing phase as the rest of the algo-
rithms steps are performed the same as the standard DCA and as explained, pre-
viously, in Chapter 1.
3.3 Conclusion
In this Chapter, we have elucidated how rough set theory could be used to select
the right features and to categorize each selected attribute to its signal category.
We have developed three rough dendritic cell algorithms, compared them and
noted the appropriate observations while testing two crucial hypotheses.
Section 3.3 – Conclusion 17
Despite of the noticed advantages of our proposed rough set methods, they
have some limitations when dealing with real-valued attributes. This prompted
our research into the use of fuzzy-rough sets for feature selection and signal cate-
gorization. This will be dealt with in the next Chapter.
Chapter 4
Chapter 4
4.1 Introduction
In this Chapter, we still focus on the DCA data pre-processing phase. We propose
novel approaches to more improve the standard DCA data pre-processing task.
This is achieved through the use of Fuzzy Rough Set Theory (FRST) (Dubois &
Prade, 1992). The proposed algorithms are seen as extensions to the rough set
algorithms which were previously presented in Chapter 3.
The rest of this Chapter is organized as follows: Section 4.2 deals with the
motivation behind the use of FRST for data pre-processing. Our developed DCA
variants are explained in details in Sections 4.3, 4.4 and 4.5. A comparison of our
proposed algorithms is presented in Section 4.6.
....
18
Section 4.3 – Conclusion 19
4.3 Conclusion
Throughout this Chapter, we focused our work on using fuzzy rough set theory for
the DCA data pre-processing phase. We have developed three fuzzy-rough DCAs
based on different methodologies for the feature selection and signal categoriza-
tion processes. These methods were compared and the appropriate observations
were noted.
5.1 Introduction
In this Chapter, we will investigate and try to solve another limitation of the DCA
which is its sensitivity to the input class data order. As previously mentioned, the
standard DCA had to be applied to ordered data sets where all class 1 are fol-
lowed by all class 2. This is to ensure high and satisfactory classification results.
While investigating this DCA limitation, we have noticed that this problem could
be linked to the DCA context assessment phase. Therefore, we propose in this
Chapter to develop an automated fuzzy dendritic cell algorithm characterized by
its stability as a binary classifier.
A study focusing on the DCA behavior stated that the algorithm gives satisfactory
and interesting classification results when it is, only, applied to databases having
ordered classes (Greensmith & Aickelin, 2008). In this Chapter, we try to study
20
Section 5.3 – Conclusion 21
carefully the DCA algorithmic phases while trying to figure out the reasons of
such a restriction.
5.3 Conclusion
In this Chapter, we have analyzed the behavior of the standard DCA while trying
to overcome its shortcoming as it is sensitive to the class data order. This inves-
tigation led us to the development of the FCDCM algorithm. FCDCM presents a
fuzzy version of the DCA and it aims at coping with the crisp separation between
the two DCs contexts. In order to fix the parameters of the algorithm, we have used
various fuzzy clustering techniques, compared them and chosen the most appro-
priate one. The experimental results demonstrate that our proposed method with
the Gustafson-Kessel algorithm (FCDCMGK ) achieves better results compared to
other methods.
In the next Chapter, we will try to investigate more the causes of the dendritic
cell algorithm sensitivity to the input class data order. We will focus, mainly, on
the dendritic cell algorithm detection phase while suggesting new automated and
more adequate solutions for the DCA.
Chapter 6
Chapter 6
6.1 Introduction
Still focusing on the DCA sensitivity to the input class data order, we have noticed
that this limitation cannot only be linked to the DCA context assessment phase,
as presented in the previous Chapter, but also to the DCA detection phase. We
hypothesize that there is a second possible cause related to the DCA sensitivity.
It is possible that the induced signal base generated by the DCA detection phase
contains disagreeable objects such as noisy, incoherent or redundant instances.
Such objects may influence the DCA behavior and, thus, maintaining the signal
base looks essential. To achieve this, we propose in this Chapter to expand a
second fuzzy DCA version based on a maintenance technique.
The quality of the database is very important to generate accurate results and to
have the possibility to learn models from the presented data. To achieve this,
the disagreeable objects - especially noisy and redundant instances which affect
22
Section 6.3 – Conclusion 23
negatively the quality of the results - have to be eliminated from the data set. To
address this situation, data maintenance is a feasible way. The goal of database
maintenance methods is to select the most representative information from a given
data set.
Many works dealing with database maintenance have been proposed in liter-
ature (Leake & Wilson, 2001). Most of them are based on updating a database
by adding or deleting instances to optimize and reduce the initial database. These
policies include different operations such as: the outdated, redundant or incon-
sistent instances may be deleted, groups of objects may be merged to eliminate
redundancy and improve reasoning power, objects may be re-described to repair
incoherencies, signs of corruption in the database have to be checked and any ab-
normalities in the database which might signal a problem has to be controlled. A
database which is not maintained can become sluggish and without accurate data
users will make uninformed decisions. The database maintenance policies may
be categorized into two main heads; the selective reduction approaches and the
deletion reduction approaches.
6.3 Conclusion
7.1 Introduction
24
Section 7.2 – Summary 25
7.2 Summary
7.3 Conclusion
In this Chapter, we have proposed a hybrid automated and maintained fuzzy den-
dritic cell immune classifier. COID-FLA-FCDCMGK can be seen as the output of
several investigations conducted on the standard DCA version. A set of hypothe-
ses has been checked and based on the results drawn for the six hypotheses, the
COID-FLA-FCDCMGK was developed. The worthy characteristics of the latter
algorithm were highlighted in this Chapter as well.
Conclusion
More precisely, we have tended to solve two main limitations of the standard
version of the dendritic cell algorithm. . . .
Secondly, . . .
Thirdly, . . .
Finally, . . . .
26
Appendix
A.1 Introduction
Fuzzy set theory was introduced, in 1965, by Zadeh (Zadeh, 1965). It is consid-
ered as a useful theory for modeling and reasoning with imprecise knowledge.
Fuzzy set theory is a mathematical theory where the fuzziness is the ambigu-
ity that can be found in the definition of a concept or the meaning of a word
(Zimmermann, 1996). Imprecision in expressions like “low frequency”, “high de-
mand” or “small number” can be called fuzziness. In this Appendix, the basics of
fuzzy sets will be introduced.
In Section A.2, the basics of fuzzy set theory will be given. In Section A.3, the
main notions of the fuzzy set membership functions will be introduced while in
Section A.4, the fundamental operations of fuzzy sets will be highlighted. Finally,
in Section A.5, the process of fuzzy logic will be described.
A.2 Conclusion
In this Appendix, we have elucidated the basics of fuzzy set theory which is a
generalization of the classical set theory. Offering a natural model, fuzzy sets are
used to handle imprecise information.
28
References
Chelly, Z., & Elouedi, Z. (2012a). Rc-dca:a new feature selection and signal
categorization technique for the dendritic cell algorithm based on rough set
theory. Proceedings of the 11th International Conference of Artificial Im-
mune Systems, Springer, ICARIS’2012, Lecture Notes in Computer Science,
7597, 152-165.
Chelly, Z., & Elouedi, Z. (2012b). Rst-dca: A dendritic cell algorithm based
on rough set theory. Proceedings of the 19th International Conference on
Neural Information Processing, Springer, ICONIP’2012, Lecture Notes in
Computer Science, 7665, 480-487.
Dubois, D., & Prade, H. (1992). Putting rough sets and fuzzy sets together.
Intelligent decision support : Handbook of applications and advances of
the rough sets theory, Springer, 11, 203-232.
Eiben, A., & Smith., J. (2007). Introduction to evolutionary computing. Natural
Computing.
Greensmith, J., & Aickelin, U. (2005). Introducing dendritic cells as a
novel immune-inspired algorithm for anomaly detection. Proceedings of
the 4th International Conference on Artificial Immune Systems, Springer,
ICARIS’2005, Lecture Notes in Computer Science, 3627, 153-167.
Greensmith, J., & Aickelin, U. (2008). The deterministic dendritic cell algorithm.
Proceedings of the 7th International Conference on Artificial Immune Sys-
tems, Springer, ICARIS’2008, Lecture Notes in Computer Science, 5132,
291-302.
Jensen, R. (2005). Combining rough and fuzzy sets for feature selection. Doctoral
Dissertation, School of Informatics University of Edinburgh.
Leake, D., & Wilson, D. (2001). Maintaining case-based reasoners: Dimensions
29
30 References