09 Chapter4

Download as pdf or txt
Download as pdf or txt
You are on page 1of 25

83

CHAPTER 4

TEXTURE FEATURE EXTRACTION

This chapter deals with various feature extraction technique based


on spatial, transform, edge and boundary, color, shape and texture features. A
brief introduction to these texture features is given first before describing the
gray level co-occurrence matrix based feature extraction technique.

4.1 INTRODUCTION

Image analysis involves investigation of the image data for a


specific application. Normally, the raw data of a set of images is analyzed to
gain insight into what is happening with the images and how they can be used
to extract desired information. In image processing and pattern recognition,
feature extraction is an important step, which is a special form of
dimensionality reduction. When the input data is too large to be processed and
suspected to be redundant then the data is transformed into a reduced set of
feature representations. The process of transforming the input data into a set
of features is called feature extraction. Features often contain information
relative to colour, shape, texture or context.

4.2 TYPES OF FEATURE EXTRACTION

Many techniques have been used to extract features from images.


Some of the commonly used methods are as follows:
84

Spatial features

Transform features

Edge and boundary features

Colour features

Shape features

Texture features

4.2.1 Spatial Features

Spatial features of an object are characterized by its gray level,


amplitude and spatial distribution. Amplitude is one of the simplest and most
important features of the object. In X-ray images, the amplitude represents the
absorption characteristics of the body masses and enables discrimination of
bones from tissues.

4.2.1.1 Histogram features

The histogram of an image refers to intensity values of pixels. The


histogram shows the number of pixels in an image at each intensity value.
Figure 4.1 shows the histogram of an image and it shows the distribution of
pixels among those grayscale values. The 8-bit gray scale image is having 256
possible intensity values. A narrow histogram indicates the low contrast
region. Some of the common histogram features are mean, variance, energy,
skewness, median and kurtosis are discussed by Myint (2001).
85

Intensity

Figure 4.1 Histogram of an image

4.2.2 Transform Features

Generally the transformation of an image provides the frequency


domain information of the data. The transform features of an image are
extracted using zonal filtering. This is also called as feature mask, feature
mask being a slit or an aperture. The high frequency components are
commonly used for boundary and edge detection. The angular slits can be
used for orientation detection. Transform feature extraction is also important
when the input data originates in the transform coordinate.

4.2.3 Edge and Boundary Features

Asner and Heidebrecht (2002) discussed edge detection is one of the


most difficult tasks hence it is a fundamental problem in image processing.
Edges in images are areas with strong intensity contrast and a jump in
intensity from one pixel to the next can create major variation in the picture
quality. Edge detection of an image significantly reduces the amount of data
86

and filters out unimportant information, while preserving the important


properties of an image. Edges are scale-dependent and an edge may contain
other edges, but at a certain scale, an edge still has no width. If the edges in an
image are identified accurately, all the objects are located and their basic
properties such as area, perimeter and shape can be measured easily.
Therefore edges are used for boundary estimation and segmentation in the
scene.

4.2.3.1 Sobel technique

Sobel edge detection technique consists of a pair of 3 3 convolution


kernels. One kernel is simply the other rotated by 90° as shown in Figure 4.2.
These kernels are designed to respond maximally to edges running vertically
and horizontally relative to the pixel grid of the image, one kernel for each of
the two perpendicular orientations. The kernels can be applied separately to
the input image, to produce separate measurements of the gradient component
in each orientation. These can then be combined together to find the absolute
magnitude of the gradient at each point and the orientation of that gradient.

-1 0 +1 +1 +2 +1

-2 0 +2 0 0 0

-1 0 +1 -1 -2 -1

Gx Gy

Figure 4.2 Masks used for Sobel operator


87

4.2.3.2 Robert technique

The Robert cross operator performs a simple, quick to compute, 2-D


spatial gradient measurement on an image. Pixel values at each point in the
output represent the estimated absolute magnitude of the spatial gradient of
the input image at that point. The operator consists of a pair of
2 2 convolution kernels as shown in Figure 4.5. One kernel is simply the
other rotated by 90°. This is very similar to the Sobel operator.

+1 0 0 +1

0 -1 -1 0

Gx Gy

Figure 4.3 Masks used for Robert operator

4.2.3.3 Prewitt technique

Prewitt operator is similar to the Sobel operator and is used for


detecting vertical and horizontal edges in images.

-1 0 +1 +1 0 -1

-1 0 +1 +1 0 -1

-1 0 +1 +1 0 -1

Gx Gy

Figure 4.4 Masks for the Prewitt gradient edge detector


88

The Prewitt operator measures two components. The vertical edge


component is calculated with kernel G x and the horizontal edge component is
calculated with kernel G y as shown in Figure 4.4. | G x | | G y | gives an

indication of the intensity of the gradient in the current pixel.

4.2.3.4 Canny technique

The Canny edge detection algorithm is known popularly as the


optimal edge detector. The Canny algorithm uses an optimal edge detector
based on a set of criteria which include finding the most edges by minimizing
the error rate, marking edges as closely as possible to the actual edges to
maximize localization, and marking edges only once when a single edge
exists for minimal response. According to Canny, the optimal filter that meets
all three criteria that can be efficiently approximated using the first derivative
of a Gaussian function. The first stage involves smoothing the image by
convolving with a Gaussian filter. This is followed by finding the gradient of
the image by feeding the smoothed image through a convolution operation
with the derivative of the Gaussian in both the vertical and horizontal
directions. This process alleviates problems associated with edge
discontinuities by identifying strong edges, and preserving the relevant weak
edges, in addition to maintaining some level of noise suppression.

Figure 4.5 Input landsat image


89

Figure 4.6 Output of the edge detection techniques

Finally, hysteresis is used as a means of eliminating streaking.


Streaking is the breaking up of an edge contour caused by the operator output
fluctuating above and below the threshold. Figure 4.6 shows the output of the
different edge detection technique of given input image as shown in
Figure 4.5.

4.2.4 Colour Features

Colour is a visual attribute of object things that results from the light
emitted or transmitted or reflected. From a mathematical viewpoint, the
colour signal is an extension from scalar-signals to vector-signals. Colour
features can be derived from a histogram of the image. The weakness of
colour histogram is that the colour histogram of two different things with the
same colour can be equal. Platt and Goetz (2004) discussed colour features
are still useful for many biomedical image processing applications such as
90

cell classification, cancer cell detection and content-based image retrieval


(CBIR) systems.

In CBIR, every image added to the collection is analyzed to


compute a colour histogram. At search time, the user can either specify the
desired proportion of each colour or submit an example image from which a
colour histogram is calculated. Either way, the matching process then
retrieves those images whose colour histograms match those of the query
most closely.

4.2.5 Shape Features

The shape of an object refers to its physical structure and profile.


Shape features are mostly used for finding and matching shapes, recognizing
objects or making measurement of shapes. Moment, perimeter, area and
orientation are some of the characteristics used for shape feature extraction
technique. The shape of an object is determined by its external boundary
abstracting from other properties such as colour, content and material
composition, as well as from the object's other spatial properties.

4.2.6 Texture Features

Guiying Li (2012) defined texture is a repeated pattern of


information or arrangement of the structure with regular intervals. In a general
sense, texture refers to surface characteristics and appearance of an object
given by the size, shape, density, arrangement, proportion of its elementary
parts. A basic stage to collect such features through texture analysis process is
called as texture feature extraction. Due to the signification of texture
information, texture feature extraction is a key function in various image
processing applications like remote sensing, medical imaging and content-
based image retrieval.
91

There are four major application domains related to texture analysis


namely texture classification, segmentation, synthesis and shape from texture.
Texture classification produces a classified output of the input image where
each texture region is identified with the texture class it belongs.

Texture segmentation makes a partition of an image into a set of


disjoint regions based on texture properties, so that each region is
homogeneous with respect to certain texture characteristics.

Texture synthesis is a common technique to create large textures


from usually small texture samples, for the use of texture mapping in surface
or scene rendering applications.

The shape from texture reconstructs three dimensional surface


geometry from texture information. For all these techniques, texture
extraction is an inevitable stage. A typical process of texture analysis is
shown in Figure 4.7.
Input Image

Pre-processing

Feature extraction

Segmentation, Classification,
Synthesis, Shape from texture

Post-processing

Figure 4.7 Various image analysis steps


92

4.3 TEXTURE FEATURE EXTRACTION

Neville et al (2003) discussed texture features can be extracted using


several methods such as statistical, structural, model-based and transform
information.

4.3.1 Structural based Feature Extraction

Structural approaches represent texture by well defined primitives


and a hierarchy of spatial arrangements of those primitives. The description of
the texture needs the primitive definition. The advantage of the structural
method based feature extraction is that it provides a good symbolic
description of the image; however, this feature is more useful for image
synthesis than analysis tasks. This method is not appropriate for natural
textures because of the variability of micro-texture and macro-texture.

4.3.2 Statistical based Feature Extraction

Statistical methods characterize the texture indirectly according to


the non-deterministic properties that manage the relationships between the
gray levels of an image. Statistical methods are used to analyze the spatial
distribution of gray values by computing local features at each point in the
image and deriving a set of statistics from the distributions of the local
features. The statistical methods can be classified into first order (one pixel),
second order (pair of pixels) and higher order (three or more pixels) statistics.
The first order statistics estimate properties (e.g. average and variance) of
individual pixel values by waiving the spatial interaction between image
pixels. The second order and higher order statistics estimate properties of two
or more pixel values occurring at specific locations relative to each other. The
most popular second order statistical features for texture analysis are derived
93

from the co-occurrence matrix. Statistical based texture features will be


discussed in section 4.4.

4.3.3 Model based Feature Extraction

Model based texture analysis such as fractal model and Markov


model are based on the structure of an image that can be used for describing
texture and synthesizing it. These methods describe an image as a probability
model or as a linear combination of a set of basic functions. The Fractal
model is useful for modeling certain natural textures that have a statistical
quality of roughness at different scales and self similarity, and also for texture
analysis and discrimination.

There are different types of models based feature extraction


technique depending on the neighbourhood system and noise sources. The
different types are one-dimensional time-series models, Auto Regressive
(AR), Moving Average (MA) and Auto Regressive Moving Average
(ARMA). Random field models analyze spatial variations in two dimensions.
Global random field models treat the entire image as a realization of a random
field, and local random field models assume relationships of intensities in
small neighbourhoods. Widely used class of local random field models are
Markov models, where the conditional probability of the intensity of a given
pixel depends only on the intensities of the pixels in its neighbourhood.

4.3.4 Transform based Feature Extraction

Transform methods, such as Fourier, Gabor and wavelet transforms


represent an image in space whose co-ordinate system has an interpretation
that is closely related to the characteristics of a texture. Methods based on
Fourier transforms have a weakness in a spatial localization so these do not
perform well. Gabor filters provide means for better spatial localization but
94

their usefulness is limited in practice because there is usually no single filter


resolution where one can localize a spatial structure in natural textures. These
methods involve transforming original images by using filters and calculating
the energy of the transformed images. These are based on the process of the
whole image that is not good for some applications which are based on one
part of the input image.

4.4 STATISTICAL BASED FEATURES

The three different types of statistical based features are first order
statistics, second order statistics and higher order statistics as shown in Figure 4.8.

Statistical based features

First order Second order Higher order


Statistics Statistics Statistics

Figure 4.8 Statistical based features

4.4.1 First Order Histogram based Features

First Order histogram provides different statistical properties such as


four statistical moments of the intensity histogram of an image. These depend
only on individual pixel values and not on the interaction or co-occurrence of
neighbouring pixel values. The four first order histogram statistics are mean,
variance, skewness and kurtosis.
95

A histogram h for a gray scale image I with intensity values in the


range I ( x, y ) 0, K 1 would contain exactly K entries, where for a typical 8-bit

grayscale image, K 28 256 . Each individual histogram entry is defined as,

h(i ) = the number of pixels in I with the intensity value I for all

0 i K . The Equation (4.1) defines the histogram as,

h( i ) cardinality ( x, y ) | I ( x, y ) i (4.1)

where, cardinality denotes the number of elements in a set. The standard


deviation, and skewness of the intensity histogram are defined in
Equation (4.2) and (4.3).

( I ( x , y ) m) 2
(4.2)
N

( I ( x , y ) m) 3
skewness (4.3)
N 3

4.4.2 Second Order Gray Level Co-occurrence Matrix Features

Some previous research works compared texture analysis methods;


Dulyakarn et al. (2000) compared each texture image from GLCM and
Fourier spectra, in the classification. Maillard (2003) performed comparison
works bewteen GLCM, semi-variogram, and Fourier spectra at the same
purpose. Bharati et al. (2004) studied comparison work of GLCM, wavelet
texture analysis, and multivariate statistical analysis based on PCA (Principle
Component Analysis). In those works, GLCM is suggested as the effective
texture analysis schemes. Monika Sharma et al (2012) discussed GLCM is
applicable for different texture feature analysis.
96

The GLCM is a well-established statistical device for extracting


second order texture information from images. A GLCM is a matrix where
the number of rows and columns is equal to the number of distinct gray levels
or pixel values in the image of that surface. GLCM is a matrix that describes
the frequency of one gray level appearing in a specified spatial linear
relationship with another gray level within the area of investigation. Given an
image, each with an intensity, the GLCM is a tabulation of how often
different combinations of gray levels co-occur in an image or image section.

Texture feature calculations use the contents of the GLCM to give a


measure of the variation in intensity at the pixel of interest. Typically, the co-
occurrence matrix is computed based on two parameters, which are the
relative distance between the pixel pair d measured in pixel number and their
relative orientation . Normally, is quantized in four directions (e.g., 0º,
45 º, 90 º and 135 º), even though various other combinations could be
possible.

GLCM has fourteen features but between them most useful features
are: angular second moment (ASM), contrast, correlation, inverse difference
moment, sum entropy and information measures of correlation. These features
are thoroughly promising.

4.4.3 Gray Level Run Length Matrix Features

Petrou et al (2006) defined gray level run length matrix (GLRLM) is


the number of runs with pixels of gray level i and run length j for a given
direction. GLRLM generate for each sample of image fragment. A set of
consecutive pixels with the same gray level is called a gray level run. The
number of pixels in a run is the run length. In order to extract texture features
gray level run length matrix are computed. For each element, (i, j ) the run
length, r of the GLRLM represents the number of runs of gray level i having
97

length j . GLRLM can be computed for any direction. Mostly five features are
derived from the GLRLM. These features are: Short Runs Emphasis (SRE),
Long Runs Emphasis (LRE), Gray Level Non-Uniformity (GLNU), Run
Length Non-Uniformity (RLNU), and Run Percentage (RPERC). These are
quite improved in representing binary textures.

4.4.4 Local Binary Pattern Features

Local binary pattern (LBP) operator is introduced as a


complementary measure for local image contrast. Lahdenoja (2005) discussed
the LBP operator associate statistical and structural texture analysis. The LBP
describes texture with smallest primitives called textons (or, histograms of
texture elements). For each pixel in an image, a binary code is produced by
thresholding, its neighbourhood with the value of the center pixel. A
histogram is then assembled to collect the occurrences of different binary
codes representing different types of curved edges, spots, flat areas, etc.

This histogram is an arrangement as the feature vector result of


applying the LBP operator. The LBP operator considers only the eight nearest
neighbours of each pixel and it is rotation variant, but invariant to monotonic
changes in gray-scale can be applied. The dimensionality of the LBP feature
distribution can be calculated according to the number of neighbours used.
LBP is one of the most used approaches in practical applications, as it has the
advantage of simple implementation and fast performance.

Some related features are Scale-Invariant Feature Transform (SIFT)


descriptor (SIFT is a distinctive invariant feature set that is suitable for
describing local textures), LPQ (Local Phase Quantization) operator, Center-
Symmetric LBP (CS-LBP) and Volume-LBP.
98

4.4.5 Auto Correlation Features

An important characteristic of texture is the repetitive nature of the


position of texture elements in the image. An autocorrelation function can be
evaluated that measures this coarseness. Based on the observation of
autocorrelation feature is computed that some textures are repetitive in nature,
such as textiles. The autocorrelation feature of an image is used to evaluate
the fineness or roughness of the texture present in the image. This function is
related to the size of the texture primitive for example the fitness of the
texture. If the texture is rough or unsmooth, then the autocorrelation function
will go down slowly, if not it will go down very quickly. For normal textures,
the autocorrelation function will show peaks and valleys. It has relationship
with power spectrum of the fourier transform. It is also responsive to noise
interference. The autocorrelation function of an image I ( x, y ) is defined in
Equation (4.4) as follows

N N
I (u , v) I (u x, v y)
u 0 v 0
P ( x, y ) N N
(4.4)
2
I (u , v)
u 0v 0

4.4.6 Co-occurrence Matrix – SGLD

Statistical methods use second order statistics to model the


relationships between pixels within the region by constructing Spatial Gray
Level Dependency (SGLD) matrices. A SGLD matrix is the joint probability
occurrence of gray levels i and j for two pixels with a defined spatial
relationship in an image. The spatial relationship is defined in terms of
distance, d and angle, . If the texture is coarse and distance d is small
compared to the size of the texture elements, the pairs of points at distance d
should have similar gray levels. Conversely, for a fine texture, if distance d is
99

comparable to the texture size, then the gray levels of points separated by
distance d should often be quite different, so that the values in the SGLD
matrix should be spread out relatively uniformly.

Hence, one of the ways to analyze texture coarseness would be, for
various values of distance d , some measure of scatter of the SGLD matrix
around the main diagonal. Similarly, if the texture has some direction, i.e., is
coarser in one direction than another, then the degree of spread of the values
about the main diagonal in the SGLD matrix should vary with the direction .
Thus texture directionality can be analyzed by comparing spread measures of
SGLD matrices constructed at various distances of d . From SGLD matrices, a
variety of features may be extracted.

From each matrix, 14 statistical measures are extracted including:


angular second moment, contrast, correlation, variance, inverse difference
moment, sum average, sum variance, sum entropy, difference variance,
difference entropy, information measure of correlation, information measure
of correlation II and maximal correlation coefficient. The measurements
average the feature values in all four directions.

4.4.7 Edge Frequency based Texture Features

A number of edge detectors can be used to yield an edge image from


an original image. An edge dependent texture description function E can be
computed using Equation (4.5) as follows

E | f (i, j ) f (i d , j ) | | f (i, j ) f (i d , j ) | | f (i, j ) f (i, j d) |


(4.5)
| f (i, j ) f (i, j d ) |

This function is inversely related to the autocorrelation function.


Texture features can be evaluated by choosing specified distances d. It varies
the distance, d , parameter from 1 to 70 giving a total of 70 features.
100

4.4.8 Primitive Length Texture Features

Coarse textures are represented by a large number of neighbouring


pixels with the same gray level, whereas a small number represents fine
texture. A primitive is a continuous set of maximum number of pixels in the
same direction that have the same gray level. Each primitive is defined by its
gray level, length and direction. Let B(a, r ) represents the number of
primitives of all directions having length r and gray level a . Assume M , N be
image dimensions, L is the number of gray levels, N r is the maximum
primitive length in the images and K is the total number of runs. It is given
by the Equation (4.6) as

L Nr
B ( a, r ) (4.6)
a 1 r 1

Then, the Equations (4.6) – (4.10) define the five features of image
texture.

L Nr
1 B ( a, r )
Short primitive emphasis = (4.7)
K a 1 r 1 r2

L Nr
1
Long primitive emphasis = B ( a, r ) 2 (4.8)
K a 1 r 1

Nr 2
L
1 2
Gray level uniformity = B ( a, r ) r (4.9)
K a 1 r 1

Nr 2
L
1
Primitive length uniformity = B ( a, r ) (4.10)
K a 1 r 1

K K
Primitive percentage = L Nr
(4.11)
MN
rB (a, r )
a 1 r 1
101

4.4.9 Law’s Texture Features

Law’s of texture observed that certain gradient operators such as


Laplacian and Sobel operators accentuated the underlying microstructure of
texture within an image. This was the basis for a feature extraction scheme
based a series of pixel impulse response arrays obtained from combinations of
1-D vectors shown in Figure 4.9. Each 1-D array is associated with an
underlying microstructure and labeled using an acronym accordingly. The
arrays are convolved with other arrays in a combinatorial manner to generate
a total of 25 masks, typically labeled as L5, E5, S5, W5 and R5 for the mask
resulting from the convolution of the two arrays.

Level L5 [ 1 4 6 4 1 ]
Edge E 5 [ 1 2 0 2 1 ]
Spot S5 [ 1 0 2 0 1 ]
Wave W 5 [ 1 2 0 2 1 ]
Ripple R 5 [ 1 4 6 4 1 ]

Figure 4.9 Five 1D arrays identified by laws

These masks are subsequently convolved with a texture field to


accentuate its microstructure giving an image from which the energy of the
microstructure arrays is measured together with other statistics. The
commonly used features are mean, standard deviation, skewness, kurtosis and
energy measurements. Since there are 25 different convolutions, altogether it
obtains a total of 125 features.

For all feature extraction methods, the most appropriate features are
selected for classification using a linear stepwise discriminant analysis.

Among the above mentioned techniques, researchers suggested the


GLCM is one of the very best feature extraction techniques. From GLCM,
102

many useful textural properties can be calculated to expose details about the
image. However, the calculation of GLCM is very computationally intensive
and time consuming.

4.5 GRAY LEVEL CO-OCCURRENCE MATRIX

In 1973, Haralick introduced the co-occurrence matrix and texture


features which are the most popular second order statistical features today.
Haralick proposed two steps for texture feature extraction. First step is
computing the co-occurrence matrix and the second step is calculating texture
feature based on the co-occurrence matrix. This technique is useful in wide
range of image analysis applications from biomedical to remote sensing
techniques.

4.5.1 Working of GLCM

Basic of GLCM texture considers the relation between two


neighbouring pixels in one offset, as the second order texture. The gray value
relationships in a target are transformed into the co-occurrence matrix space
by a given kernel mask such as 3 3 , 5 5 , 7 7 and so forth. In the
transformation from the image space into the co-occurrence matrix space, the
neighbouring pixels in one or some of the eight defined directions can be
used; normally, four direction such as 0°, 45°, 90°, and 135° is initially
regarded, and its reverse direction (negative direction) can be also counted
into account. It contains information about the positions of the pixels having
similar gray level values.

Each element (i, j ) in GLCM specifies the number of times that the
pixel with value i occurred horizontally adjacent to a pixel with value j . In
Figure 4.8, computation has been made in the manner where, element (1, 1) in
the GLCM contains the value 1 because there is only one instance in the
103

image where two, horizontally adjacent pixels have the values 1 and 1.
Element (1, 2) in the GLCM contains the value 2 because there are two
instances in the image where two, horizontally adjacent pixels have the values
1 and 2.

Figure 4.10 Creation of GLCM from image matrix

Element (1, 2) in the GLCM contains the value 2 because there are
two instances in the image where two, horizontally adjacent pixels have the
values 1 and 2. The GLCM matrix has been extracted for input dataset
imagery. Once after the GLCM is computed, texture features of the image are
being extracted successively.

4.6 HARALICK TEXTURE FEATURES

Haralick extracted thirteen texture features from GLCM for an


image. The important texture features for classifying the image into water
body and non-water body are Energy (E), Entropy (Ent), Contrast (Con),
Inverse Difference Moment (IDM) and Directional Moment (DM).
104

Andrea Baraldi and Flavio Parmiggiani (1995) discussed the five


statistical parameter energy, entropy, contrast, IDM and DM, which are
considered the most relevant among the 14 originally texture features
proposed by Haralick et al. (1973). The complexity of the algorithm also
reduced by using these texture features.

Let i and j are the coefficients of co-occurrence matrix, M i, j is


the element in the co-occurrence matrix at the coordinates i and j and N is
the dimension of the co-occurrence matrix.

4.6.1 Energy

Energy (E) can be defined as the measure of the extent of pixel pair
repetitions. It measures the uniformity of an image. When pixels are very
similar, the energy value will be large. It is defined in Equation (4.12) as

N 1 N 1
E M 2 i, j (4.12)
i 0 j o

4.6.2 Entropy

This concept comes from thermodynamics. Entropy (Ent) is the


measure of randomness that is used to characterize the texture of the input
image. Its value will be maximum when all the elements of the co-occurrence
matrix are the same. It is also defined as in Equation (4.13) as

N 1 N 1
Ent M i, j ( ln(M (i, j ))) (4.13)
i 0 j o

4.6.3 Contrast

The contrast (Con) is defined in Equation (4.14), is a measure of


intensity of a pixel and its neighbour over the image. In the visual perception
105

of the real world, contrast is determined by the difference in the colour and
brightness of the object and other objects within the same field of view.

N 1 N 1
2
Con i j M i, j (4.14)
i 0 j o

4.6.4 Inverse Difference Moment

Inverse Difference Moment (IDM) is a measure of image texture as


defined in Equation (4.15). IDM is usually called homogeneity that measures the
local homogeneity of an image. IDM feature obtains the measures of the closeness
of the distribution of the GLCM elements to the GLCM diagonal. IDM has a range
of values so as to determine whether the image is textured or non-textured.

N 1 N 1
1
IDM 2
M i, j (4.15)
i 0 j o 1 i j

4.6.5 Directional Moment

Directional moment (DM), as the name signifies, this is a textural


property of the image computed by considering the alignment of the image as
a measure in terms of the angle and it is defined as in Equation (4.16)

N 1 N 1
DM M i, j i j (4.16)
i 0 j o

The Table 4.1 shows some of the texture features extracted using
GLCM, to classify an image into water body and non-water body region.
106

Table 4.1 Texture features extracted using GLCM

Energy Entropy Contrast IDM DM


0.2398 6.8042 0.124 6.15E+04 237.2482
0.1949 7.0086 0.1904 6.01E+04 324.5662
0.3168 6.4868 0.2488 5.99E+04 389.9733
0.1524 6.7707 0.2025 5.95E+04 334.9738
0.7568 5.5702 0.12 6.28E+04 272.9614
0.1655 7.3033 0.1999 5.93E+04 286.0215
0.313 6.6852 0.1464 6.12E+04 270.9697
0.2236 6.9529 0.1739 6.10E+04 289.1159
0.5483 5.8905 0.1019 6.26E+04 229.464
0.5583 5.9409 0.1524 6.26E+04 253.8671
0.5143 6.1439 0.0794 6.31E+04 197.9385
0.2486 6.6115 0.1654 6.06E+04 261.9178
0.1608 6.88 0.1993 6.05E+04 302.5752
0.4855 5.9474 0.0953 6.28E+04 211.5594
0.1613 6.9496 0.1639 6.07E+04 283.433
0.2853 6.4627 0.2106 6.01E+04 339.5505
0.1477 7.0368 0.3293 5.73E+04 444.9242
0.316 5.9372 0.1803 6.03E+04 281.7755
0.3046 6.4706 0.1998 6.03E+04 339.1179
0.2796 6.4406 0.2019 6.02E+04 322.4984
0.573 6.0185 0.1416 6.21E+04 288.6223
0.1729 7.2134 0.1497 6.05E+04 234.1717
0.3145 6.804 0.1592 6.09E+04 274.1024
0.7637 5.3457 0.0753 6.35E+04 184.061
0.6113 5.8042 0.138 6.19E+04 264.5534
0.7586 5.3523 0.0594 6.37E+04 164.2614
0.3124 6.2919 0.1397 6.15E+04 281.9666
0.5817 6.0175 0.1585 6.14E+04 290.4872
0.1226 7.1201 0.2642 5.73E+04 337.0597
0.1993 7.2553 0.2249 5.89E+04 365.6291
0.7293 5.5209 0.1787 6.16E+04 279.578
0.5257 6.4206 0.091 6.30E+04 211.6174
0.3006 6.4985 0.1749 6.10E+04 318.4937
0.1576 6.8883 0.1738 6.04E+04 291.3021
0.1929 7.1205 0.2127 5.86E+04 294.5526
0.1727 7.1763 0.2284 5.80E+04 277.6838
0.8759 5.0943 0.0371 6.46E+04 137.2638
0.285 6.7064 0.1587 6.09E+04 300.2948
0.1382 7.4154 0.19 5.94E+04 287.7632
0.3316 6.7746 0.1325 6.16E+04 268.2331
107

4.7 APPLICATION OF TEXTURE

Texture analysis methods have been utilized in a variety of


application domains such as automated inspection, medical image processing,
document processing, remote sensing and content-based image retrieval.

4.7.1 Remote Sensing

Texture analysis has been extensively used to classify remotely


sensed images. Land use classification where homogeneous regions with
different types of terrains (such as wheat, bodies of water, urban regions, etc.)
need to be identified is an important application.

4.7.2 Medical Image Analysis

Image analysis techniques have played an important role in several


medical applications. In general, the applications involve the automatic
extraction of features from the image which is then used for a variety of
classification tasks, such as distinguishing normal tissue from abnormal tissue.
Depending upon the particular classification task, the extracted features
capture morphological properties, colour properties, or certain textural
properties of the image.

4.8 SUMMARY

This chapter detailed the gray level co-occurrence matrix based


feature extraction to obtain energy, entropy, contrast, inverse difference
moment and directional moment. These texture features are served as the
input to classify the image accurately. Effective use of multiple features of the
image and the selection of a suitable classification method are especially
significant for improving classification accuracy. The chapter 5 discusses
classification techniques for improving accuracy along with their applications.

You might also like