DIP Notes

Download as pdf or txt
Download as pdf or txt
You are on page 1of 22

Digital Image Processing

1. Explain the fundamental steps in digital image processing which can be applied to
images.
Image processing is the process of transforming an image into a digital form and
performing certain operations to get some useful information from it. The image
processing system usually treats all images as 2D signals when applying certain
predetermined signal processing methods.
Fundamental Image Processing Steps
 Image Acquisition: Image acquisition is the first step in image processing. This step is
also known as preprocessing in image processing. It involves retrieving the image from a
source, usually a hardware-based source.
 Image Enhancement: Image enhancement is the process of bringing out and
highlighting certain features of interest in an image that has been obscured. This can
involve changing the brightness, contrast, etc.
 Image Restoration: Image restoration is the process of improving the appearance of an
image. However, unlike image enhancement, image restoration is done using certain
mathematical or probabilistic models.
 Color Image Processing:Color image processing includes a number of color modeling
techniques in a digital domain. This step has gained prominence due to the significant use
of digital images over the internet.
 Wavelets and Multiresolution Processing: Wavelets are used to represent images in
various degrees of resolution. The images are subdivided into wavelets or smaller regions
for data compression and for pyramidal representation.
 Compression:Compression is a process used to reduce the storage required to save an
image or the bandwidth required to transmit it. This is done particularly when the image
is for use on the Internet.
 Morphological Processing: Morphological processing is a set of processing operations
for morphing images based on their shapes.
 Segmentation: Segmentation is one of the most difficult steps of image processing. It
involves partitioning an image into its constituent parts or objects.
 Representation and Description: After an image is segmented into regions in the
segmentation process, each region is represented and described in a form suitable for
further computer processing. Representation deals with the image’s characteristics and
regional properties. Description deals with extracting quantitative information that helps
differentiate one class of objects from the other.
 Recognition:Recognition assigns a label to an object based on its description.
2. Define histogram equalization. Explain the procedure for histogram equalization.
Histogram Equalization is a computer image processing technique used to improve
contrast in images. It accomplishes this by effectively spreading out the most frequent
intensity values, i.e. stretching out the intensity range of the image. This method usually
increases the global contrast of images when its usable data is represented by close
contrast values. This allows for areas of lower local contrast to gain a higher contrast.
The histogram equalization process for digital images consists of four steps:
1. Find the running sum of the histogram values
2. Normalize the values from step1 by dividing by total number of pixels.
3. Multiply the values from step2 by the maximum gray level value and round.
4. Map the gray-level values to the results from step 3, using a one-toone
correspondence.
3. What is the purpose of image restoration? Explain the model of image degradation
and restoration process using suitable block diagram.
Ans: The purpose of image restoration is to "compensate for" or "undo" defects which
degrade an image. Degradation comes in many forms such as motion blur, noise, and
camera misfocus. In cases like motion blur, it is possible to come up with an very good
estimate of the actual blurring function and "undo" the blur to restore the original image.
In cases where the image is corrupted by noise, the best we may hope to do is to
compensate for the degradation it caused. In this project, we will introduce and
implement several of the methods used in the image processing world to restore images.
The Fig. 1 shows, the degradation process is modeled as a degradation function that,
together with an additive noise term, operates on an input image f(x, y) to produce a
degraded image g(x, y). Given g(x, y), some knowledge about the degradation function
H, and some knowledge about the additive noise term η(x, y), the objective of restoration
is to obtain an estimate f(x, y) of the original image. the estimate should be as close as
possible to the original input image and, in general, the more we know about H and η, the
closer f(x, y) will be to f(x, y).

The degraded image is given in the spatial domain by

g (x, y) = h (x, y) * f (x, y) + η (x, y)

where h (x, y) is the spatial representation of the degradation function and, the symbol *
indicates convolution. Convolution in the spatial domain is equal to multiplication in the
frequency domain, hence

G (u, v) = H (u, v) F (u, v) + N (u, v)

where the terms in capital letters are the Fourier transforms of the corresponding terms in
above equation.
Fig. 1 model of the image degradation/restoration process.

4. Describe arithmetic coding with an example for compression of image.


Unlike the variable-length codes described previously, arithmetic coding generates nonblock
codes. In arithmetic coding, which can be traced to the work of Elias, a one-to-one
correspondence between source symbols and code words does not exist. Instead, an entire
sequence of source symbols (or message) is assigned a single arithmetic code word. The code
word itself defines an interval of real numbers between 0 and 1. As the number of symbols in
the message increases, the interval used to represent it becomes smaller and the number of
information units (say, bits) required to represent the interval becomes larger. Each symbol
of the message reduces the size of the interval in accordance with its probability of
occurrence. Because the technique does not require, as does Huffman's approach, that each
source symbol translate into an integral number of code symbols (that is, that the symbols be
coded one at a time), it achieves (but only in theory) the bound established by the noiseless
coding theorem.

Fig.1 Arithmetic coding procedure


Figure 1 illustrates the basic arithmetic coding process. Here, a five-symbol sequence or
message, a1a2a3a3a4, from a four-symbol source is coded. At the start of the coding process,
the message is assumed to occupy the entire half-open interval [0, 1). As Table.2 shows, this
interval is initially subdivided into four regions based on the probabilities of each source
symbol. Symbol ax, for example, is associated with subinterval [0, 0.2). Because it is the first
symbol of the message being coded, the message interval is initially narrowed to [0, 0.2).
Thus in Fig. 5.1 [0, 0.2) is expanded to the full height of the figure and its end points labeled
by the values of the narrowed range. The narrowed range is then subdivided in accordance
with the original source symbol probabilities and the process continues with the next
message symbol.

FIG.2

5. Differentiate between image enhancement and image restoration.


6. Explain the theory of sampling of an image.
Digitizing the co-ordinate value is called sampling. ince an analogue image is continuous
not just in its co-ordinates (x axis), but also in its amplitude (y axis), so the part that deals
with the digitizing of co-ordinates is known as sampling. In digitizing sampling is done on
independent variable. In case of equation y = sin(x), it is done on x variable.

When looking at this image, we can see there are some random variations in the signal
caused by noise. In sampling we reduce this noise by taking samples. It is obvious that more
samples we take, the quality of the image would be more better, the noise would be more
removed and same happens vice versa. However, if you take sampling on the x axis, the
signal is not converted to digital format, unless you take sampling of the y-axis too which is
known as quantization.Sampling has a relationship with image pixels. The total number of
pixels in an image can be calculated as Pixels = total no of rows * total no of columns. For
example, let’s say we have total of 36 pixels, that means we have a square image of 6X 6.
As we know in sampling, that more samples eventually result in more pixels. So it means
that of our continuous signal, we have taken 36 samples on x axis. That refers to 36 pixels
of this image. Also the number sample is directly equal to the number of sensors on CCD
array.
7. Define DFT. State and prove the following properties:
(i) Correlation (ii) Periodicity
The discrete-time Fourier transform (DFT) represents an image as a sum of complex
exponential of varying magnitudes, frequency and phases. Square functions as an image
representation of a square, while the natural images are the image representation with rich
local covariance. We can perform DFT for all these image types to improve their
qualities.
i. Circular Correlation
The Complex correlation property states

Here rxy(l) is circular cross correlation which is given as

ii. Periodicity
Let x(n) and x(k) be the DFT pair then if
x(n+N) = x(n) for all n then,
X(k+N) = X(k) for all k
Thus periodic sequence xp(n) can be given as
8. Define compression and explain the general compression system model.
Compression: It is the process of reducing the size of the given data or an image. It will help us
to reduce the storage space required to store an image or File.
Image Compression Model: There are two Structural model and they are broadly Classified as
follows
1.An Encoder
2.ADecoder.
An Input image f(x,y) is fed in to encoder and create a set of symbols and after transmission over
the channel ,the encoded representation is fed in to the decode.
A General Compression system model:
The General system model consist of the following components, They are broadly classified as
1.Source Encoder
2.Channel Encoder
3.Channel
4.Channel Decoder
5.Souce Decoder

The Source Encoder Will removes the input redundancies. The channel encoder will increase the
noise immunity of the source encoder’s output. If the channel between encoder and decoder is
noise free then the channel encoder and decoder can be omitted

MAPPER: It transforms the input data in to a format designed to reduce the inter pixel
redundancy in the input image.
QUANTIZER: It reduce the accuracy of the mapper’s output.
SYMBOL ENCODER: It creates a fixed or variable length code to represent the quantizer’s
output and maps the output in accordance with the code
SYMBOL DECODER: The inverse operation of the source encoder’s symbol will be performed
and maps the blocks.
9. Explain the elements of an image processing system.
Image Processing System is the combination of the different elements involved in the
digital image processing. It consists of following components:-
 Image Sensors: Image sensors senses the intensity, amplitude, co-ordinates and other
features of the images and passes the result to the image processing hardware. It
includes the problem domain.
 Image Processing Hardware: Image processing hardware is the dedicated hardware
that is used to process the instructions obtained from the image sensors. It passes the
result to general purpose computer.
 Computer: Computer used in the image processing system is the general purpose
computer that is used by us in our daily life.
 Image Processing Software: Image processing software is the software that includes
all the mechanisms and algorithms that are used in image processing system.
 Mass Storage: Mass storage stores the pixels of the images during the processing.
 Hard Copy Device: Once the image is processed then it is stored in the hard copy
device. It can be a pen drive or any external ROM device.
 Image Display: It includes the monitor or display screen that displays the processed
images.
 Network: Network is the connection of all the above elements of the image processing
system.

10. Explain about the discrete cosine transform and write its applications.
Discrete cosine transform (DCT) is a transform that is mainly used in compression
algorithms. It transforms data points in a spatial domain into a frequency domain. This
makes it easier to find the repetition of patterns. Like any other transform, it is also
invertible. This means we can return the actual data points if the transforms are given.
Several formulas define DCT, but the most common one is given below:
𝜋𝑘(2𝑛 + 1)
𝑋(𝑘) = 2 𝑥(𝑛)cos ( )
2𝑁

Applications

As mentioned before, DCT does a great job of highlighting the frequencies. Which can
further show redundancies. This makes it perfect for compression applications. Some of
the applications of DCT are as follows:

 Image compression algorithms like JPEG and HEIF.


 Audio file formats like MP3, AAC, and more.
 Video file formats like MPEG.
 Scientists and Engineers also use them for digital signal processing, telecommunications,
and more.
11. Explain about Huffman coding by taking an example.
Huffman coding is a lossless data compression algorithm. The idea is to assign variable-
length codes to input characters, lengths of the assigned codes are based on the
frequencies of corresponding characters. The most frequent character gets the smallest
code and the least frequent character gets the largest code. The variable-length codes
assigned to input characters are Prefix Codes, means the codes (bit sequences) are
assigned in such a way that the code assigned to one character is not prefix of code
assigned to any other character. This is how Huffman Coding makes sure that there is no
ambiguity when decoding the generated bit stream. Let us understand prefix codes with a
counter example. Let there be four characters a, b, c and d, and their corresponding
variable length codes be 00, 01, 0 and 1. This coding leads to ambiguity because code
assigned to c is prefix of codes assigned to a and b. If the compressed bit stream is 0001,
the de-compressed output may be “cccd” or “ccb” or “acd” or “ab”.
Decoding:
• Decoding is an exact reverse procedure of encoding. In Huffman's tree the leaves
represent the character and from the given code word we can find out the original
characters.
• In above example, if we decode bits,
• 0111111111110
• The resulting string would be “aeed”
Example
12. Define image segmentation. Give classification. Explain region based segmentation.
Image segmentation is a branch of digital image processing which focuses on partitioning
an image into different parts according to their features and properties. The primary goal
of image segmentation is to simplify the image for easier analysis. In image
segmentation, you divide an image into various parts that have similar attributes. The
parts in which you divide the image are called Image Objects.
The types of image segmentation techniques:
1. Thresholding Segmentation
2. Edge-Based Segmentation
3. Region-Based Segmentation
4. Watershed Segmentation
5. Clustering-Based Segmentation Algorithms
6. Neural Networks for Segmentation
Region-Based Segmentation
Region-based segmentation algorithms divide the image into sections with similar features.
These regions are only a group of pixels and the algorithm find these groups by first locating
a seed point which could be a small section or a large portion of the input image. After
finding the seed points, a region-based segmentation algorithm would either add more pixels
to them or shrink them so it can merge them with other seed points. Based on these two
methods, we can classify region-based segmentation into the following categories:
Region Growing
In this method, start with a small set of pixels and then start iteratively merging more pixels
according to particular similarity conditions. A region growing algorithm would pick an
arbitrary seed pixel in the image, compare it with the neighbouring pixels and start increasing
the region by finding matches to the seed point. When a particular region can’t grow further,
the algorithm will pick another seed pixel which might not belong to any existing region.
One region can have too many attributes causing it to take over most of the image. To avoid
such an error, region growing algorithms grow multiple regions at the same time.
Region Splitting and Merging
As the name suggests, a region splitting and merging focused method would perform two
actions together – splitting and merging portions of the image. It would first split the image
into regions that have similar attributes and merge the adjacent portions which are similar to
one another. In region splitting, the algorithm considers the entire image while in region
growth, the algorithm would focus on a particular point. The region splitting and merging
method follows a divide and conquer methodology. It divides the image into different
portions and then matches them according to its predetermined conditions.

13. Define boundary of a region.


The boundary (also called border or contour) of a region R is the set of pixels in the region
that have one or more neighbors that are not in R.
14. Explain about Hough transform.
The HT is a feature extraction method in image analysis, computer vision, and digital image
processing . It uses a voting mechanism to identify bad examples of objects inside a given class of
forms. This voting mechanism is carried out in parameter space. First, object candidates are
produced as local maxima in an accumulator space, using the HT algorithm.The traditional HT
was concerned with detecting lines in an image, but it was subsequently expanded to identifying
locations of arbitrary shapes, most often circles or ellipse. The Hough approach is effective for
computing a global description of a feature(s) from (potentially noisy) where the number of
solution classes does not need to be provided before. For example, the Hough approach for line
identification is motivated by the assumption that each input measurement reflects its contribution
to a globally consistent solution (e.g., the physical line which gave rise to that image point).

A line can be described analytically in a variety of ways. One of the line equations uses the
parametric or normal notion: xcosθ+ysinθ=r. where r is the length of a normal from the origin to
this line and θ is the orientation as given in Figure.

The known variables (i.e., xi,yi) in the Image are constants in the parametric line equation,
whereas r and are the unknown variables we seek. Points in cartesian image space correspond to
curves in the polar Hough parameter space if we plot the potential (r, θ) values specified by each.
The Hough transformation for straight lines is this point-to-curve transformation. Collinear spots
in the cartesian image space become clearly obvious when examined in the Hough parameter
space because they provide curves that overlap at a single (r, θ) point. Where a and b are the
circle’s center coordinates, and r is the radius. Because we now have three coordinates in the
parameter space and a 3-D accumulator, the algorithm’s computing complexity increases. (In
general, the number of parameters increases the calculation and the size of the accumulator array
polynomially.) As a result, the fundamental Hough approach described here is only applicable to
straight lines.
15. a) What do you mean by image histogram?
b) Write down the steps of the Histogram Equalization method of image
enhancement.
c) Write down the differences between Histogram Equalization and Histogram
Matching.

We can define the histogram of an image as a 2D bar plot. The horizontal axis
represents the pixel intensities. The vertical axis denotes the frequency of each intensity.

Steps Involved
1. Get the input image
2. Generate the histogram for the image
3. Find the local minima of the image
4. Divide the histogram based on the local minima
5. Have the specific gray levels for each partition of the histogram
6. Apply the histogram equalization on each partition

The comparison between histogram equalization, histogram matching is as follows:


Histogram Equalization
 Histogram Equalization is an image processing technique.
 It enhances contrast by adjusting the intensities of the image.
 It ensures that an image contains a uniform distribution of intensities.
Histogram Specification
 It is also known as histogram matching.
 It refers to the generalized version of Histogram Equalization.
 It is a transformation technique which ensures that the histogram of an image matches a
specified histogram.
16. a)Define Discrete Fourier Transform and its inverse in two-dimension.
b) Calculate Fourier Transform of a continuous unit impulse located at origin.

The discrete Fourier transform (DFT) of an image f of size 𝑴 × 𝑵 is an image F of same size
defined as:
𝑴 𝟏𝑵 𝟏
𝒖𝒏 𝒗𝒏
𝒋𝟐𝝅(
𝑭(𝒖, 𝒗) = 𝒇(𝒎, 𝒏) 𝒆 𝑴 𝑵
𝒎 𝟎𝒏 𝟎
In the sequel, we note F the DFT so that F[f]=F

The Fourier transform of a continuous-time function x(t) can be defined as,



𝒋𝝎𝒕
𝑿(𝝎) = 𝒙(𝒕)𝒆 𝒅𝒕

The unit impulse function is defined as,
δ(t)= 1 for t=0\0for t≠0
If it is given that
x(t) = δ(t)
Then, from the definition of Fourier transform, we have,
∞ ∞
𝒋𝝎𝒕 𝒋𝝎𝒕
𝑿(𝝎) = 𝒙(𝒕)𝒆 = 𝜹(𝒕)𝒆 𝒅𝒕
∞ ∞
As the impulse function exists only at t= 0. Thus,
∞ ∞
𝒋𝝎𝒕 𝒋𝝎𝒕 𝒋𝝎𝒕
𝑿(𝝎) = 𝜹(𝒕)𝒆 𝒅𝒕 = 𝟏. 𝒆 𝒅𝒕 = 𝒆 |𝒕 𝟎 =𝟏
∞ ∞

𝑭𝑻
𝑭[𝜹(𝒕)] = 𝟏 𝒐𝒓 𝜹(𝒕) ↔ 𝟏
That is, the Fourier transform of a unit impulse function is unity.
17. What is understood by spatial domain representation ?
In computer imaging, a two-dimensional image can be represented by a two-dimensional
function f{x, y) where x and y are the spatial coordinates in the 2- D image plane and f is
either a scalar describing the gray level or a vector containing three components
describing the intensities of the three primary colors at the location represented by
coordinates x and y.
18. a)Write down the types of redundancies that might be present in a typical image.
b) Define Run-Length code for a binary image.

In digital image compression, three basic data redundancies can be identified and
exploited: coding redundancy, interpixel redundancy, and psychovisual redundancy.
Coding Redundancy: In this, we utilize formulation to show how the gray-level
histogram of an image also can provide a great deal of insight into the construction of
codes to reduce the amount of data used to represent it.
Interpixel redundancy: Interpixel redundancy is due to the correlation between the neighboring
pixels in an image. That means neighboring pixels are not statistically independent. The gray
levels are not equally probable. The value of any given pixel can be predicated from the value of
its neighbors that is they are highly correlated.
Psychovisual Redundancy: Certain information simply has less relative importance than other
information in normal visual processing. This information is said to be psychovisually redundant.
It can be eliminated without significantly impairing the quality of image perception.

Run-length coding Method that works by counting the number of adjacent pixels with the
same grey-level value. This count, called the run length, is then coded and stored. Basic
methods are used primarily for binary images, but can be used for more complex images
that have been pre-processed by thresholding to reduce the number of gray levels to
two.Basic concept is to code each contiguous group of 1’s or 0’s encountered in a left to
right scan of a row by its length, and establishing a convention for determining the length
of the run.
The most common approaches for determining the value of the run are:-
(i) specify the value of the first run of each row, or
(ii) (ii) assume that the row begins with a white run, whose run-length
may in fact be zero.
19. Differentiate between spatial domain enhancement and frequency domain
enhancement.

20. Compute the Size of 800x600 image at 240 pixels per inch?
If the image is 800 pixels by 600 pixels at 240 pixels per inch, then the image size is (800
/ 240) by (600 / 240) or 31/3 by 21/2 inches.
21. What is the difference between lossy and lossless compression ?

22. Describe the region growing technique for image segmentation and mention the
problem associate with it.
Region Growing
In this method, you start with a small set of pixels and then start iteratively merging more
pixels according to particular similarity conditions. A region growing algorithm would pick
an arbitrary seed pixel in the image, compare it with the neighbouring pixels and start
increasing the region by finding matches to the seed point.
When a particular region can’t grow further, the algorithm will pick another seed pixel which
might not belong to any existing region. One region can have too many attributes causing it to
take over most of the image. To avoid such an error, region growing algorithms grow multiple
regions at the same time.
You should use region growing algorithms for images that have a lot of noise as the noise
would make it difficult to find edges or use thresholding algorithms.
Disadvantage: 1. The computation is consuming, no matter the time or power.
2. Noise or variation of intensity may result in holes or over segmentation.
3. This method may not distinguish the shading of the real images.
23. Difference between Huffman coding and Arithmetic coding.

24. Explain quantization.


Quantization:
It discusses the effects produced when the number of bits used to represent the
grey level in an image is decreased .this is illustrated by reducing the grey level required
to represent a 1024*1024,512 image. The 256,128,and 64 level image are visually identical
for all practical purposes the 32 level images has developed a set of rigid like structure in
areas of smooth grey I+Im lines.this effect caused by the user insufficient number of grey
levels in smooth areas of digital image is called a false contouring.this is visible in images
displayed using 16 or lesser gray level values.
25. Describe the fundamental steps in image processing?
Digital image processing encompasses a broad range of hardware, software and theoretical
underpinnings.

The problem domain in this example consists of pieces of mail and the objective is to read the
address on each piece. Thus the desired output in this case is a stream of alphanumeric characters.
The first step in the process is image acquisition that is acquire a digital image .To do so requires an
imaging sensor and the capability to digitize the signal produced by the sensor. After the digital
image has been obtained the next step deals with preprocessing that image. The key function of this
is to improve the image in ways that increase the chances for success of the other processes. The next
stage deals with segmentation. Broadly defined segmentation partitions an input image into its
constituent parts or objects. The key role of this is to extract individual characters and words from the
background, The output of the segmentation stage usually is raw pixel data, constituting either the
boundary of a region or all the points in the region itself. Choosing a representation is only part of the
solution for transforming raw data into a form suitable for subsequent computer processing.
Description also called feature selection deals with extracting features that result in some quantitative
information of interest that are basic for differentiating one class of object from another. The last
stage involves recognition and interpretation. Recognition is the process that assigns a label to an
object based on the information provided by its descriptors. Interpretation involves assigning
meaning to an ensemble of recognized objects. Knowledge about a problem domain is coded into an
image processing system in the form of knowledge database. This knowledge may be simple as
detailing regions of an image where the information of interest is known to be located thus limiting
the search that has to be conducted in seeking that information. The knowledge base also can be quite
complex such as an interrelated list of all major possible defects in a materials inspection problem or
an image database containing high resolution satellite images of a region in connection with change
detection application. Although we do not discuss image display explicitly at this point it is important
to keep in mind that viewing the results of image processing can take place at the output of any step .
26. Write short notes on image segmentation.
 Segmentation subdivides on image in to its constitute regions or objects. The level to
which the subdivides is carried depends on the problem being solved .
 Examples: In autonomous air to ground target acquisition applications identifying
vehicles on a road is of interest.
 The first step is to segment the road from the image and then to segment the elements
of the road down to objects of a range of sizes that correspond potential vehicles.
 In target acquistition ,the system designer has no control of the environment. So the
usual approach is to focus on selecting the types of sensors most likely to enhance the
objects of interest
 Example is the use of infrared imaging to detect objects with a strong heat
signature,such as tanks in motion.
 Segmentation algorithms for monochrome images are based on one of the two
basic properties of gray level values . They are discontinuity and similarity.Based on the
first category ,the approach is based on abrupt changes in gray level and the areas of
interest based on this category are detection of isolated points and detection of lines and
edges in an image.Based on the second category the approach is based on thresholding,
region growing and region splitting and merging .
The concept of segmenting an image based on discontinuity or similarity of the
gray level values of its pixels is applicable to both static and dynamic images.

Short answer type question


Define Image?
An image may be defined as two dimensional light intensity function f(x, y) where x and y
denote spatial co-ordinate and the amplitude or value of f at any point (x, y) is called intensity or
gray scale or brightness of the image at that point.
What are the steps involved in DIP?
1. Image Acquisition
2. Preprocessing
3. Segmentation
4. Representation and Description
5. Recognition and Interpretation
Define sampling and quantization?
Sampling means digitizing the co-ordinate value (x, y).Quantization means digitizing the
amplitude value. Several rods are connected to one nerve end. So it gives the overall picture of
the image.This is also known as thin lightvision.
Write Any Four Applications Of Dip?

1. Remote sensing
2. Image transmission and storage for business application
3. Medical imaging
4. Astronomy

Explain Spatial Filtering?


Spatial filtering is the process of moving the filter mask from point to point in an image.
For linear spatial filter, the response is given by a sum of products of the filter
coefficients, and the corresponding image pixels in the area spanned by the filter mask.

What is meant by pixel?


A digital image is composed of a finite number of elements each of which has a particular
location or value. These elements are referred to as pixels or image elements or picture

elements or pels elements.

Write short notes on neighbors of a pixel.


The pixel p at co-ordinates (x, y) has 4 neighbors (ie) 2 horizontal and 2 vertical
neighbors whose co-ordinates is given by (x+1, y), (x-1,y), (x,y-1), (x, y+1). This is
called as direct neighbors. It is denoted by N4(P)
Four diagonal neighbors of p have co-ordinates (x+1, y+1), (x+1,y-1), (x-1, y-1),
(x-1, y+1). It is denoted by ND(4).
Eight neighbors of p denoted by N8(P) is a combination of 4 direct neighbors and

4 diagonal neighbors.

Specify the objective of image enhancement technique.


The objective of enhancement technique is to process an image so that the result is

more suitable than the original image for a particular application.

List the 2 categories of image enhancement.


· Spatial domain refers to image plane itself & approaches in this category are
based on direct manipulation of picture image.
· Frequency domain methods based on modifying the image by fourier transform.
Write the steps involved in frequency domain filtering.
1. Multiply the input image by (-1) to center the transform.
2. Compute F(u,v), the DFT of the image from (1).
3. Multiply F(u,v) by a filter function H(u,v).
4. Compute the inverse DFT of the result in (3).
5. Obtain the real part of the result in (4).
6. Multiply the result in (5) by (-1)
Define spatial filtering
Spatial filtering is the process of moving the filter mask from point to point in an
image. For linear spatial filter, the response is given by a sum of products of the filter
coefficients, and the corresponding image pixels in the area spanned by the filter mask.
What is segmentation?
Segmentation subdivides on image in to its constitute regions or objects. The level
to which the subdivides is carried depends on the problem being solved .That is egmentation
should when the objects of interest in application have been isolated.
Define region growing?
Region growing is a procedure that groups pixels or subregions in to layer regions
based on predefined criteria. The basic approach is to start with a set of seed points and
from there grow regions by appending to each seed these neighbouring pixels that have
properties similar to the seed.
What is image compression?
Image compression refers to the process of redundancy amount of data required to
represent the given quantity of information for digital image. The basis of reduction process is
removal of redundant data.
Define is coding redundancy?
If the gray level of an image is coded in a way that uses more code words than
necessary to represent each gray level, then the resulting image is said to contain coding
redundancy.
Define interpixel redundancy?
The value of any given pixel can be predicted from the values of its neighbors.
The information carried by is small. Therefore the visual contribution of a single pixel to
an image is redundant. Otherwise called as spatial redundant geometric redundant or
interpixel redundant.

Eg: Run length coding

What is run length coding?


Run-length Encoding, or RLE is a technique used to reduce the size of a repeating
string of characters. This repeating string is called a run; typically RLE encodes a run of
symbols into two bytes, a count and a symbol. RLE can compress any type of data
regardless of its information content, but the content of data to be compressed affects the
compression ratio. Compression is normally measured with the compression ratio:
Define psycho visual redundancy?
In normal visual processing certain information has less importance than other information. So
this information is said to be psycho visual redundant.
Define Huffman coding
 Huffman coding is a popular technique for removing coding redundancy.
 When coding the symbols of an information source the Huffman code yields the
smallest possible number of code words, code symbols per source symbol.
Define arithmetic coding
In arithmetic coding one to one corresponds between source symbols and code
word doesn’t exist where as the single arithmetic code word assigned for a sequence of source
symbols. A code word defines an interval of number between 0 and 1.

You might also like