Updated DIP UNIT1
Updated DIP UNIT1
Updated DIP UNIT1
TWO MARKS
1. What is meant by digital image processing?
Processing of digital image f(x, y)by means of digital computer is known as digital image
processing. It can be defined as the process of analyzing and manipulating images using a
computer.
15. List color models involved in hardware and their applications? (NOV/DEC-
2011)(A/M-18) (or) List the application of color models (NOV/DEC-2018)
1) RGB model – used for color monitor
2) CMY model – used for color printing
3) HIS model – used for color image
EC 8093 - DIP UNIT-1 DIGITAL IMAGE FUNDAMENTALS
17. What do you mean by zooming and shrinking in Digital image? (MAY/JUNE-2011)
Zooming results from oversampling and shrinking results from under sampling.
Zooming produces large size of image and shrinking produces small size of image.
If the ratio is larger, poor brightness discrimination because a large percentage of change
in intensity is needed.
23. What is meant by Mach band effect? (MAY/JUNE-2009)& (May 2015)(Nov/Dec 16)
Mach band effect proved that perceived brightness is not a simple function of intensity.
Although the intensity of strips is constant, we actually perceive a brightness pattern and are
called as Mach band effect. This is known an s Mach band effect.
.
31. State the two important properties of unitary transforms [NOV 2013]
i) Energy compaction – few transform coefficients have large magnitude.
ii) Energy conservation – preserve the norm of input vectors.
rotating in opposite direction and superimposed by each other. Then the multiplication of two
images will give a new image.
36. Define nearest neighbor interpolation?
The assignment of gray level to new locations is given by gray level of the previous location.
This process is known as nearest neighbor interpolation.
Halftone printing means larger or smaller filled circles of blacking for new paper printing. It is
represented in matrix format as,
2X2 matrix form 4X4 matrix form
48. Mention the difference between a monochrome and a grayscale image. [NOV 2013]
Monochrome Image Grayscale Image
Each bit is stored as a single bit as a ‘0’ or a ‘1’ Each pixel is stored as a byte with a value
between 0 and 255
49. Compare RGB and HSI color image models. [NOV 2014]
RGB Model HSI Model
RGB means Red, Green and Blue colour. HSI means Hue, Saturation, intensity of colour.
It represents color. It decides the type of colour.
It’s formed by either additive or subtractive It numerically represents the average of the
model. equivalent RGB value.
It’s subjective process. It’s objective process.
EC 8093 - DIP UNIT-1 DIGITAL IMAGE FUNDAMENTALS
52. Find the intensity component value in HIS color model for Black RGB image.(A/M
2021)
Th e intensity components is given by
𝑅+𝐺+𝐵
I= where R – Red, G – Green, B - Blue
3
53. What is the function of an image sensor? (May/June 17)
The purpose of the image sensor is to capture the light and convert it to electrical charges and
then to binary data.
The image sensors employed by most digital cameras, a CCD (Charge Coupled Device) how
ever some cameras use Complementary Metal Oxide Semiconductor (CMOS) technology
used.
54. When is fine and coarse sampling used? (Or) Define the characteristics of image
sampling. (May/June 17)
To achieve highest visual quality and at the same time the lowest memory requirement, we can
perform fine sampling of an image in the neighborhood of sharp grey level transition and coarse
sampling in the smooth area of an image.
56. What is the memory required to store a 512*512 RGB image of 8 bit resolution?
(APR/MAY-19)
Memory size = Horizontal pixels X Vertical pixels X resolution / 8 /1024
Memory Size = 5
EC 8093 - DIP UNIT-1 DIGITAL IMAGE FUNDAMENTALS
Step 5: Wavelets:
• Used in image data compression and pyramidal representation.
Step 6: Compression
• Techniques Compression for o reducing the storage required to save an image.
o Reducing the size of the image to transmit it ("JPEG Standard"), with suitable bandwidth
required for transmission.
2. Briefly explain the elements or components of image processing system? Explain the
function of each element in details (N/D-12)&(N-14)(May/June 17)(N/D-17)(N/D-18)
5) Mass storage:
It provides adequate storage of images, 1000 or million of images is stored. For example: if a
image of size 1024X1024 pixels in which the intensity of each pixel is a 8bit quantity, that
requires 1byte of storage space. It is possible in mass storage device.
There are three types of storage used in DIP.
i) Short time storage – for digital processing
ii) Online storage – for relatively fast recall.
iii) Archival storage – used for infrequent access. Example: magnetic tapes and optical disks.
6) Image display: Today image displays are TV color monitors, the monitors are driven by
the output of images. In some cases, it is necessary to have stereo displays.
7) Hard copy: It is used to record images. It includes laser printers, film cameras, heat
sensitive devices, ink jet printers such as optical and CD – ROM disk.
3. What is visual perception model and explain. How this is analogous to a DIP
system.(NOV/DEC 2014) (OR)Explain in detail about the elements of visual perception or
elements of human visual system? (A/M 2021)
Vision is the most advanced human sense. So images play the most important role in
human perception.
Human visual perception is important because the selection of image processing technique
based on visual judgments.
Cornea: It is a tough and transparent tissue which covers the anterior portion of the eye.
Sclera: Sclera is the continuous of the cornea which covers the remaining portion of the eye.
Choroid:
The choroid is directly below the sclera. It consists of networks of blood vessels which
bring nutrition to the eye.
Injury to choroid is highly pigmented and helps to reduce the amount of light entering into
eyes and avoid backscatter into the eye.
In the anterior of choroid is divided into ciliary body and Iris diaphragm.
It is diaphragm varies in diameter 2-8 mm.
EC 8093 - DIP UNIT-1 DIGITAL IMAGE FUNDAMENTALS
The front of Iris has visible pigment of the eye and back contains black pigment.
Eye lens:
The lens is made up of fibrous cells and it is suspended by fibers of ciliary body.
The lens contains 60 – 70% water and about 6% of fat and more proteins.
It is colored by yellow pigmentation that increases the vision of eye.
Excessive dust deposition in the lens will commonly leads to cataracts. The cataracts lead
to poor color discrimination and loss of clear vision.
The lens will absorb 8% of light visible spectrum of IR and UV rays are absorbed by
proteins of the lens.
Retina:
The innermost membrane of the eye is retina.
It occupies the posterior portion of the eye. When eye focus down the object the image is
formed in the retina.
The vision is formed in the Retina due to reception in it.
1) Cones:
There are 6 to 7 million of cones which are located in the central portion of the Retina
called fovia.
EC 8093 - DIP UNIT-1 DIGITAL IMAGE FUNDAMENTALS
It is highly sensitive to color and gives fine details of the image because it is directly
connected to the nerve.
This cone vision is called as photopic vision or bright – light vision.
2) Rods:
There are 75 to 150 million of Rods in the eye.
It gives only low details of the image. Because it is connected to single nerve. It is sensitive to
low levels of illumination. The Rod vision is known as scotopic vision or dim – light vision.
The area in which receptors are not present is known as blind spot.
Distribution of cones and Rods in the Retina:
The lens is controlled by ciliary fibers. The muscles of lens become thicker in order to
focus the nearer objects.
Focal length – The distance between centre of lens and retina is known as focal length. The
focal length varies from 14 to 17mm.
Image formation:
The eye focus the far objects the lens in the lowest refractive power. When the eye focus
the nearer object then lens has high refractive power.
Let us take an example; a man is looking at a tree which is 15m height at distance of 100m.
EC 8093 - DIP UNIT-1 DIGITAL IMAGE FUNDAMENTALS
Subjective brightness:
The subjective brightness is the logarithmic function of light incident on the eye.
DIP SYSTEM:
In the human eye, the lens has a fixed focal length and focusing at various distance is achieved by
varying the distance between the lens and imaginary plane. The focal length needed to achieve
proper focus is obtained by varying the shape of the lens.
Subjective brightness:
Brightness is a subjective description of light perception that is practically impossible to
measure.
It is a key factor in describing color sensation.
The subjective brightness is the logarithmic function of light incident on the eye.
EC 8093 - DIP UNIT-1 DIGITAL IMAGE FUNDAMENTALS
In fig, the long curve ‘A’ represents the range of intensities that the visual system can adapt.
The short intersecting curve (B—C) represent the range of subjective brightness at that level.
Range of Brightness:
In photopic vision, the range is about 106. The transition from the scotopic to photopic
vision is gradual over the approximate range from 0.001 to 0.1 milliambert.
Brightness discrimination: (Discrimination – treat something in a different way).
The eyes ability to discriminate (i.e.,) differentiate various intensity levels is very
important in presenting image processing.
Experimental setup:
This setup is used to characterize the brightness discrimination.
The setup is a flat area which is a diffuser such an s opaque glass. It is illuminated by a light
source with a variable light intensity, I.
Case (i):
The intensity of the field is increase to an amount ∆I.
The total illumination of the field = I +∆I
Case (ii) – In this case, the background illumination is kept constant and the other source
intensity is increase from very low to high range.
Mach band effect proved that perceived brightness is not a simple function of intensity. Although
the intensity of stripes is constant, we actually perceive the brightness pattern that has strongly
scalloped bands called as Mach band. This effect is called as mach band effect.
The three basic quantities that are used to describe the quality of a chromatic light source.
They are: - 1.Radiance 2. Luminance 3.Brightness
4. With a neat diagram explain image sensing and acquisition and also explain image
acquisition using sensors arrays (Nov/Dec 16)(APR/MAY-18)
The most common sensor of this type is the photodiode, which is constructed of silicon
materials and whose output voltage waveform is proportional to light.
The use of a filter in front of a sensor improves selectivity.
An arrangement used in high precision scanning, where a film negative is mounted onto a
drum whose mechanical rotation provides displacement in one dimension.
The single sensor is mounted on a lead screw that provides motion in the perpendicular
direction. Since mechanical motion can be controlled with high precision, this method is an
inexpensive (but slow) way to obtain high-resolution images.
EC 8093 - DIP UNIT-1 DIGITAL IMAGE FUNDAMENTALS
. Sensor strips mounted in a ring configuration are used in medical and industrial imaging
to obtain crosssectional (“slice”) images of 3-D objects.
This is the basis for medical and industrial computerized axial tomography (CAT) imaging.
CCD cameras: widely used in modern applications: private consumers, industry, astronomy…
CCD: Charge Couple Device
CCD vs CMOS
CCD: when exposure complete, transfers each pixel’s charge packet sequentially to a
common output structure, which converts the charge to a voltage, buffers it and sends it off-chip.
CMOS imager: the charge-to-voltage conversion takes place in each pixel
EC 8093 - DIP UNIT-1 DIGITAL IMAGE FUNDAMENTALS
(iv) A Simple Image Formation Model
Construction:
It is made up of a cylindrical glass envelope.
The envelope has an electron gun at one end and a faceplate at the other end.
The glass tube is surrounded by a yoke which contains two coils, namely
1) Focusing coil
2) Horizontal and vertical deflection coils.
EC 8093 - DIP UNIT-1 DIGITAL IMAGE FUNDAMENTALS
Working Principle:-
In the above fig, the target is a thin photoconductive layer of either selenium or antimony
compounds.
This is deposited on transparent conducting film on the inner surface of face plate. This
conductive coating is known as single electrode connected to DC supply through the load
resistance RL.
The beam from the electron gun is focused on surface of photoconductive layer by combine’s
action of uniform magnetic field of external coil and electrostatic field of grid no: 3.
Grid no: 4 provides deaccelerating field on the photoconductive layer so that the electron
becomes low velocity to prevent secondary emission.
Deflection of beam for scanning target is obtained by vertical and horizontal deflecting coils.
Charge image:
The photo layer has a thickness of about 0.0001cm.
Before light falls on it, it behaves like an insulator with resistance of 20MΩ (dark).
When light is focused the photon energy allows more electrons to go to conduction band.
When the right light falls, the resistance of that portion reduces to 2MΩ. Each point of the
gun side of photolayer varies with potential with respect to DC supply.
Storage action:
Even though the light falls continuously on target, each element of photo coating is
scanned at interval equal to frame time.
This result in storage action and change in resistance at any point of the photoconductive
layer depends on the time between two successive scanning and intensity of incident light.
The storage time for all point in the target is same and proportional to the light intensity.
Signal current:
After beam scans, the target plate it encounters the positive potential of gun faced
photolayer.
Then sufficient numbers of electrons are deposited on the target.
This sudden change of potential on each element while the beam scans causes a current
EC 8093 - DIP UNIT-1 DIGITAL IMAGE FUNDAMENTALS
flow in the signal electrode of voltage is varied across RL.
Consequently the output of RL is directly proportional to the light intensity variations.
Applications:
It is used in slides, pictures and closed circuit TV etc (fast movement).
It has wide application in educations, medicine, industry, aerospace and oceanography.
It is short tube with length 12 – 20cms of diameter 1.5 – 4 cm
Its lifetime is between 5000 to 20,000 hours.
5.2 Explain in detail about the working principle of digital camera?(Nov/Dec 16)
Components of digital camera:
To use a digital camera, the object and shoot to capture the image is same as conventional
camera.
The big difference is that the digital camera is not containing a role of film instead of it store
capture image in digital form.
The images are stored on a magnetic or optical disk.
Photo sites:-
The sensor consists of an array of these “tiny solar cells” which are light sensitive spots are
called photo sites.
Photo sites are usually square in shape and laid out in rows and columns.
The size of photo sites ranges from 3 to 11.8 microns square.
Once the sensor converts the light into electron, it reads the value of each cell in the image.
CCD:
A CCD transports the charge across the chip and reads it at one corner of the array.
ADC:
An analog to digital converter then turns each pixel values into a digital value by
measuring the amount of charge at each photo site and converting the measurement to
binary form.
The CMOS sensor uses several transistors to amplify and move the charge using more
traditional wires.
The CMOS signal is digital, it needs no ADC.
Working principle:
Although the principle may be same as film camera, the inner workings of digital camera
are different.
The imaging being performed either by a charge coupled device (CCD) or CMOS
(Complementary Metal Oxide Semiconductor) sensors.
Each sensor element converts light into a vigil proportional to the brightness which is
passed into an ADC which translates the fluctuations of the CCD into discrete binary code.
The digital output of ADC is sent to a digital signal processor (DSP) which adjusts contrast
and details, and compresses the image before sending it to storage medium.
The brighter the light – the higher the voltage of the brighter the resulting computer pixel.
The more elements – the higher the resolution of greater the detail can be captured.
The CCD or CMOS sensors are fixed in place and it can go on taking photos for the
lifetime of camera.
Saving photos:
PC serial port – now – USB mass storage used.
Common method – used is cellular networks method is MMS (Multimedia service).
Email attachment – send a picture.
Card reader – several types of media – high speed transfer data.
Printing photos:
EC 8093 - DIP UNIT-1 DIGITAL IMAGE FUNDAMENTALS
Thermal printer:
File formats:
Joint Photography Experts Group standard (JPEG), tagged image file format (TIFF), AVI, and
DV.
Batteries: Digital cameras have high power requirements, and overtime have become smaller,
resulting in an ongoing need to develop a battery to fit in the camera and to power it for a
reasonable length of time.
Uses:
Digital camera is a camera that takes video or still photographs, or both digitally by recording
images via sensor.
Many compact digital still cameras can record sound and moving video as well as still
photographs.
Digital cameras are of 35 mm film cameras.
Applications:-
With in the constraints of correct exposure various parameters can be changed including
exposure, aperture light metering, white balance and equivalent sensitivity.
6. Explain in detail about image sampling and Quantization?(Or)How the continuous image
can be converted into digital image using technique? (Or) Describe how image is digitized
by sampling and quantization & explain about checker board effect and false contouring
with neat sketch. Nov 11(MAY/JUNE-2011) (May 2015)(Nov/Dec 16)(OR) Explain in detail
about the phenomenon of image sampling. Illustrate how aliasing happen if sampling
theorem is violated (N/D-17)(APR/MAY-19)
To convert this continuous image into digital form we have to sample the function in both
coordinates and in amplitude.
Digitizing the amplitude value F is known as Quantization.
EC 8093 - DIP UNIT-1 DIGITAL IMAGE FUNDAMENTALS
Let us consider a following continuous image AB.
The amplitude value of above figure corresponds to the line segment AB as shown below.
To sample above function, take equally spaced squares that is samples each is assigned by a
vertical tip marks.
The generation of digital image from the sensed data is done by converting sensed data into
digital form. This can be done by two processes.
1. Sampling
2. Quantization
To convert the continuous image f(x, y) into digital form a gray level is added to the right side of
the samples. Each sample assigns one gray scale value.
Then continuous image is converted into digital form image with various intensity levels.
Representation of digital:-
The digital image has ‘H’ rows and ‘N’ columns. The digital image is represented by,
Origin 1 2 3 . . . . . . . . . N–1 y
0
. . . . .
1
. . . . .
2
. . . . .
. One pixel
. . . . .
.
. . . . .
N-1
EC 8093 - DIP UNIT-1 DIGITAL IMAGE FUNDAMENTALS
Image f(x, y)
x
Dynamic range:
The values given by the gray level scale is known as dynamic range.
If the dynamic range of pixel is high, then image will have high contrast. The image with
low dynamic range will have low contrast.
The spatial level resolution of 2w image is known as minute detail of the image.
A picture is divided into number of vertical lines. A pair of vertical line width is “1w” and
space between the vertical lines is “w”. Then width of the line pad is “2w”.
Aliasing effect:
If the function with unlimited duration is sampled at a rate equal or greater than twice its
highest frequency. Then it is possible to recover the original image.
If the sampling is not satisfied, then the function is under sampled then aliasing is
introduced in the image. The addition of all unwanted frequency components are called aliasing
frequencies.
Moiré pattern effect:
Consider a two identical periodic pattern of equally spaced vertical sheets as shown in the
figure above two patterns are rotating in opposite direction and superimposed by each other.
Then the multiplication of two images will give a new image. Hence the new image will
looks like a thin roof.
7. Explain in detail about color image fundamental? (Or)Explain about various color
models in image processing? Write short notes RGB color model. What is HIS color model?
How an RGB image is converted into HIS? Explain(APR/MAY-19)
Color models:-
Definition:-
A color model is a specification of a coordinate system and a subspace within that system
where each color is represented by a single point.
The color model can also be called as color space or color system. It is used to specify the
colors easily in a standard and accepted way.
Classification:
1. Hardware – oriented color models: RGB – (Red, Green, Blue) model for color monitors
and color video cameras
– CMY (Cyan, Magenta, yellow) model.
– HIS (Hue, Saturation, Intensity) model.
2. Application – oriented color models.
– One example is the creation of color graphics for animation. RGB color model, all the colors
appear in their primary spectral components of red, green and blue.
Color image processing is divided into two major areas: full – color and pseudo color processing.
1. RGB model
2. HSI model
In the first category, the images typically are required with a full – color sensor such as
color TV Camera.
In the second category, the problem is one of assigning a color to a particular monochrome
intensity or range of intensities.
Chrome light:
Three basic quantities are used to describe the quality of a chromatic light source.
1. Radiance
2. Luminance
3. Brightness.
CIE: for the purpose of standardization, (commission international del’ Eclairage the
international commission on illumination) in 1931.
Valid values of RGB:
Blue = 435.8 nm, Green = 543.1nm, Red = 7.nm.
EC 8093 - DIP UNIT-1 DIGITAL IMAGE FUNDAMENTALS
Red, blue green colors are at three corners and cyan, magenta and yellow (CMY) colors are at
three corners.
Black is at the origin. White is at the corner farthest from the origin.
All values of R, G and B are assumed to be normalized in the range [0, 1] and this color cube is a
unit cube.
Pixel depth:
The number of bits used to represent each pixel in the RGB space is called as “Pixel
depth”.
EG: - Pixel depth of each RGB color pixel
= 3x number of bits / plane
= 3x 8 = 24
Applications:
The application of RGB color model is: - 1. Color monitors 2. Video color cameras
etc.
white 60
Cyan Red
Blue
Magenta
1. RG sector (0≤H≤120˚)
B = I (1-S) HSI Model
And G = 3I – (R+B)
GB sector: (120˚≤H≤240˚)
H = H - 120˚
RGB component is:-
R = I (1-S)
B = 3I – (R+G)
BR Sector: (240˚≤H≤360˚)
H = H - 240˚
G = I (1-S)
R = 3I- (G+B)
Advantages of HSI Model:-
1. They describe color in terms that are suitable for human interpretation.
2. It can be used as an ideal tool for describing image processing algorithms based on color
description.
If pixel value is zero, we print nothing if pixel value is 4, we print all four dots. 4X4 dithered
matrix is given by,
+ - Display
Another method of suppressing contouring effects to add a small amount of uniformly distributed
pseudorandom noise to the luminance samples before quantization. this pseudorandom noise is
also called as “Dither”.
Colors such as pink (red + white) and lavender (violet + white) are less saturated. The pure
spectrum colors are fully saturated.
Hue and Saturation together called as chromaticity.
The amount of Red, green and blue needed to form any color is called as Tristimulus values. It is
denoted as x, y and z
1. Connectivity:
Connectivity between pixels is a fundamental concept of digital image processing.
Conditions:
Two pixels are said to be connected if a) they are neighbors and b) their gray levels satisfy a
specified similarity criterion (Eg: if their gray levels are equal).
Definition:
Let ‘S’ represents a subset of pixels in an image.
Two pixels p& q are said to be connected in ‘S’ if there exists a path between pand q which
consists of all pixels in S.
Connected components:
For any pixel in S, the set of pixel that are connected to P is called as connected components.
Connected set:
If the pixel p has only one connected component then set S is called as connected set.
2. Region:
Let R represents a subset of pixel in an image.
If R is a connected set, it is called a region of the image.
3. Boundary:
Boundary is also known as border or contour.
The boundary of a region is defined as the set of pixels in the region that has one or more
neighbors which are not in the same region.
EC 8093 - DIP UNIT-1 DIGITAL IMAGE FUNDAMENTALS
The boundary of the finite region forms a closed path. Therefore it is called as global
concept.
4. Edge:
Edges are formed by the pixels with derivative values that are higher than a present threshold.
Thus, edges are considered as gray-level or intensity discontinuities. Therefore it is called as local
concept.
6. Linear operation:
H is said to be a linear operator if, for any two images f and g and any two scalars a and b,
H (af bg) aH ( f ) bH (g)
7.Neighbors of a Pixel.
Three types of neighbors are defined for a pixel. They are
1. 4 Neighbors
2. D Neighbors
3. 8 Neighbors
EC 8093 - DIP UNIT-1 DIGITAL IMAGE FUNDAMENTALS
EC 8093 - DIP UNIT-1 DIGITAL IMAGE FUNDAMENTALS
8. Adjacency:
Two pixels are connected if they are neighbors and their gray levels satisfy some specified
criterion of similarity.
For example, in a binary image two pixels are connected if they are 4-neighbors and have same
value (0/1)
1. 4-adjacency: Two pixels p and q with values from v are 4-adjacent if q is in the set N4 (p).
2. 8-adjacency: Two pixels p and q with values from v are 8-adjacent if q is in the set N8 (p).
3. m-adjacency (mixed): two pixels p and q with values from v are m-adjacent if: q is in N4 (p)
orq is in ND (P)
9. Path:
A digital path (or curve) from pixel p with coordinate (x,y) to pixel q with coordinate (s,t) is a
sequence of distinct pixels with coordinates (x0, y0), (x1, y1), ..., (xn, yn), where (x0, y0)= (x,y),
(xn, yn)= (s,t)
The system is said to be FIR or IIR if its impulse response are finite or infinite region of support.
EC 8093 - DIP UNIT-1 DIGITAL IMAGE FUNDAMENTALS
3. Shift invariant:
A system is called spatially or shift invariant if the translational of input causes translation of
output.
For a shift invariant system:-
Z – Transform:-
9. Circulant matrices:-
A matrix ‘C’ is called as circulant matrix. If each of its row and columns is the circular shift of
previous row.
EC 8093 - DIP UNIT-1 DIGITAL IMAGE FUNDAMENTALS
Definition:
Image transforms refer to a class of unitary matrices used for representing images. One
dimensional discrete Fourier transform (1D – DFT) .
The original function f(x) can be obtained using the inverse DFT,
f(x) = ∑𝑢=0
𝑁−1 𝐹(𝑈)𝑒𝑖2𝜋𝑢𝑥/𝑁 for x= 0,1,2,……..N-1 ---------------------------- > (2)
The equation (1)&(2 )are discrete fourier transform pair.
If F(U,V) is given, f(x,y) can be obtained via the inverse DFT, given by the expression
f(x,y) =∑𝑀−1 ∑𝑁−1 ( 𝑢𝑥 𝑣𝑦
𝑖2𝜋( + ) for x= 0,1,2 .......... M-1;
𝑈=0 𝑉=0 𝐹 )
𝑈, 𝑉 𝑒 𝑀 𝑁
When images are sampled in a square array, M=N and the discrete transform pair in equation (3)
& (4) become 𝑢𝑥+𝑣𝑦
)
F(U,V) = ∑
1 𝑁−1 ∑𝑁−1 ( ) −𝑖2𝜋( for u, v= 0,1,2. . . . . .N-1; ----------------- > (5)
𝑓 𝑥, 𝑦 𝑒 𝑁
𝑁 𝑥=0 𝑦=0
and
f(x,y) =∑𝑁−1 ∑𝑁−1 ( 𝑖2𝜋(
𝑢𝑥+𝑣𝑦
) for x, y= 0,1,2. . . . . .N-1; --------------------- > (6)
𝑈=0 𝑉=0 𝐹 𝑈, 𝑉 )𝑒 𝑁
The 1/N term is included in equation (6) to compensate for the removal of the same in equation 6)
Properties of DFT:
The properties are
i. Separablity vi. rotation xi. average value
ii. convolution vii. Periodicity xii. correlation
iii. Linearity viii. Conjugate symmetry xiii. sampling
iv. Translation ix. Distributivity xiv. Laplacian
v. modulation x. scaling
i) Separablity
i) 2-D DFT pair can be expressed in the separable form as F(U,V )
1 for u, v= 0,1,2...........N-1;
F(U,V) = ∑𝑁−1 𝑒−𝑖2𝜋𝑢𝑥/𝑁 ∑𝑦=0
𝑁−1 𝑓(𝑥, 𝑦)𝑒−𝑖2𝜋𝑣𝑦/𝑁
𝑁 𝑥=0
ii) Convolution
It is an important fourier transform relationshipthat links the spaial and frequency domains.
Convolution in one domain is the equivalent of multiplication in other domain.
F[f(x,y)* g(x,y)] = F(U,V) G(U,V)
[f(x,y)* g(x,y)] = F-1 [F(U,V) G(U,V)]
[f(x,y)* g(x,y)] =𝐹−1 [F(U,V)* G(U,V)]
iv) Translation: It states that multiplying f(x,y) by the exponential term .if the image moved
from one place to another place, the resulting fourier spectrum changes with a phase shift but
the magnitude of the image remains same
𝑢𝑥0+𝑣𝑦0
F[f(x-x0, y-y0)] = F(U,V)𝑒 −𝑗2𝜋( 𝑁 )
f(x-x y-y ) = −1[F(U,V) −𝑗2𝜋(𝑢𝑥0+𝑣𝑦0)
0, 0 𝐹 )𝑒 𝑁 ]
vi) Rotation
This property states that rotating f(x,y) by an angle 𝜃0 rotates F(U,V) by the same angle and
rotating F(U,V) by some angle rotates by the same angle.
r=xcos𝜃, c=xsin𝜃, 𝑢 = 𝑤𝑐𝑜𝑠∅, 𝑣 = 𝑤𝑠𝑖𝑛∅
f(r, 𝜃 + 𝜃0)= [F(w, ∅ + ∅0)]
If the image is rotated by an angle𝜃 then F(U,V) is rotated by the same angle and viceversa.
vii) Periodicity
This property defines the implied symmetry of the fourier spectrum that results from certain
theoretical considerations.
F(U,V) = F(U+N,V)= F(U,V+N)= F(U+N ,V+N)
ix) Distributivity
The fourier transform and its inverse are distributive over addition but not over
multiplication
F[f(x,y) +g(x,y)]=F[f(x,y)] +f[g(x,y)]
F[f(x,y) .g(x,y)]≠F[f(x,y)] .f[g(x,y)]
x) Scaling: The scaling properties of DFT when two scalars a and b are used
a f(x,y) =a F(u,v)
f (ax,by)= 1 F(𝑈,𝑉)
𝑎𝑏 𝑎 𝑏
xiii) Sampling
It is defined as the process of dicretizing a continuous function.
xiv) Laplacian
It is used for outlining the edges of image.
2) 2D – DCT:-
The discrete cosine transform is also called cosine transform.it is mainly used for image
compression
Two dimensional discrete Cosine transform:-
MXN Cosine transform is defined as,
C = {c (u, v)}
1
α(u) (𝑜𝑟) α(v) = √ for u=0;x=0,………N-1
N
2
=√ for u=1, 2, 3; x=0….N-1
𝑁
Cosine transform is derived from fourier transform by assuming the function asmirrored about
orgin, making it is an even function which means symmetric about orgin
Properties:
1. Orthogonal property:
The cosine transform is real and orthogonal. C = C*, C-1 = CT
2. It is not the real part of unitary DFT. However the Cosine Transform of the sequence is
related to DFT of its symmetric.
3. Fast Fourier Transform:
The Cosine transform is calculated in O (N log 2N) operations.
It requires O (N log 2N) operations for odd and even sequences. But the operations of
summation of odd and even sequence into general sequence requires O (N2 log 2N)
operations.
4. It has excellent energy compaction for highly correlated data.
5. Basis image:The basis vectors of cosine transform are the eigen vectors of the
Symmetric Tridiagonal matrix is denoted by,
EC 8093 - DIP UNIT-1 DIGITAL IMAGE FUNDAMENTALS
Step 2: The Fourier spectrum of discrete function is given by 4*4 DFT kernel is
𝐹(0) 1 1 1 1 𝑓(0)
1 1 −𝑗 −1 𝑗
(
𝐹(1)
)= ( ) l𝑓(1)
𝐹(2) 4 1 −1 1 −1 𝑓(2)
𝐹(3) 1 𝑗 −1 −𝑗 𝗁𝑓(3))
12. Determine the DCT matrix for N=4. [NOV 2013]
DCT for N=4
(2𝑛+1)𝜋𝑘
X(k) = α(K) ∑𝑁−1 𝑥(𝑛) cos 0≤ 𝐾 ≤N-1 -------------------- > (1)
𝑛=0 2𝑁
1
α(K) = √ if k=0
𝑁
EC 8093 - DIP UNIT-1 DIGITAL IMAGE FUNDAMENTALS
𝛼
=√ if k 0
𝑘
for N=4
(2𝑛+1)𝜋𝑘
X(k)= α(K) ∑3𝑛=0 𝑥(𝑛) cos ---------------- > (2)
8
sub k=0,1,2,3
k=0
1
X (0) =√ ∑3 (𝑛) cos
(2𝑛+1)𝜋𝑘
4 𝑛=0 𝑥 8
= 1 ∑3 𝑥(𝑛)(1)
2 𝑛=0
= 1 [𝑥(0) + 𝑥(1) + 𝑥(2) + 𝑥(3)]
2
X (0) = 0.5 𝑥(0) + 0.5 𝑥(1) + 0.5 𝑥(2) + 0.5 𝑥(3)]
k=1
1
X (1) =√ ∑3 (𝑛) cos
(2𝑛+1)𝜋
2 𝑛=0 𝑥 8
1 𝜋 3𝜋 5𝜋 7𝜋
=√ {𝑥(0) cos + 𝑥(1) cos + 𝑥(2) cos + 𝑥(3) cos
2 8 8 8 8
X (1) = 0.6532𝑥(0) + 0.2706𝑥(1) − 0.2706 𝑥(2) − 0.6532𝑥(3)
k=2
1
X (2) =√ ∑3 𝑥(𝑛) cos
(2𝑛+1)𝜋2
2 𝑛=0 8
1
=√ ∑3 𝑥(𝑛) cos
(2𝑛+1)𝜋
2 𝑛=0 4
1 𝜋 3𝜋 5𝜋 7𝜋
=√ {𝑥(0) cos + 𝑥(1) cos + 𝑥(2) cos + 𝑥(3) cos
2 4 4 4 4
X (2) = 0.5𝑥(0) − 0.5𝑥(1) − 0.5𝑥(2) − 0.5𝑥(3)
k=3
1
X (3) =√ ∑3 ( )
(2𝑛+1)𝜋3
2 𝑛=0 𝑥 𝑛 cos 8
1
=√ ∑3 (𝑛) cos
(2𝑛+1)3𝜋
2 𝑛=0 𝑥 8
1 3𝜋 9𝜋 15𝜋 21𝜋
=√ {𝑥(0) cos + 𝑥(1) cos + 𝑥(2) cos + 𝑥(3) cos
2 8 8 8 8
X (3) = 0.2706𝑥(0) − 0.6533𝑥(1) + 0.6533𝑥(2) − 0.2706𝑥(3)
EC 8093 - DIP UNIT-1 DIGITAL IMAGE FUNDAMENTALS