Color Feature Based Object Localization in Real Time Implementation
Color Feature Based Object Localization in Real Time Implementation
Color Feature Based Object Localization in Real Time Implementation
e-ISSN: 2278-0661,p-ISSN: 2278-8727, Volume 20, Issue 2, Ver. III (Mar. - Apr. 2018), PP 31-37
www.iosrjournals.org
Abstract: One of the important task in pattern detection and recognition is object localization. This important
task is performed to reduce the searching time to the interest object. In this research we demonstrate our novel
method of object localization based on color feature. Our novel method is a combination of histogram of s-RGB
and histogram of Hue. In the training phase, we use these histograms to determine the color dominant in the
initial Region of Interest (ROI). Then this information is used to label the interest object. We apply the row and
column density function of pixels to reduce noise and localize the interest object. The comparison result with
some processes, our system gives a best result and takes a short computation time of 48.37 ms, in the video rate
of 15 frames per second (fps).
Keywords - color feature; histogram of s-RGBH; object localization; pattern recognition
----------------------------------------------------------------------------------------------------------------------------- ----------
Date of Submission: 02-04-2018 Date of acceptance: 17-04-2018
----------------------------------------------------------------------------------------------------------------------------- ----------
I. Introduction
One of the task in pattern detection and recognition is object localization. This important task is
required to reduce the searching time to the interest object. We don’t require to run a window to search the
object by scanning the whole of image frame. We only label the pixels in initial step to localize the interest
object based on the feature. There are some features implementing for this task, such as color feature, texture
feature, orientation feature and shape feature.
Some methods for object localization, such as matching method to match these objects with data stored
in a database [1][2]. Clustering algorithm as in [3][4]. Clustering methods require the several clusters as the
output of the system[5]. Segmentation system similar to [6]. The segmentation system or methods always need
an origin point for the algorithm. Spatial Data Analysis involves illustration such relationships between objects
based on their spatial properties i.e. positions, area coverage, density [7][8][9][10].
\Ozuysal et.al. [1] proposed a layered approach to object detection. They trained an estimator for the
bounding box dimensions by using Support Vector Machine (SVM) classifier. Object in their database rotated at
constant angular velocity and recovered its value by using the time of capture of a full rotation. This method had
a high precision both at 0.5 overlap and 0.7 overlap. Li et.al [2] proposed a two-step domain adaptation for
weakly supervised object localization: classification adaptation and detection adaptation. In the classification
adaptation step, they trained a classification network using the given weak image-level labels. They trained the
classification network to recognize the presence of a certain object category in an image. In the detection
adaptation step, they used the classification network to collect class-specific object proposals and applied
multiple instance learning to mine confident candidates.
Marszalek et.al. [3] used the Harris-Laplace or the Laplacian interest point detector to find a sparse set
of salient image features. A featured shape mask similarity was used to measure similarity between two features.
And to evaluate the shape masks, they used a bag-of-keypoints representation and a non-linear Support Vector
Machine (SVM) with x2 kernel. With this method, they were able to detect and localize object in vary position
well. However, they required to train some positive and negative samples. Bostanci et.al. [4] proposed
nethermost stretching trees which were often used for network routing so that minimize the number of hops for
packages. The graph in their algorithm was built using the 2D positions of the point features acquired by the
FAST algorithm. By this method, they were able to detect objects in stuck together. In this experiment,
however, they used a one color background.
Self-Taught Localization (in brief STL) proposed by Bazzani et.al. [5] was to generate bounding boxes
that are very likely to contain objects. Their proposed approach relied on the idea of covering out regions of an
image provided as input to a deep network. The reduction in recognition score caused by the covering out is
embedded into an agglomerative clustering method which merges regions for object localization.
A graph search algorithm was used by Sun et.al. [6] to split the graph of line fragments into smaller
ones using certain node and edge weighting functions. They encoded how likely a line fragment or line fragment
pair is to belong to an object. They used generic appearance and shape cues to score each cycle. These score are
not constricted to linearly additive measures and can easily yield multiple hypotheses that share some edges.
Sigal [7] initialized object localization by performing object detection. In object detection, he
performed object’s part detection before detected the whole of object. He implemented graphical models and its
inference method to detect the object’s part and loose-limbed body model to localize the object. These method
were able to detect and localize in multi-frame for single target and single frame for multi-target. However, it
required a complex computation. Murphy et.al. [8] implemented feature dictionary with the bank of 13 filters
and patch classifier of gentleBoost algorithm as local image feature. And gist of an image was as global image
feature. With this method, they were able to detect and localize object well. However, they implemented the
standard technique of object detection using sliding window classifiers applied to local features which required a
large computation time. Leitner et.al. [9] investigated a novel approach to spatial perception. Their method
combined the two calibration tasks, defining the camera parameters and precise kinematics, into one machine
learning problem, removing the need for prior calibration. They applied machine learning approaches: Artificial
Neural Networks (ANN) and Genetic Programming (GP). As the result, the accuracy achievement of ANN
technique was better than GP.
To localize a head as object from background, Mudjirahardjo et al [10] used Euclidean distance
function of the determined pixels as object and the surrounding pixels. The pixel was a part of object, when the
distance was less than a threshold. They computed the distance in RGB and HSV color space. Mudjirahardjo et
al. [11] extracted the interest object in dynamic background by using velocity histogram based on Harris corner
detector, Lucas-Kanade tracker and shift histogram based on motion history image (MHI).
In this paper, we demonstrate and evaluate our novel method. Our novel method is segmentation
method. It doesn’t require artificial intelligence as learning algorithm.
III. Method
This section explains our method to develop the object localization. First, we initialize the region of
interest (ROI) of the object manually in the first frame. Second, we determine the dominant color based on
histogram of s-RGB and histogram of Hue. Third, we calculate the density function of pixels to localize the
object.
s RGB ( x, y ) I R ( x, y ) I G ( x, y ) I B ( x, y ) (1)
where s-RGB(x,y) is sum of RGB intensity at pixel coordinate (x,y), IR(x,y), IG(x,y), and IB(x,y) are red, green and
blue intensity at pixel coordinate (x,y), respectively. When we use 8 bit to code a color intensity, then we can get
the s-RGB value of 0-765. Second, we divide the s-RGB value into 16 bins, then create the histogram of s-RGB
as shown in Fig. 3.
Third, from this histogram we determine modus bin of s-RGB by using equation (2),
mods-RGB arg maxhistogram _ s RGB (2)
bin
This modus value restricts the pixels having bin of s-RGB for further processing.
I P ( the interest object) if mod s RGB th1 I s RGB ( x, y ) mod s RGB th1
AND mod Hue th2 I Hue ( x, y ) mod Hue th2
(8)
pixel
I (non interest object) otherwise
P
where th1 and th2 are s-RGB threshold and Hue threshold value respectively; Is-RGB(x,y) and IHue are intensity of
s-RGB and Hue at pixel coordinate (x,y) respectively.
The row density function, fy, and its properties is shown in Fig. 4. By applying a threshold, d, we can determine
two coordinates of row. This calculation is also applied for column density function, fx, to determine two
coordinates of column. These two coordinates of row and two coordinates of column are used to localize the
interest object.
V. Conclusion
In this paper we use s-RGBH feature as color feature to determine the color dominant in RGB and
HSV space to label the interest object. Training phase is performed in the first frame in the initial ROI. Together
with row and column density function of pixels, we develop an object localization. The row and column density
function effectively to reduce noise. The comparison result with some processes, our system gives a best result
and takes a short computation time of 48.37 ms, as suitable for real time application.
To improve the object detection and localization, our future work is to add another feature, such as
texture feature and shape feature into our system.
Statement
In this research, I involved author’s face as the object of this study, as mentioned it in section 4.
Acknowledgments
This work was supported by DRPM Ditjen Penguatan Riset dan Pengembangan Kemristekdikti Indonesia, Grant
No. 063/SP2H/LT/DRPM/IV/2017.
References
[1]. M. Ozuysal, V. Lepetit, P. Fua. Post Estimation for Category Specific Multiview Object Localization. Proceeding of Computer
Vision and Pattern Recognition (CVPR2009), (Florida, USA. 2009) pp. 1-8.
[2]. D. Li, J.B. Huang, Y. Li, S. Wang, M.H. Yang. Weakly Supervised Object Localization with Progressive Domain Adaptation.
Proceeding of Computer Vision and Pattern Recognition (CVPR 2016), (Las Vegas, USA. 2016) pp. 3512-3520.
[3]. M. Marszalek, C. Schmid. Accurate Object Localization with Shape Masks. Proceeding of Computer Vision and Pattern
Recognition (CVPR 2007), (Minnesota, USA. 2007) pp. 1-8.
[4]. E. Bostanci, B. Bostanci. Object Localization and Spatial Analysis Using Computer Vision. International Journal of Machine
Learning and Computing, Vol. 1, No. 2, June 2011. Pp. 1-5.
IOSR Journal of Computer Engineering (IOSR-JCE) is UGC approved Journal with Sl. No.
5019, Journal no. 49102.