Face Identification Using Kinect Technology: Ana-Andreea Nagâţ, Cătălin-Daniel Căleanu
Face Identification Using Kinect Technology: Ana-Andreea Nagâţ, Cătălin-Daniel Căleanu
Face Identification Using Kinect Technology: Ana-Andreea Nagâţ, Cătălin-Daniel Căleanu
=
=
M
n
n
M
1
1
and
n
are the eigenfaces:
M n A
n
M
k
k nk n
,..., 1 ,
1
= = =
=
(2)
The weights form a vector [ ]
' 2 1
... , ,
M
T
= that
describes the contribution of each eigenface in
representing the input face image. Further, the method
determines which face class provides the best description
of an input face image by calculating the face class k that
minimizes the Euclidian distance:
( )
2 2
k k
= (3)
where
k
is a vector describing the kth face class.
It will always return the most similar object. It has as
parameters the images used for training, each of them
having the same size, the labels corresponding to the
images and the criteria for recognizer training. The
algorithm returns the name of the person form the
database having the most similarities with the one in front
of the sensor. If the person is not recognized (not enough
similarities are found) the algorithm returns an empty
string, which means the person is unknown. In order to
increase the recognition rate, multiple faces (from
different angles) of the same person could be added to the
database.
IV. EXPERIMENTAL EVALUATION
The human identity recognition application was
implemented using Microsoft Visual Studio 2010 and
Microsoft Kinect SDK v1.4. In the following we will refer
only to the face based identification part of our system.
The first component of the application displays Kinect
data as following: color data (the top-left corner), depth
data (the top-right corner), skeletal tracking data (the
bottom-left corner), sensors elevation angle (the bottom-
right corner), as is depicted in fig .2.
When the person is unknown, a label named
Unknown person appear above the persons head. The
person remains unknown until it is added into the data
based and the recognition algorithm is started (fig. 3).
In order to add an unknown person to the database, its
name must be filled in the corresponding text box and the
Add face button should be pressed. Now the person
exists in the database and can be recognized by the
A.Nagt and C.-D. Caleanu Face Identification Using Kinect Technology
170
application (fig. 4). The name of the recognized person
appears above its head but also in the right side of the
application.
It is very important that the person to be recognized to
stay at the optimum distance from the sensor, that is
between 1m and 3m. The optimum distance is about 1.5
m.
The application shows real time capabilities when
running on a mid-level laptop, HP Compaq 8510w,
configured as Intel Core 2 Duo T7500 processor, 2GB
DDR2 RAM and 256MB NVIDIA Quadro FX 570M
GPU with Microsoft Windows 7 32 bit.
Video sequence and the final application could be
downloaded from http://www.ea.etc.upt.ro/Kinect.html
V. CONCLUSION AND FUTURE RESEARCH DIRECTIONS
This paper has presented a possibility to implement a
real time system performing face based human
identification using the Kinect technology. A solution for
obtaining the face images using the combination of the 3D
information provided by the skeletal tracker and 2D
information provided by the RGB camera was proposed.
Then, an appearance-based method (eigenface - a PCA
approach) in conjunction with a distance based classifier
solution was chosen for implementing the feature
extraction and classification stages. In order to perform
these operations, a .NET wrapper was employed to enable
calling OpenCV image processing library functions.
Future work should address the possibility to use the
newly introduced Microsoft Face Tracking SDK. Using it
is possible to calculate the head pose and the face
expression in real time (fig. 5).
We would also like to improve the classification stage
by following the recent trends in computational
intelligence: the use of biologically inspired architectures,
e.g. reservoir computing [31], liquid state machines
(LMS) [32] and echo state networks (ESN) [33]. For
example, [34] and [35] present a face recognition/facial
expression application that uses ESN and LSM
architectures respectively, achieving high recognition rates
and robustness to noise. These approaches use either 2D
or 3D information in recognition.
Lastly, we would like to include other biometric traits
(skeleton, voice, hair color, eye color and skin color, as
well as the existence of beard, moustache and glasses) for
further increasing the human identification accuracy [37].
Figure 2. Kinect data acquisition.
Figure 3. Detected person not present in the facial database.
Figure 4. Successfully recognized person. In order to increase the
recognition rate, multiple faces of the same person could be added to the
database.
171
8th IEEE International Symposium on Applied Computational Intelligence and Informatics May 2325, 2013 Timisoara, Romania
Figure 5. SDK Face Tracking offering a Candide-3 3D face model
[36].
REFERENCES
[1] J. Yang, L. Nanni (Eds.), State of the art in Biometrics, InTech,
July, 2011.
[2] W. Lin, Video Surveillance, InTech, 2011.
[3] KinectTM for Windows, http://www.microsoft.com/en-
us/kinectforwindows
[4] W. Zhao, R. Chellappa, Face Processing: Advanced Modelling
and Methods, Academic Press, 2005.
[5] Z. Lei, S. Liao, M. Pietika inen, S.Z. Li, Face Recognition by
Exploring Information Jointly in Space, Scale and Orientation,
Image Processing, IEEE Transactions on , vol. 20, no. 1, pp. 247-
256, Jan. 2011.
[6] L. A. Schwarz, A. Mkhitaryan, D. Mateus, N. Navab, Estimating
Human 3D Pose from Time-of-Flight Images Based on Geodesic
Distances and Optical Flow IEEE Conference on Automatic Face
and Gesture Recognition (FG), Santa Barbara, USA, March 2011.
[7] G. Medioni and R. Waupotitsch, Face recognition and modelling
in 3D, IEEE International Workshop on Analysis and Modelling
of Faces and Gestures (AMFG 2003), October 2003, pp. 232233.
[8] C. Hesher, A. Srivastava, and G. Erlebacher, A novel technique
for face recognition using range images, Seventh International
Symposium on Signal Processing and Its Applications, 2003.
[9] R. Labayrade, D. Aubert, J.-P. Tarel, "Real time obstacle detection
in stereovision on non flat road geometry through "v-disparity"
representation", Intelligent Vehicle Symposium, vol.2, 17-21 June
2002, pp. 646- 651.
[10] T. Lyes, K. Hawick, Implementing stereo vision of gpu-
accelerated scientific simulations using commodity hardware, in
Proc. International Conference on Computer Graphics and Virtual
Reality (CGVR11). Number CGV4047, Las Vegas, USA, 2011,
pp76-82.
[11] J. Wang, P. Huang, C. Chen, W. Gu, J. Chu, "Stereovision aided
navigation of an Autonomous Surface Vehicle," Advanced
Computer Control (ICACC), 2011 3rd International Conference
on, 18-20 Jan. 2011, pp.130-133.
[12] E. N. K. Kollorz, J, Penne, J, Hornegger, A, Barke, Gesture
recognition with a Time-Of-Flight camera, International Journal
of Intelligent Systems Technologies and Applications, vol. 5, pp.
334-343, no 3/4, doi 10.1504/IJISTA.2008.021296, 2008.
[13] V. den Bergh, L.Van Gool, Combining RGB and ToF cameras
for real-time 3D hand gesture interaction , Applications of
Computer Vision (WACV), 2011 IEEE Workshop.
[14] P. Breuer, C. Eckes, S Muller, Hand Gesture Recognition with a
novel IR Time-of-Flight Range Camera - A pilot study,
Proceedings of Mi-rage 2007, Computer Vision / Computer
Graphics Collaboration Techniques and Applications, pp. 247
260.
[15] [PMD2012] PMD[vision] CamCube 3.0,
http://www.pmdtec.com/products-services/pmdvisionr-
cameras/pmdvisionr-camcube-30
[16] J. Salvi, S. Fernandez, T. Pribanic, X. Llado, A state of the art in
structured light patterns for surface profilometry, Pattern
Recognition, vol. 43, no. 8, pp. 2666-2680, August 2010.
[17] D. Scharstein, R. Szeliski, High-Accuracy Stereo Depth Maps
Using Structured Light, in Proceedings of the IEEE Conference
on Computer Vision and Pattern Recognition, 2003, pp. 195-202.
[18] M. Gupta, A. Agrawal, A. Veeraraghavan, S.G. Narasimhan,
"Structured light 3D scanning in the presence of global
illumination," Computer Vision and Pattern Recognition (CVPR),
2011 IEEE Conference on, 20-25 June 2011 pp.713-720.
[19] J. Kramer, N. Burrus, F. Echtler, D. Herrera, M. Parker, Hacking
the Kinect, Apress, 2012.
[20] S. Kean, Y. C. Hall, P. Perry, Meet the Kinect, Apress, 2011.
[21] J. Webb, J. Ashley, Beginning Kinect Programming with
Microsoft Kinect SDK, Apress, 2012.
[22] D. Catuhe, Programming with the Kinect for Windows Software
Development Kit, Microsoft Press, 2012.
[23] Kinect SDK, http://www.microsoft.com/en-
us/kinectforwindows/develop/developer-downloads.aspx
[24] T. Acharya and A. Ray, Image Processing: Principles and
Applications, New York: Wiley, 2005.
[25] Emgu CV, http://www.emgu.com/wiki/index.php/Main_Page
[26] G. Bradski, A. Kaebler, Learning OpenCV, OReilly, 2008
[27] G. Bradski, Using the Kinect Depth Camera with OpenCV,
http://opencv.willowgarage.com/wiki/Kinect
[28] D. Solis, Illustrated C# 2012, Apress, 2012.
[29] M. Turk and A. Pentland, Eigenfaces for recognition, Journal of
Cognitive Neuroscience, vol.3, No.1, 1991.
[30] M. Turk and A. Pentland, Face recognition using eigenfaces,
Proc. IEEE Conf. on Computer Vision and Pattern Recognition,
1991, pp. 586-591.
[31] K. Vandoorne, J. Dambre, D. Verstraeten, B. Schrauwen, P.
Bienstman, Parallel Reservoir Computing Using Optical
Amplifiers, Neural Networks, IEEE Transactions on, vol. 22, no.
9, pp.1469-1481, 2011.
[32] [H. Jaeger, W. Maass, J. Principe, Introduction to the special
issue on echo state networks and liquid state machines, Neural
Networks, vol. 20, no. 3, pp. 287-289, 2007.
[33] S.P. Chatzis, Y. Demiris, Echo State Gaussian Process, Neural
Networks, IEEE Transactions on, vol.22, no.9, pp.1435-1445,
Sept. 2011.
[34] A. Woodward, T. Ikegami, A Reservoir Computing approach to
Image Classification using Coupled Echo State and Back-
Propagation Neural Networks, IVCNZ2011, International
Conference on Image and Vision Computing New Zealand, 2011.
[35] B. J. Grzyb, E. Chinellato, G. M. Wojcik, A. Kaminski, Facial
Expression Recognition based on Liquid State Machines built of
Alternative Neuron Models, Proceedings of International Joint
Conference on Neural Networks, Atlanta, Georgia, USA, June 14-
19, 2009.
[36] J. Ahlberg, CANDIDE - a parameterized face,
http://www.icg.isy.liu.se/candide/
[37] A. Dantcheva, J-L Dugelay, P. Elia, Person Recognition using a
bag of facial soft biometrics (BOFSB), MMSP 2010, IEEE
International Workshop on Multimedia Signal Processing, October
4-6, 2010.
A.Nagt and C.-D. Caleanu Face Identification Using Kinect Technology
172