Indian Sign Language Converter Using Convolutional Neural Networks
Indian Sign Language Converter Using Convolutional Neural Networks
Indian Sign Language Converter Using Convolutional Neural Networks
2
C. Image Resizing these regions again. Although the segmentation accuracy is
In digital imaging and graphics, image resizing is done to not 100% in our case, and some regions are marked wrongly,
resize a digital image. We can change the total number of the result provided by the algorithm is sufficient enough to
pixels using image resizing. We have resized each image to a give us a good result while classifying it. Hard labelling has
resolution of 224x224px. We have used python software in been avoided in order to give us a quicker result and for the
order to achieve this. This has been done in order to reduce whole process to be conducted automatically as well.
the computational time as well as to have a uniform dataset to
be used for training.
D. Image Flipping
A flipped image or reversed image is basically the image
that is produced due to the mirror-reversal of the original
image across the vertical or horizontal axis. Image flipping is
done because the webcam automatically flips the images
while capturing it. Hence, flipping across the vertical axis is
required once again to regain the original photo.
3
After obtaining a good accuracy of 96% in the testing
N 85% Moderate
phase, we further applied the algorithm on 20 real time
images for each sign. Since there were a total of 26 signs, the
total number of real time images taken into consideration
were 20×26=520. O 90% Good
TABLE 1.
A 95% Good
Q 80% Moderate
B 90% Good
R 75% Poor
C 95% Good
S 85% Moderate
D 85% Moderate
T 75% Poor
E 85% Moderate
U 100% Excellent
F 80% Moderate
V 100% Excellent
G 95% Good
W 95% Good
H 90% Good
X 90% Good
I 100% Excellent
Y 85% Moderate
J 80% Moderate
Z 90% Good
K 75% Poor
4
TABLE II. CONFUSION MATRIX FOR THE 26 INDIAN SIGN LETTERS segmentation approaches can also be used in order to
completely segment the hand from the rest of the image.
REFERENCES
[1] R.M. Gurav, P.K. Kadbe, “Real time finger tracking and contour detection
for gesture recognition using opencv,” International Conference on
Industrial Instrumentation and Control 2015 (ICIC 2015), pp. 974-977,
2015.
[2] Pinaki Pratim Acharjya, Ritaban Das & Dibyendu Ghoshal, “Study
and Comparison of Different Edge Detectors for Image Segmentation,”
2012, Global Journal of Computer Science and Technology Graphics
& Vision, Volume 12, Issue 13, Version 1.0, Year 2012.
[3] Y.Ramadevi, T.Sridevi, B.Poornima, B.Kalyani, “Segmentation and object
recognition using edge detection techniques,” 2010, International Journal
of Computer Science & Information Technology (IJCSIT), Vol 2,No
6, December 2010.
[4] Pranali Loke, Juilee Paranjpe, Sayli Bhabal, Ketan Kanere, “Indian Sign
Language Converter System Using An Android App,” 2017, International
Conference on Electronics, Communication and Aerospace Technology-
ICECA, 978-1-5090-5686-6/17.
Table 2 depicts the confusion matrix for the 26 Indian sign [5] Muttaki Hasan, Tanvir Hossain Sajib and Mrinmoy Dey, “A Machine
letters. The 520 real time images have been considered in the Learning Based Approach for the Detection and Recognition of Bangla
confusion matrix. Sign Language,” IEEE - 978-1-5090-5421-3/16.
[6] Farhad Yasir, P.W.C. Prasad and Abir Alsadoon, “SIFT Based Approach
on Bangla Sign Language Recognition,” IEEE 8th International Workshop
V. CONCLUSION AND FUTURE WORK on Computational Intelligence and Applications, November 6-7, 2015.
[7] S. Karishma, V. Lathasree, “Fusion of skin color detection and background
In this paper, we have proposed a method for the subtraction for hand gesture segmentation,” International Journal of
recognition and classification of the 26 Indian sign language Engineering Research and Technology, vol. 3, no. 2, 2014.
alphabets by using CNNs. We observed that MobileNet is an [8] M. K. Ahuja and A. Singh, “Static vision-based hand gesture recognition
using principal component analysis,” in IEEE 3rd International Conference
effective approach for the classification of large amounts of on MOOCs, Innovation and Technology in Education 2015(MITE 2015),
data. We considered previous work for the recognition of sign IEEE, 2015, pp. 402–406.
language and came to a conclusion that MobileNet would be [9] Malladi Sai Phani Kumar, Veerapalli Lathasree and S.N. Karishma,
efficient to classify the hand signs with a high accuracy. “Novel Contour Based Detection and GrabCut Segmentation for Sign
Language Recognition,” International Conference on Wireless
However, letters H and J are dynamic gestures in the Indian Communications, Signal Processing and Networking (WiSPNET)-IEEE,
Sign Language whereas our method is applicable for static 978-1-5090-4442-9/17.
gesture recognition. Hence, these letters have been denoted by [10] Andrew G. Howard, Menglong Zhu, Bo Chen, Dmitry Kalenichenko,
Weijun Wang, Tobias Weyand, Marco Andreetto, Hartwig Adam,
a single frame. In the future, we would like to click the real “MobileNets: Efficient Convolutional Neural Networks for Mobile Vision
time images with a cellular camera instead of the webcam Applications,” arXiv:1704.04861v1 [cs.CV] 17 Apr 2017.
and get the output on the cell phone itself, since MobileNet is [11] C. Rother, V. Kolmogorov, and A. Blake, “Grabcut: Interactive foreground
extraction using iterated graph cuts,” in ACM Transactions on Graphics
especially suited for cellular operations [10]. Better (TOG), vol. 23, ACM, 2004, pp. 309–314.