Paper 4
Paper 4
Paper 4
Abstract: In this work, a novel model for recognition of transfer learning heads to the use of labeled data of successful
handwritten Kannada characters using transfer learning from intelligent systems for the training purpose of new systems
Devanagari handwritten recognition system is presented. The with similar characteristics that has a smaller data corpus[2],
objective is to use the knowledge of large data corpus of
several attempts using transfer learning by Lee[3],
Devanagari recognition system as training data to perform the
Navarretta[4], Cao[5], Zoph[6] proved to be successful in
recognition of handwritten Kannada characters that has a smaller
data corpus. The transfer of knowledge for recognition is carried
realization of higher accuracies. Therefore in this paper, an
out using deep learning network architecture to VGG19 NET. attempt made towards using the labelled dataset of Devanagiri
VGG19 NET is defined of five blocks of hidden layers, two dense character recognition system toward Kannada handwritten
fully connected layers and an output layer. Each block (except recognition system.
block1) consists of four convolution layers along with a max Kannada handwritten character recognition is one of
pooling layer. In proposed classification framework, Devanagari the challenging and unresolved research problems. In spite of,
character set consists of totally 92000 images with 46 classes and a couple of attempts based on machine learning architectural
Kannada character set is built with 81654 for training and 9401 for
models Chacko[7], Coates[8], Deng [9] and quite a few deep
testing, for about 188 classes with each class comprising of 200-500
learning architectures, Acharya [10], Chen [11], Wu[12]
sample image. A total of 1,23,654 data samples is employed for
training with VGG19 NET. For experimentation 9401 samples of
limitations of these system remained unsolved. The problem
about 188 classes built of about 40-100 samples in each classes is of handwritten character recognition system is a complex
used and for which accuracy close to 90% is achieved. Validated problem irrespective of script type. The various barriers that
accuracy after evaluation in 10 epochs with VGG19 NET, it has comes in the way of achieving higher recognition rates
recorded an accuracy of 73.51% with a loss of 16.18%. includes indefinite number of handwriting styles,
unconstrained environment maintained while writing, varied
Keywords— Deep learning, neural network, transfer orientation in writing etc. Along with these, the complexity of
learning, Kannada character recognition, handwritten South Indian script Kannada, its consonants/vowels modifiers,
characters recognition, handwritten characters, convolution compound characters etc., improves the odds of these systems.
neural networks. Along with this availability of small size data corpus is one of
the drawback for these types of system. Thus in proposed
I. INTRODUCTION
work, the uncertainty issue of datasets is handled using
transfer learning model based on Devanagari character
In the recent times, evolution of deep learning system is found
recognition systems. In the recent years there are several
dominantly in most of successful intelligent systems. The idea
significant number of researches that are reported in this area,
of transfer learning has been integrated from the problem of
a detailed review of these works are discussed subsequently.
large datasets requirements for deep learning systems. Though
Siddiqua et al[13] proposed a method for Kannada character
deep learning systems are very powerful and robust towards
recognition in scene images using transfer learning. The
achieving higher accuracies these systems are dependent on
numbers of neural networks are trained and tested using 1700
very large datasets with proper labelling of training data [1].
Kannada scene characters that are extracted from the char74k
This requirement of large data corpus for its implementation
datasets. The modifications are made to Alexnet by placing
has restrained many research attempts. Later, the notion of
Authorized licensed use limited to: UNIVERSITY OF ROCHESTER. Downloaded on September 21,2020 at 15:43:04 UTC from IEEE Xplore. Restrictions apply.
Proceedings of the Second International Conference on Inventive Research in Computing Applications (ICIRCA-2020)
IEEE Xplore Part Number: CFP20N67-ART; ISBN: 978-1-7281-5374-2
batch normalization that results in accuracy of 96%. The [19] proposed the model the transfer knowledge from one
experimentations carried out with VGG19 for better recognition system to another for handwritten number
performance. The authors have focused only on recognition recognition in scripts including Hindi, Arabic, and Bangla.
First the system is trained with numerals from the three one-
Kannada aksharamala characters. Kunte [14] explored an idea
of-a-kind scripts and the best resulting weights is saved. In
to handle different font sizes and types to recognize Kannada addition, because of over fitting problem the CNN layers are
characters in printed document. The pre-processing methods paused for weight updation. Meanwhile, classification layers
are applied to remove noise, skew corrections and binarized. used feature vectors acquired from another language to detect
Horizontal, vertical projection profile and connected distinctive digits. In this work convolutional neural network is
components segments the document image into line, words used for transfer learning which results in decreased time and
and characters. Hu’s invariant moments and Zernike moments competitive accuracy and also it indicates that Independence
of model-particular training reduces the re-training time within
are extracted which are further applied to Radial basis function
the target task notably.
neural classifiers. The experiments are performed with 50 In addition, Aneja et al [20] carried out comparisons of pre-
samples as training and 20 document samples as testing of trained models for handwritten recognition of Devanagari
about 2500 characters in each class. A method for character alphabets using transfer learning for Deep Convolution Neural
recognition from historical handwritten Kannada documents Network (DCNN). This work implements AlexNet, VGG,
using transfer learning proposed by Chandrakala et al [15] . Inception ConvNet and DenseNet as fixed feature extractors.
The pre-processing approaches are applied to enhance contrast The experiments performed outcomes that InceptionV3
and to remove the degradations and images are semi performs better accuracy than the AlexNet because of
automatically segmented. For feature extraction, stochastic regularization imposed by smaller convolution filter sizes and
gradient descent with momentum (SGDM) algorithm is used. highest number of layers. The DenseNet model accomplished
The experiments are carried out to compare the performances poorest because of the structure of its architecture. Tang et al
of SVM and DCCN. The dataset considered of 1260 character [21] proposed CNN-primarily based transfer learning
images that belongs to 118 different classes. Text-Line approach applied on ancient Chinese character recognition.
segmentation is proposed for the historical document images First of all, a CNN model is trained by printed Chinese
by Ravi et al [16]. Three methods are proposed by applying character samples, and this network structure and weights are
horizontal projection and connected components. The method utilized to initialize another CNN model which recognizes the
3 presents good results with connected components and historical characters. The new model is refined by a few
bounding box that are applied to the sample images consists of historical or handwritten Chinese character samples and used
217 lines from which 178 correctly extracted. Cireşan et al to test the target characters. Transfer learning with CNN fine-
[17] explored an idea on character recognition tasks of Latin tuning shows the use of an additional adaptation layer, the
and Chinese characters by transfer learning with Deep Neural amount of samples selected for transfer learning from the
Networks (DNN). Minimal retraining is employed to target domain, the selection of samples in the source domain
recognize uppercase letters from the DNN trained on digits. and the schemes for updating the network parameters. This
Work focused on training weights to all layers except the last work concluded that it is viable to combine traditional transfer
layer that are utilized for different task by transferring learning methods with CNN-based transfer learning method.
knowledge from already learnt task and further for any new Oquab et al., [22] simple transfer learning procedure are
task prediction only retraining of classification is required. applied on challenging benchmark datasets of relatively
Containing similar structures that differentiate each with smaller size .The author has additionally demonstrated that
structures such as horizontal lines and dots to prevent over
high potential of the mid-level extracted features from an
fitting of these networks and improving accuracy even for a
varied and challenging dataset by using Dropout layer and Image Net-trained CNNs. Although the performance of this
data increment model Zhao et al [18] came up with the model increased as augmentation carried out on the source
method of transfer learning for the deep learning problem task data using only 12% of the Image Net corpus results in
with a small sample datasets and a version of the convolution the higher outcomes on the Pascal VOC 2012 datasets
neural network combined with global average pooling classification and recognition challenges. Asha et al [23]
(TLCNN-GAP) based on transfer learning. Transferring the proposed character recognition system for documents written
knowledge from a pre-trained CNN model on a large data set
in Kannada language. Pre-processing is carried out to enhance
to a small sample dataset with the aid of adjusting the full-pool
layer, convolution layers and the softmax classifier .This documents by noise removal and contrast enhancement
method of TLCNN-GAP efficaciously reduces range of techniques. Boundaries are applied to each line of the
parameters passed and optimizes generalization capacity of document, vertical segmentation is performed to segment each
network and stability by reducing the over-fitting. Tushar et al. word and further segmentation is carried out for character
Authorized licensed use limited to: UNIVERSITY OF ROCHESTER. Downloaded on September 21,2020 at 15:43:04 UTC from IEEE Xplore. Restrictions apply.
Proceedings of the Second International Conference on Inventive Research in Computing Applications (ICIRCA-2020)
IEEE Xplore Part Number: CFP20N67-ART; ISBN: 978-1-7281-5374-2
extraction from the word. For features extraction The idea of adapting datasets of a successful and efficient
Convolutional Neural Network (CNN) model is used on system through transfer learning to carry out recognition of a
Chars74K dataset and achieved 98% accuracy for the system with inadequate data sets is the inclination of proposed
document containing non-overlapping lines of characters. work. The idea of using the knowledge from deep learning
It is noticed in most of the works the number of classes to model to build another system through transfer learning is
which data is labelled is limited and focused on Roman scripts shown in figure 2, which represents the block diagram of
dominantly than other south Indian scripts. A common one of proposed system. In this work, feature extraction and
the measure or crucial parameter that can improvise the classification is carried out using the VGG NET 19, which is a
recognition accuracies of improvise and uncertain system like 19 layer deep convolutional neural network. VGG Net 19 has
handwritten character recognition system is manifestation of been proved efficient towards classification of cross domain
datasets. The composition of datasets with labelling is one of images of up to 1000 categories of datasets. In our proposed,
the vital stages for this type of systems and this step greatly VGG NET 19 is used for classification of about 188 classes of
influences the success of the systems. Therefore handling handwritten Kannada characters. Figure 3 represents the
uncertainty of data is one of the major challenges for architecture of VGG NET 19 deep network. The details of 188
handwritten character recognition systems. classes proposed for classification is as shown in table 1, the
details of datasets used is shown in table 2.
The paper is organized as follows: proposed methodology is
explained in section II. Section III explains experimental
analysis. Section IV and section V concludes the paper.
Authorized licensed use limited to: UNIVERSITY OF ROCHESTER. Downloaded on September 21,2020 at 15:43:04 UTC from IEEE Xplore. Restrictions apply.
Proceedings of the Second International Conference on Inventive Research in Computing Applications (ICIRCA-2020)
IEEE Xplore Part Number: CFP20N67-ART; ISBN: 978-1-7281-5374-2
Authorized licensed use limited to: UNIVERSITY OF ROCHESTER. Downloaded on September 21,2020 at 15:43:04 UTC from IEEE Xplore. Restrictions apply.
Proceedings of the Second International Conference on Inventive Research in Computing Applications (ICIRCA-2020)
IEEE Xplore Part Number: CFP20N67-ART; ISBN: 978-1-7281-5374-2
Table 1 and table 2 shows the number of sample data. Table 1 training and testing. Table 2 gives the overall number of
shows the number of sample data present in each class for training and sample datasets present for 188 classes.
Authorized licensed use limited to: UNIVERSITY OF ROCHESTER. Downloaded on September 21,2020 at 15:43:04 UTC from IEEE Xplore. Restrictions apply.
Proceedings of the Second International Conference on Inventive Research in Computing Applications (ICIRCA-2020)
IEEE Xplore Part Number: CFP20N67-ART; ISBN: 978-1-7281-5374-2
In the presented methodology, classification of handwritten with a output layer at the end which labels the features
Kannada characters is performed using VGG NET 19 with abstracted to 188 classes. Table 2 presents the details of
architecture as shown in figure 3. Overall architecture is datasets used for training and testing of Kannada handwritten
comprised in four blocks of hidden layers, two dense fully characters to VGG NET 19.
connected layers and an output layer. Initially input layer is Further the details of Devanagari handwritten character set
fed with raw inputs of images with dimensions of about 32x32 used for transfer learning includes a huge data samples of
and number of channels as 3. Further four blocks of hidden about 92000 with number of classes as 46. Each data sample is
layers for feature abstraction are defined with block 1 and of dimension 32x32 with 3 channels. Training is including the
block 2 consisting of convolution layer 1, convolution layer 2 transfer of knowledge from samples of Devanagari
and a max pooling layer, block3 and block 4 are comprised of handwritten character set along with the training samples of
four convolution layers followed by a max pooling layer. about 31654 of Kannada handwritten character set.
Finally two dense fully connected layers are present along
Authorized licensed use limited to: UNIVERSITY OF ROCHESTER. Downloaded on September 21,2020 at 15:43:04 UTC from IEEE Xplore. Restrictions apply.
Proceedings of the Second International Conference on Inventive Research in Computing Applications (ICIRCA-2020)
IEEE Xplore Part Number: CFP20N67-ART; ISBN: 978-1-7281-5374-2
Authorized licensed use limited to: UNIVERSITY OF ROCHESTER. Downloaded on September 21,2020 at 15:43:04 UTC from IEEE Xplore. Restrictions apply.
Proceedings of the Second International Conference on Inventive Research in Computing Applications (ICIRCA-2020)
IEEE Xplore Part Number: CFP20N67-ART; ISBN: 978-1-7281-5374-2
Authorized licensed use limited to: UNIVERSITY OF ROCHESTER. Downloaded on September 21,2020 at 15:43:04 UTC from IEEE Xplore. Restrictions apply.