Journal Pone 0255615
Journal Pone 0255615
Journal Pone 0255615
RESEARCH ARTICLE
Readers can access the data in the same manner as a medium for data acquisition for research in psychology and behavioural science [6]. It has
as the authors. also been used as a tool for predicting personality type [1, 2], and understanding trends and
Funding: The author(s) received no specific backgrounds of users online [7, 8]. There has also been interest regarding how people express
funding for this work. sentiments when dealing with catastrophic events such as natural disasters, extreme political
Competing interests: The authors have declared viewpoints [9], and terrorism [10]. For instance, in the case of a terror attack in Kenya [10],
that no competing interests exist. Twitter became a crucial channel of communication between the government, emergency
response team and the public.
Sentiment analysis involves the use of natural language processing (NLP) [11] methods to
systematically study affective states and emotion understanding of individuals or social groups
[12–14]. We note that deep learning, which is a machine learning method, has been promi-
nently used for NLP tasks. Apart from research in psychology, sentiment analysis has a num-
ber of applications such as understanding customer behaviour [15], clinical medicine [16],
building better prediction model for trading stocks [17], and elections such as the US Presi-
dential campaign in 2012 [18]. Recently, there has been a trend of using deep learning-based
language models [19] with training data from Twitter for sentiment analysis [20, 21]. One of
the earliest works began using NLP methods such as n-grams with hash tags for building train-
ing data and machine learning methods such as Adaboost for sentiment classification [20].
Deep learning methods such as convolutional neural networks have been used for sentiment
analysis on Twitter [22].
The coronavirus disease 2019 (COVID-19) [23–26] global pandemic has been a catastrophic
event with major impact on the world’s economy which created rise in unemployment, psy-
chological issues, and depression. The abrupt social, economic and travel changes has moti-
vated research from various fields [27–29], where computational modelling with machine
learning has been prominent [30]; however, it had a number of challenges due to the testing
and reporting of cases [31]. Deep learning models have played a significant role in forecasting
COVID-19 infection treads for various parts of the world [32–34].
During rise of COVID-19 cases, and stricter lock downs, people have been expressing dif-
ferent sentiments in social media such as Twitter. Social media has played a significant role
during COVID-19 which has driven researchers for analysis with NLP and machine learning
methods. A study that used sentiment analysis via deep learning reported that World Health
Organisation (WHO) tweets have been unsuccessful in providing public guidance [35]. There
has been a study on sentiment analysis to study the effect of nationwide lockdown due to
COVID-19 outbreak in India where it was found that people took the fight against COVID19
positively and majority were in agreement with the government for the initial nation-wide
lockdown [36]. Social media posts and tweets brings another level of understanding when
combined with sentiment analysis. An example of topic modelling examined tweets during the
COVID-19 pandemic identified themes such as ‘origin of the virus’, and ‘the economy’ [37].
Topic modelling has been used with Twitter based sentiment analysis during early stages of
COVID-19 and sentiments such as fear was dominant [38]. In region specific studies, tweets
from the United States was used to determine the network of dominant topics and sentiments
[39]. Further work was done in case of China via bi-directional encoder representations from
transformers (BERT) language model for trend, topic, and sentiment analysis [40]. Further
region specific studies include community sentiment analysis in Australia [41] and sentiment
analysis in Nepal [42], where majority positive sentiments were found with elements of fear. In
Spain, sentiment analysis reviewed how digital platforms created an impact during COVID-19
[43]. A study of cross-language sentiment analysis of European Twitter messages during the
first few months of the COVID-19 pandemic found that the lockdown announcements corre-
late with a deterioration of moods, which recovers within a short time span [44].
The sentiment analysis examples from Europe and India show how social media can play a
powerful role in understanding the psychology and the human condition Recent advance-
ments of deep learning models as a tool for building robust language models have provided
further motivation in understanding the temporal nature of the sentiments during and after
the first peak of COVID-19 in India. Different countries had different peaks where rise and fall
of new cases implemented lock-downs which directly affected the economy and employment.
India is special in this way (until February 2021), where a single nation-wide peak was seen
and only certain states had multiple peaks (Delhi and Maharashtra) [33].
In this paper, we use deep learning based language models via long short-term memory
(LSTM) recurrent neural networks for sentiment analysis via tweets with a focus of rise of
novel cases in India. We use LSTM and bidirectional LSTM (BD-LSTM) model with global
vector (GloVe) for word representation for building a language model. Moreover, we use the
BERT model to compare the results from LSTM and BD-LSTM models and then use the best
model for COVID-19 sentiment analysis for the case of India. We use three datasets, which
include India, along with the state of Maharashtra (includes Mumbai) and Delhi. We compare
the monthly sentiments expressed covering the major peak of new cases in 2020. We present a
framework that focuses on multi-label sentiment classification, where more than one senti-
ment can be expressed at once. We use Senwave COVID-19 sentiment dataset [45] which fea-
tures 10,000 tweets collected worldwide and hand-labelled by 50 experts for training LSTM
models.
We highlight that there is no study that uses language models for sentiment analysis during
the rise of novel COVID-19 cases. Our framework compares the different types of sentiments
expressed across the different months in relation to the rise of the number of cases, which
impacted the economy and had different levels of lock downs. This had an effect on the psy-
chology of the population given stress and fear. Hence, the study is a way to quantify and vali-
date emotional and psychological conditions given uncertainty about the pandemic. The
major contribution of the paper is in using state-of-art sentiment analysis methods for under-
standing the public behaviour in terms of psychological well being in relation to the rise of
novel COVID-19 infections. This study presents a novel framework that makes use of informa-
tion from social media for understanding public behavior during a major disruptive event of
the century.
The rest of the paper is organised as follows: Section 2 presents a background of related
work, and Section 3 presents the proposed methodology with data analysis. Section 4 presents
experiments and results. Section 5 provides a discussion and Section 6 concludes the paper
with discussion of future work.
2 Related work
Word embedding is the process of feature extraction from text for NLP tasks such as sentiment
analysis [46–48]. Word embedding can be obtained using methods where words or phrases
from the vocabulary are mapped to vectors of real numbers. The process generally involves a
mathematical embedding from a large corpus with many dimensions per word to a vector
space with a lower dimension that is useful for machine learning or deep learning models for
text classification tasks [46]. Basic word embedding methods such as bag of words [49] and
term frequency inverse document frequency [50] do not have context awareness and semantic
information in embedding. This is also a problem for skip-grams [51] that use n-grams (such
as bigrams and tri-grams) to develop word embedding, and in addition allow adjacent
sequences of words tokens to be “skipped” [52].
Over the last decade, there has been phenomenal progress in the area of world embedding
and language models. Mikolov et al. [53] proposed word2vec embedding which uses a feedfor-
ward neural network model to learn word associations from a text dataset which can detect
synonymous words or suggest additional words given a partial sentence. It uses continuous
bag-of-words (CBOW) or continuous skip-gram model architectures to produce a distributed
representation of words. The method is used to create a large vector which represent each
unique word in the corpus where semantic information and relation between the words are
preserved. It has been shown that for two sentences that do not have much words in common,
their semantic similarity can be captured using word2vec [53]. The limitation of word2vec is
that it does not well represent the context of a word. Pennington et al. [54] for obtaining vector
representations for words by mapping words into a meaningful space where the distance
between words is related to semantic similarity. GloVe uses matrix factorization to constructs
a large matrix of co-occurrence information to obtain representation that showcase linear sub-
structures of the word vector space. The embedding feature vectors with top list words that
match with certain distance measures. GloVe can be used to find relations between words
such as synonyms, company-product relations. Due to the awareness in ethics in machine
learning, there has been a major focus on ethical issues in NLP. A recent study showed that
GloVe can have gender biased information; hence, a gender neutral GloVe method has been
proposed [55].
There are some studies that review the effectiveness of word embedding methods. Ghannay
et al. [56] provided an evaluation of word embedding methods such as GloVe [54], skip-gram,
and continuous space language models (CSLM) [57]. The authors reported that skip-gram and
GloVe outperformed CSLM in all the language tasks. Wang et al. [58] evaluated word embed-
ding methods such as GloVe for applications of biomedical text analysis where it was found
that word embedding trained from clinical notes and literature better captured word
semantics.
3 Methodology
3.1 LSTM and BERT language models
Recurrent neural networks (RNN) in general feature a context memory layer to incorporate
previous state and current inputs for propagating information to the next state, and eventually
the output layer for decision making. Canonical RNNs (simple RNNs) feature several different
architectures, besides the Elman RNN [59, 60] for modelling temporal sequences and dynam-
ical systems [61–63]. One of the major challenges in training simple RNNs is due to the archi-
tectures properties of unfolding in time for long-term dependency problems. Backpropagation
through time (BPTT), which is an extension of the backpropagation algorithm, has been
prominent for training simple RNNs [64]. Due to problem of learning long-term dependencies
given vanishing and exploding gradients with simple RNNs [65], long short-term memory
(LSTM) recurrent neural networks have been developed [66]. LSTM networks have better
capabilities for learning long-term dependencies in temporal data using memory cells and
gates.
In the last decade, with the deep learning revolution, several LSTM architectures have been
developed. Bidirectional LSTM models [67] process information in two directions, rather than
making use of only previous context state for determining the next states which are based on
bidirectional RNNs [68]. In this way, two independent LSTM models allow both backward
and forward information about the sequence at every time step. This enables better access to
long range state information which have been useful for word embedding [67] and several
other sequence processing problems [67, 69, 70].
A transformer is an extended LSTM model that adopts the mechanism of attention which
mimics cognitive attention to enhance important parts of the data while fading the rest [71].
Transformers also use an encoder-decoder architecture and have mostly been used for NLP
tasks such as translation and text summarising [71, 72]. In comparison to conventional RNNs,
transformers do not require data to be processed in a sequential order since the attention oper-
ation provides context for any position in the input sequence. BERT is a pre-trained trans-
former-based model for NLP tasks which has been used to better understand user behaviour
in Google search engine [73]. BERT can be used for a number of other NLP applications such
as clinical data processing [74]. The original BERT [73] has two models: 1.) BERT-base fea-
tures 12 encoders with 12 bidirectional self-attention heads, and 2.) BERT-large features 24
encoders with 16 bidirectional self-attention heads. These are pre-trained from unlabeled data
extracted from a corpus with 800 million words and English Wikipedia with 2,500 million
words, respectively. Word2vec and GloVe are content-free models that generate a single word
embedding representation for each word, whereas BERT takes into account the context for
each occurrence of a given word which makes BERT one of the best language models. Hence,
BERT is suitable for our sentiment analysis tasks using tweets during rise of COVID-19 cases
in India.
Fig 1. Framework: Twitter-based sentiment analysis for COVID-19 with model development via LSTM, BD-LSTM and BERT. After training, the best
model is chosen for prediction of COVID-19 tweets from India. Note that the framework features multi-label classification which has the ability to provide
more than two outcomes at once; i.e. a tweet can be optimistic and joking at the same time.
https://doi.org/10.1371/journal.pone.0255615.g001
Table 1. Examples of word transformation that features changing symbols and emoji’s to words that can have
semantic representation.
Original Phrase/Emoji Transformed Word
omg oh my god
btw by the way
socialdistancing social distancing
smiling face
sad face
bed
fire
wink
laugh
https://doi.org/10.1371/journal.pone.0255615.t001
(emoji’s) and abbreviations are present in tweets, since language in social media has been rap-
idly evolving. Hence, we have to transform them for building the language model. Table 1 We
note that since this is a case of India, where English is used in combination with some key
indigenous Indian languages (such as Hindi). Hence, we transform certain words, emotions
and character symbols expressed in other languages as well. Note that we use hand-labelled
sentiment dataset (Senwave COVID-19 dataset) [45] which features 11 different sentiments
labelled by a group of 50 experts for 10,000 tweets worldwide during COVID-19 pandemic in
2020. In this dataset, “official report” is classified as a sentiment although it is a topic.
The next step involves converting each word into its corresponding GLoVe embedding vec-
tor, where each word is converted into a vector (300 dimensions in our case). We select
GLoVe embedding since it has shown good results for language models with sentiment analy-
sis in the literature [54]. The aforementioned vector from each word thus obtained is passed
on to the respective LSTM models for training once the model architecture has been defined
as shown in Fig 1. We first evaluate our trained models (LSTM, BD-LSTM, and BERT) and
then use it for COVID-19 sentiment analysis using test data (India, Maharashtra, and Delhi).
As shown in Fig 1, the trained model is used to classify 11 sentiments such as “anxious”, “sad”,
and “joking”.
It is essential to have the right metric to evaluate the model for the given application. The
outcome of a misclassification in a multi-label classification is no longer a correct or an incor-
rect instance as opposed to binary classification or multi-class classification [75]. In multi-label
classification, a prediction that features a subset of the actual classes are considered better than
a prediction that contains none of the actual classes; hence, this needs to be captured by the
metric that captures the loss or gives a score. As given in the literature [75], multi-label classifi-
cation evaluation is typically based on 1.) binary cross-entropy (BCE) loss [76] which is a soft-
max activation combined with a cross-entropy loss; 2.) Hamming loss [75] which calculates
loss generated in the bit string of class labels using exclusive or (XOR) operation between the
actual and predicted labels and then computes the average over the instances of the dataset.
The Jaccard coefficient score [77] provides a measure of the overlap between actual and pre-
dicted labels with their attributes capturing similarity and diversity. Furthermore, label rank-
ing average precision (LRAP) score [78] finds the percentage of the higher-ranked labels that
resemble true labels for each of the given samples. F1-score conveys the balance between the
precision and the recall. F1-score has been prominent for understanding class imbalance prob-
lems [79]. In application to multi-label classification, two types of F1-scores are predominantly
used. F1-macro is computed using the F1-score per class/label and then averaging them,
whereas F1-micro computes the F1-score for the entire data considering the total true
positives, false negatives and false positives [80]. We use a combination of these scores to eval-
uate our model results for the first phase of the framework where the respective models are
trained.
Table 2. Training performance for BD-LSTM, LSTM and BERT model using Senwave COVID-19 training dataset.
Note that except for the BCE and Hamming loss, higher scores shows better performance.
Metric BD-LSTM LSTM BERT
BCE Loss 0.281 0.255 0.372
Hamming Loss 0.163 0.157 0.142
Jaccard Score 0.417 0.418 0.510
LRAP Score 0.503 0.511 0.766
F1 Macro 0.434 0.430 0.530
F1 Micro 0.495 0.493 0.587
https://doi.org/10.1371/journal.pone.0255615.t002
January 2021, and since then it was gradually raising. In March 2021, the cases began rising
faster and by 22nd March, India had 47 thousand daily new cases which was moving towards a
second peak [82].
In the first six months, the states of Maharashtra (population of about 124 million), Delhi
(population of 19 million), and Tamil Nadu (population of about 8 million) led COVID-19
infections [83]. Note that the city of Mumbai is in the state of Maharashtra, which in terms of
the population can be compared to some of the highly populated countries. In the later half of
the year, Delhi cases reduced but still remained in leading 8 states [33]. In 2021, Maharashtra
continued as state of highest infections and in March, it was featuring more than half of new
cases on a weekly basis and Delhi contained the situation with less than a thousand daily cases.
Hence, our focus is to study whole of India with two states of interest that includes Maharash-
tra and Delhi.
We note that the proposed framework can be applied to any part of the world; however, we
are choosing the case of India to show the effectiveness of the framework. The final step is in
applying the different datasets from COVID-19 in India which include, nation-wide COVID-
19 sentiments, and two major states with COVID-19 cases. The trend in the cases shows that
both states had a major peak followed by minor ones, whereas India as a whole had a single
major peak which was around mid-September 2020 with close to 97,000 novel cases per day
during the peak as shown in the next section (Fig 2).
4 Results
In this section, we provide results of the implementation of our proposed framework using
Twitter dataset for COVID-19 in India.
Fig 2. Novel COVID-19 tweets and cases in India, Maharashtra, and Delhi. The red bars indicate the number of
tweets while the black bars indicate the number of novel cases.
https://doi.org/10.1371/journal.pone.0255615.g002
Fig 3. Bi-grams and tri-grams for case of India. Note that a combination of emoji’s such as “backhand—index—pointing”, “smiling—face”, and words
have been used. (a) Bi-grams. (b) Tri-grams.
https://doi.org/10.1371/journal.pone.0255615.g003
is an emotion icon (emoji) used predominantly in social media, and hence been processed as a
text during prepossessing (Table 1), which is actually is not a bi-gram from a semantic view-
point. In order to provide better understanding of the context of the tweets, we give examples
in Table 3. In the case of tri-grams (Panel b), we find more information in tweets such as
“backhand—index—pointing” which is an emoji; hence, we provide some of the tweets that
fall in this category in Table 4.
Fig 4 shows number of occurrence of a given sentiment in relation to the rest of the senti-
ments for 10,000 hand-labelled tweets in Senwave dataset [45] used for training.
Table 3. Selected examples of processed tweets that are captured in most prominent bi-grams.
Month Tweet Bi-gram
(Emoji)
March “releasing today at 6:00 pm on my youtube channel! let’s fight this together folded hands i “folded hand”
need your support guys”
July “that’s really shameful and heinous folded hands” “folded hand”
August “please applaud this corona economy warrior. folded hands kudos.” “folded hand”
March “india this backhand index pointing down” “index
pointing”
May “corona: to everyone backhand index pointing down” “index
pointing”
July “backhand index pointing right. . . .. . .‥lockdown time. . .‥ #picture #instagood “index
#instadaily #insta” pointing”
https://doi.org/10.1371/journal.pone.0255615.t003
sentiment was included in the Senwave dataset, and hence we use it for prediction. Fig 5 pres-
ents distribution of sentiments predicted by the LSTM and BERT models for the respective
datasets for the entire time-span of interest (March to September 2020). In the case of India
(Panel a), we notice that the “optimistic”, “annoyed” and “joking” sentiments are the most
expressed which also follows in the case of Maharashtra and Delhi (Panel b and c). We notice
that the BERT model seems to capture more sentiments expressed when compared to the
LSTM model, particularly the “optimistic”, “anxious”, and “annoyed” sentiments. We find that
negative sentiments such as “pessimistic”, “anxious” and “sad” have been least expressed. We
find “optimistic” sentiment as the most prominent sentiment which is followed by “annoyed”
and “joking”. It is simpler to label “optimistic” and “thankful” as a positive sentiment, but it
becomes increasing difficult when it comes to the sentiments “joking” and “surprise”, when
context information is not available. Hence, it is insightful to review the sentiments such as
“joking” and “surprise” in relation to other sentiments expressed. We find a similar trend in
the case of Delhi and Maharashtra (Panel b and c) which are subsets of data from India (Panel
a), the only major difference is the number of tweets which is significantly lower due to their
respective population. Since the BERT model provided the best results for the training data,
we provide the results by the BERT model henceforward.
Next, we review the sentiments predicted using a heatmap to examine number of occur-
rence of a given sentiment in relation to the rest of the sentiments (Figs 6–8). These heatmaps
essentially indicate how two sentiments have been expressed at once and provides more
insights regarding the context of the negative and positive sentiments given in Fig 5. Note that
in Fig 5, we found that the third most prominent sentiment has been “joking”; however, we
were not sure if it is positive or negative sentiment. As shown in Fig 6, we find that most tweets
that are associated with “joking” are either “optimistic” (8680) or “annoyed” (10323), and
some are also “thankful” (889). A much lower portion (below 500) are either “empathetic” or
Table 4. Selected examples of processed tweets that are captured in most prominent tri-grams.
Month Tweet Tri-gram (Emoji)
May “google doc of resources for amphan and covid. backhand index pointing “backhand—index—
down. retweet for greater reach.” pointing”
August “#covid tips backhand index pointing down” “backhand—index—
pointing”
September “backhand index pointing right registers highest single-day cases in the “backhand—index—
world—95,529” pointing”
https://doi.org/10.1371/journal.pone.0255615.t004
Fig 4. Heatmap showing number of occurrence of a given sentiment in relation to the rest of the sentiments for 10,000 tweets from Senwave
dataset [45] used for training.
https://doi.org/10.1371/journal.pone.0255615.g004
“official report” while “joking”. Next, we review the sentiment “optimistic” and find that
majority are either “thankful” (14230) or “joking” (8680). The “optimistic” tweets with nega-
tive sentiments are relatively minor, such as “pessimistic” (772), “denial” (143), and “sad”
(349). Furthermore, a significant portion are also “empathetic” (2533). It is not common for
one to make a statement that is optimistic and pessimistic at the same time; hence, this could
be a wrong prediction by the model. However, it seems the model is making the right predic-
tion when we look at the heatmap for the hand-labelled training datasets (Fig 4), where such
Fig 5. Distribution of sentiments predicted for the respective datasets by the LSTM and BERT models. (a) India.
(b) Maharashtra. (c) Delhi.
https://doi.org/10.1371/journal.pone.0255615.g005
Fig 6. Heatmap showing number of occurrence of a given sentiment in relation to the rest of the sentiments for the India dataset using the BERT
model.
https://doi.org/10.1371/journal.pone.0255615.g006
combinations of sentiments have been labelled by experts. We show examples of tweets of this
case and compare with those that are optimistic and thankful in Table 5.
Figs 7 and 8 show the number of occurrence of a given sentiment in relation to the rest of
the sentiments for the case of Maharashtra and Delhi, which follows a similar pattern when
compared to case of India (Fig 6)). We note that the Senwave dataset which shows tweets
worldwide (Fig 4) follow a similar pattern than the case in Indian datasets when it comes to
sentiments such as “joking” and being “optimistic” or “joking” and “annoyed”. Senwave
dataset also features cases of being optimistic and pessimistic at the same time (226 cases). This
Fig 7. Heatmap showing number of occurrence of a given sentiment in relation to the rest of the sentiments for the Maharashtra dataset using the
BERT model.
https://doi.org/10.1371/journal.pone.0255615.g007
could be due to sentiments expressed in two opposing sentences in a tweet. We can infer that
since such patterns are part of the training data, it cannot be an error in predictions when
looking at the Indian datasets.
Fig 9 provides a visualisation of the distribution of tweets with number of combination sen-
timents. We find that around 60% of the tweets have a singular sentiment. Moreover, about
25% of the tweets have two sentiment attached them and 14% have no sentiment attached to
Fig 8. Heatmap showing number of occurrence of a given sentiment in relation to the rest of the sentiments for the Delhi dataset using the BERT
model.
https://doi.org/10.1371/journal.pone.0255615.g008
them. Furthermore, a small number of tweets have 3 or more emotions attached to them
which indicates that much often, people do not show multiple emotions at the same time.
Finally, we present results that visualises the trend of sentiments expressed over time in the
three datasets. This is one of the key findings which can enable a understanding of the reaction
in terms of emotions expressed by the population given rise and fall of COVID-19 novel cases
as shown in Fig 2. Figs 10–12 presents visualisation of the monthly sentiments of India, Maha-
rashtra and Delhi, respectively.
Table 5. Selected example that show cases of tweets that are “optimistic” and also “pessimistic”, along with cases
that are “optimistic” and also “thankful”.
Month Tweet Sentiment
combination
May “don’t care china only care indian covid19 news” “optimistic—
pessimistic”
August “dear uddhav ji other than covid problem, what ever wrong is happening is not “optimistic—
good for your government” pessimistic”
September “sir, plz look in have benefits to sr citizen.” “optimistic—
pessimistic”
April “thank you switzerland smiling face with smiling eyes especially zermatt for “optimistic—
showing” solidarity for india flag” thankful”
May “big thanks to the cfpc.. . .thoroughly enjoyed it!” “optimistic—
thankful”
June “doctors, activists urge pm to promote plant-based diet—india news—times of “optimistic—
india” thankful”
https://doi.org/10.1371/journal.pone.0255615.t005
5 Discussion
Humans can express more than one sentiment at a time; however, there are variations in the
number of sentiments that can be expressed by facial expressions [85], when compared to spo-
ken or written emotions such as tweets. As shown in Fig 9, majority of the tweets have one sen-
timent, both in hand-labelled (Panel a) and predicted datasets (Panel b). This is followed by
two sentiments, while a minority have 3 sentiments. We note that there is a significant portion
of tweets with no sentiments and there are no tweets of more than three sentiments at a time.
The study of emotions with technology has gained much interested in last decade which gave
much progress in understanding sentiments. Trampe et al. [86] presented a study of everyday
emotional experiences through an experience sampling smartphone application that moni-
tored real-time emotions of more than 11,000 participants and found the group experienced at
least one emotion 90% of the time with joy as the the most frequent emotion, followed by love
and anxiety. The type of emotion would be highly dependent on the study region which fea-
tured Europeans with majority French participants. Cowen and Keltner reported [87] twenty
seven distinct categories of emotions bridged by continuous gradients by data from emotion-
ally evocative short videos with varying situational content. These studies contributed to better
understanding of emotions given historical viewpoints about context and definitions of emo-
tions and associations between them [88–91]. However, we did not find any study that reviews
the number of emotions that can be expressed at a time in relation to catastrophic events that
keep changing with time, such as the rise of COVID-19 cases.
We revisit the case of Indian dataset (Fig 10), where the monthly tweets did not sharply
increase with number of novel cases (Fig 2) with a nationwide peak of novel cases (Fig 2, Panel
a). The number of tweets gradually increased with peak of tweets in July (Fig 2, Panel a). When
India had a peak of novel cases, we found that the number of tweets significantly lowered.
Hence, people have been alarmed by rising cases, but a month before the peak of novel cases
was seen, the tweets were reduced. Moreover, we find that the “optimistic”, “annoyed” and
“joking” tweets are mostly dominating the monthly tweets for India (Fig 10) and Maharashtra
(Fig 11), with a mix of annoyed sentiments in case of Delhi (Fig 12). There is significantly
lower number of negative sentiments for the respective datasets (Figs 10–12).
We note that a limitation of the framework is due to the Senwave training data which con-
sidered tweets worldwide during COVID-19 by a group of experts; however, there can be limi-
tations on how experts perceive humour in different regions. Humour is expressed differently
Fig 9. Distribution of tweets with number of combination of sentiment types (labels) for Senwave training
dataset, and predictions by BERT and LSTM models. (a) Senwave hand-labelled sentiments (Worldwide). (b)
Predicted (India) using BERT. (c) Predicted (India) using LSTM.
https://doi.org/10.1371/journal.pone.0255615.g009
Fig 10. Monthly COVID-19 sentiments in India using the BERT model.
https://doi.org/10.1371/journal.pone.0255615.g010
Fig 11. Monthly COVID-19 sentiments in Maharashtra using the BERT model.
https://doi.org/10.1371/journal.pone.0255615.g011
Fig 12. Monthly COVID-19 sentiments in Delhi using the BERT model.
https://doi.org/10.1371/journal.pone.0255615.g012
in different cultural groups, i.e. a tweet that may be humorous in USA may not be taken as
humorous in India due to cultural and language dimensions. There are several studies about
the effect of humour in changing cultural or regional settings [92–94]. A good example is that
in orthodox Chinese literature, humour was not expressed due to religious taboo in ancient
Buddhism which later eased with Zen Buddhism [95, 96]; however, the Hindu literature had a
different or eased attitude towards humour as expressed in ancient Hindu texts [97]. Although
historic textual composition cannot be related to how tweets are expressed in India, it is
important to note the cultural and language differences in how humour has been expressed.
Another limitation of the study is regarding the uncertainty of the data. Our results are
based on the tweets expressed; however, we note that only a small fraction of the population
generally express their views on Twitter. Moreover, we need to be aware that due to restric-
tions on freedom of speech and political biases, not everyone can express themselves freely.
Social media networks have been active in monitoring how people express themselves in order
to limit the rise of anti-vaccine sentiments. Furthermore, our framework’s training data is
based on tweets expressed worldwide, whereas the test data is based on tweets from India dur-
ing the pandemic. There is a large language diversity in India and at times, people express
themselves with a combination of languages with geographic and culture specific jargon.
Although, we convert them to English, certain limitations in capturing the context will exist.
We note that there has not been much work in uncertainty quantification for the predic-
tions and there are different level of uncertainties, particularly in model parameters and data.
We note that the training data is hand-labelled, and at times two or three sentiments have
been expressed at once. This could be something open to interpretation by experts as it is hard
to formally detect more than one sentiment from a tweet of only thirty words. Hence, the
expert labelled training dataset adds to uncertainty in model predictions. In future work,
Bayesian deep learning can provide a way to address uncertainty in model predictions
[98–100].
6 Conclusions
We presented a study with novel deep learning models for sentimental analysis during the rise
of COVID-19 infections. We selected tweets from India as our case study and reviewed tweets
from specific regions that included Maharastra and Delhi. We took advantage of COVID-19
dataset of 10,000 hand-labelled tweets for training the respective deep learning models. Our
investigation revealed that majority of the tweets have been “optimistic”, “annoyed” and “jok-
ing” that expresses optimism, fear and uncertainty during the rise of the COVID-19 cases in
India. The number of tweets significantly lowered towards the peak of new cases. Furthermore,
the optimistic, annoyed and joking tweets mostly dominated the monthly tweets with much
lower number of negative sentiments expressed. We found that most tweets that have been
associated with “joking” were either “optimistic” or “annoyed”, and minority of them were
also “thankful”. In terms of the “annoyed” sentiments in tweets, mostly were either “surprised”
or “joking”. These predictions generally indicate that although the majority have been optimis-
tic, a significant group of population has been annoyed towards the way the pandemic was
handled by the authorities. The major contribution of the paper is the framework which pro-
vides sentiment analysis in a population given the rise of the COVID-19 cases. The framework
can be used by officials for better COVID-19 management through policies and projects, such
as support for depression and mental health issues.
Future work can use the framework for different regions, countries, ethnic and social
groups to understand their behaviour given multiple peaks of novel cases. The framework can
be extended to understand reactions towards vaccinations with the rise of anti-vaccine senti-
ments given fear, insecurity and unpredictability of COVID-19. Finally, the framework can
incorporate topic modelling with sentiment analysis which can provide more details for
emerging topics during the rise of COVID-19 cases in relation to various government proto-
cols such as lock-downs and vaccination plans.
Author Contributions
Conceptualization: Rohitash Chandra.
Investigation: Rohitash Chandra.
Methodology: Rohitash Chandra, Aswin Krishna.
Project administration: Rohitash Chandra.
Software: Rohitash Chandra, Aswin Krishna.
Supervision: Rohitash Chandra.
Validation: Aswin Krishna.
Visualization: Aswin Krishna.
Writing – original draft: Rohitash Chandra.
Writing – review & editing: Rohitash Chandra.
References
1. Golbeck J, Robles C, Edmondson M, Turner K. Predicting personality from twitter. In: 2011 IEEE third
international conference on privacy, security, risk and trust and 2011 IEEE third international confer-
ence on social computing. IEEE; 2011. p. 149–156.
2. Quercia D, Kosinski M, Stillwell D, Crowcroft J. Our twitter profiles, our selves: Predicting personality
with twitter. In: 2011 IEEE third international conference on privacy, security, risk and trust and 2011
IEEE third international conference on social computing. IEEE; 2011. p. 180–185.
3. Bittermann A, Batzdorfer V, Müller SM, Steinmetz H. Mining Twitter to detect hotspots in psychology.
Zeitschrift für Psychologie. 2021;. https://doi.org/10.1027/2151-2604/a000437
4. Lin J. On building better mousetraps and understanding the human condition: Reflections on big data
in the social sciences. The ANNALS of the American Academy of Political and Social Science. 2015;
659(1):33–47. https://doi.org/10.1177/0002716215569174
5. Coppersmith G, Dredze M, Harman C. Quantifying mental health signals in Twitter. In: Proceedings of
the workshop on computational linguistics and clinical psychology: From linguistic signal to clinical
reality; 2014. p. 51–60.
6. Murphy SC. A hands-on guide to conducting psychological research on Twitter. Social Psychological
and Personality Science. 2017; 8(4):396–412. https://doi.org/10.1177/1948550617697178
7. Zhou Y, Na JC. A comparative analysis of Twitter users who Tweeted on psychology and political sci-
ence journal articles. Online Information Review. 2019; 43(7):1188–1208. https://doi.org/10.1108/
OIR-03-2019-0097
8. Wang W, Hernandez I, Newman DA, He J, Bian J. Twitter analysis: Studying US weekly trends in work
stress and emotion. Applied Psychology. 2016; 65(2):355–378. https://doi.org/10.1111/apps.12065
9. Alizadeh M, Weber I, Cioffi-Revilla C, Fortunato S, Macy M. Psychology and morality of political
extremists: evidence from Twitter language analysis of alt-right and Antifa. EPJ Data Science. 2019;
8(17): https://doi.org/10.1140/epjds/s13688-019-0193-9
10. Garg P, Garg H, Ranga V. Sentiment analysis of the Uri terror attack using Twitter. In: 2017 Interna-
tional conference on computing, communication and automation (ICCCA). IEEE; 2017. p. 17–20.
11. Manning C, Schutze H. Foundations of statistical natural language processing. MIT press; 1999.
12. Liu B, Zhang L. A survey of opinion mining and sentiment analysis. In: Mining text data. Springer;
2012. p. 415–463.
13. Medhat W, Hassan A, Korashy H. Sentiment analysis algorithms and applications: A survey. Ain
Shams engineering journal. 2014; 5(4):1093–1113. https://doi.org/10.1016/j.asej.2014.04.011
14. Hussein DMEDM. A survey on sentiment analysis challenges. Journal of King Saud University-Engi-
neering Sciences. 2018; 30(4):330–338. https://doi.org/10.1016/j.jksues.2016.04.002
15. Ordenes FV, Theodoulidis B, Burton J, Gruber T, Zaki M. Analyzing customer experience feedback
using text mining: A linguistics-based approach. Journal of Service Research. 2014; 17(3):278–295.
https://doi.org/10.1177/1094670514524625
16. Greaves F, Ramirez-Cano D, Millett C, Darzi A, Donaldson L. Use of sentiment analysis for capturing
patient experience from free-text comments posted online. Journal of medical Internet research. 2013;
15(11):e239. https://doi.org/10.2196/jmir.2721
17. Mittal A, Goel A. Stock prediction using Twitter sentiment analysis. Standford University, CS229 (2011
http://cs229.stanford.edu/proj2011/GoelMittal-StockMarketPredictionUsingTwitterSentimentAnalysis.
pdf). 2012; 15.
18. Wang H, Can D, Kazemzadeh A, Bar F, Narayanan S. A system for real-time Twitter sentiment analy-
sis of 2012 us presidential election cycle. In: Proceedings of the ACL 2012 system demonstrations;
2012. p. 115–120.
19. Zhang L, Wang S, Liu B. Deep learning for sentiment analysis: A survey. Wiley Interdisciplinary
Reviews: Data Mining and Knowledge Discovery. 2018; 8(4):e1253.
20. Kouloumpis E, Wilson T, Moore J. Twitter sentiment analysis: The good the bad and the OMG! In: Pro-
ceedings of the Fifth International Conference on Weblogs and Social Media; 2011. p. 538–541
21. Giachanou A, Crestani F. Like it or not: A survey of Twitter sentiment analysis methods. ACM Comput-
ing Surveys. 2016; 49(2):1–41. https://doi.org/10.1145/2938640
22. Severyn A, Moschitti A. Twitter sentiment analysis with deep convolutional neural networks. In: Pro-
ceedings of the 38th international ACM SIGIR conference on research and development in information
retrieval; 2015. p. 959–962.
23. Gorbalenya AE, Baker SC, Baric RS, de Groot RJ, Drosten C, Gulyaeva AA, et al. The species Severe
acute respiratory syndrome-related coronavirus: classifying 2019-nCoV and naming it SARS-CoV-2.
Nature Microbiology. 2020; 5:536–544. https://doi.org/10.1038/s41564-020-0695-z
24. Monteil V, Kwon H, Prado P, Hagelkrüys A, Wimmer RA, Stahl M, et al. 536–544 Inhibition of SARS-
CoV-2 infections in engineered human tissues using clinical-grade soluble human ACE2. Cell. 2020;
181: 905–913 https://doi.org/10.1016/j.cell.2020.04.004 PMID: 32333836
25. Organization WH, et al. Coronavirus disease 2019 (COVID-19): situation report, 72; 2020. Available
from: https://apps.who.int/iris/handle/10665/331685.
26. Cucinotta D, Vanelli M. WHO declares COVID-19 a pandemic. Acta bio-medica: Atenei Parmensis.
2020; 91(1):157–160.
27. Siche R. What is the impact of COVID-19 disease on agriculture? Scientia Agropecuaria. 2020;
11(1):3–6.
28. Richards M, Anderson M, Carter P, Ebert BL, Mossialos E. The impact of the COVID-19 pandemic on
cancer care. Nature Cancer. 2020; 1(6):565–567. https://doi.org/10.1038/s43018-020-0074-y
29. Tiwari A, Gupta R, Chandra R. Delhi air quality prediction using LSTM deep learning models with a
focus on COVID-19 lockdown. arXiv preprint arXiv:210210551. 2021.
30. Shinde GR, Kalamkar AB, Mahalle PN, Dey N, Chaki J, Hassanien AE. Forecasting models for coro-
navirus disease (COVID-19): a survey of the state-of-the-art. SN Computer Science. 2020; 1(4):1–15.
31. Rahimi I, Chen F, Gandomi AH. A review on COVID-19 forecasting models. Neural Computing and
Applications. 2021; p. 1–11. https://doi.org/10.1007/s00521-020-05626-8 PMID: 33564213
32. Zeroual A, Harrou F, Dairi A, Sun Y. Deep learning methods for forecasting COVID-19 time-Series
data: A Comparative study. Chaos, Solitons & Fractals. 2020; 140:110121. https://doi.org/10.1016/j.
chaos.2020.110121
33. Chandra R, Jain A, Chauhan DS. Deep learning via LSTM models for COVID-19 infection forecasting
in India. CoRR. 2021;abs/2101.11881.
34. Tiwari A, Gupta R, Chandra R. Delhi air quality prediction using LSTM deep learning models with a
focus on COVID-19 lockdown. CoRR. 2021;abs/2102.10551.
35. Chakraborty K, Bhatia S, Bhattacharyya S, Platos J, Bag R, Hassanien AE. Sentiment Analysis of
COVID-19 tweets by Deep Learning Classifiers—A study to show how popularity is affecting accuracy
in social media. Applied Soft Computing. 2020; 97:106754. https://doi.org/10.1016/j.asoc.2020.
106754
36. Barkur G, Vibha GBK. Sentiment analysis of nationwide lockdown due to COVID 19 outbreak: Evi-
dence from India. Asian journal of psychiatry. 2020; 51:102089. https://doi.org/10.1016/j.ajp.2020.
102089
37. Abd-Alrazaq A, Alhuwail D, Househ M, Hamdi M, Shah Z. Top concerns of tweeters during the
COVID-19 pandemic: infoveillance study. Journal of medical Internet research. 2020; 22(4):e19016.
https://doi.org/10.2196/19016
38. Xue J, Chen J, Chen C, Zheng C, Li S, Zhu T. Public discourse and sentiment during the COVID 19
pandemic: Using Latent Dirichlet Allocation for topic modeling on Twitter. PloS one. 2020; 15(9):
e0239441. https://doi.org/10.1371/journal.pone.0239441
39. Hung M, Lauren E, Hon ES, Birmingham WC, Xu J, Su S, et al. Social network analysis of COVID-19
Sentiments: application of artificial intelligence. Journal of medical Internet research. 2020; 22(8):
e22590. https://doi.org/10.2196/22590 PMID: 32750001
40. Wang T, Lu K, Chow KP, Zhu Q. COVID-19 Sensing: Negative sentiment analysis on social media in
China via Bert Model. Ieee Access. 2020; 8:138162–138169.
41. Zhou J, Yang S, Xiao C, Chen F. Examination of community sentiment dynamics due to COVID-19
pandemic: A case study from Australia. arXiv preprint arXiv:200612185. 2020;.
42. Pokharel BP. Twitter sentiment analysis during COVID-19 outbreak in Nepal. Available at SSRN
3624719. 2020.
43. de Las Heras-Pedrosa C, Sánchez-Núñez P, Peláez JI. Sentiment analysis and emotion understand-
ing during the COVID-19 pandemic in Spain and its impact on digital ecosystems. International Journal
of Environmental Research and Public Health. 2020; 17(15):5542; https://doi.org/10.3390/
ijerph17155542 PMID: 32751866
44. Kruspe A, Häberle M, Kuhn I, Zhu XX. Cross-language sentiment analysis of European Twitter mes-
sages during the COVID-19 pandemic. arXiv preprint arXiv:200812172. 2020.
45. Yang Q, Alamro H, Albaradei S, Salhi A, Lv X, Ma C, et al. SenWave: Monitoring the Global Senti-
ments under the COVID-19 Pandemic. arXiv preprint arXiv:200610842. 2020.
46. Li Y, Yang T. Word embedding for understanding natural language: a survey. In: Guide to big data
applications. Springer; 2018. p. 83–104.
47. Kutuzov A, Øvrelid L, Szymanski T, Velldal E. Diachronic word embeddings and semantic shifts: a sur-
vey. arXiv preprint arXiv:180603537. 2018;.
48. Ruder S, Vulić I, Søgaard A. A survey of cross-lingual word embedding models. Journal of Artificial
Intelligence Research. 2019; 65:569–631. https://doi.org/10.1613/jair.1.11640
49. Zhang Y, Jin R, Zhou ZH. Understanding bag-of-words model: a statistical framework. International
Journal of Machine Learning and Cybernetics. 2010; 1(1-4):43–52. https://doi.org/10.1007/s13042-
010-0001-0
50. Ramos J, et al. Using TF-IFD to determine word relevance in document queries. In: Proceedings of
the first instructional conference on machine learning. vol. 242.; 2003. p. 29–48.
51. Goodman JT. A bit of progress in language modeling. Computer Speech & Language. 2001;
15(4):403–434. https://doi.org/10.1006/csla.2001.0174
52. Guthrie D, Allison B, Liu W, Guthrie L, Wilks Y. A closer look at skip-gram modelling. In: LREC. vol. 6.;
2006. p. 1222–1225.
53. Mikolov T, Sutskever I, Chen K, Corrado G, Dean J. Distributed representations of words and phrases
and their compositionality. arXiv preprint arXiv:13104546. 2013.
54. Pennington J, Socher R, Manning CD. GloVe: Global Vectors for Word Representation. In: Empirical
Methods in Natural Language Processing (EMNLP); 2014. p. 1532–1543.
55. Zhao J, Zhou Y, Li Z, Wang W, Chang KW. Learning gender-neutral word embeddings. arXiv preprint
arXiv:180901496. 2018.
56. Ghannay S, Esteve Y, Camelin N, Deléglise P. Evaluation of acoustic word embeddings. In: Proceed-
ings of the 1st Workshop on Evaluating Vector-Space Representations for NLP; 2016. p. 62–66.
57. Schwenk H. Continuous space language models. Computer Speech & Language. 2007; 21(3):492–
518. https://doi.org/10.1016/j.csl.2006.09.003
58. Wang Y, Liu S, Afzal N, Rastegar-Mojarad M, Wang L, Shen F, et al. A comparison of word embed-
dings for the biomedical natural language processing. Journal of biomedical informatics. 2018; 87:12–
20. https://doi.org/10.1016/j.jbi.2018.09.008 PMID: 30217670
59. Elman JL, Zipser D. Learning the hidden structure of speech. The Journal of the Acoustical Society of
America. 1988; 83(4):1615–1626. https://doi.org/10.1121/1.395916
60. Elman JL. Finding structure in time. Cognitive Science. 1990; 14:179–211. https://doi.org/10.1207/
s15516709cog1402_1
61. Omlin CW, Giles CL. Constructing deterministic finite-state automata in recurrent neural networks. J
ACM. 1996; 43(6):937–972. https://doi.org/10.1145/235809.235811
62. Omlin CW, Giles CL. Training second-order recurrent neural networks using hints. In: Proceedings of
the Ninth International Conference on Machine Learning. Morgan Kaufmann; 1992. p. 363–368.
63. Chandra R, Omlin CW. The Comparison and Combination of Genetic and Gradient Descent Learning
in Recurrent Neural Networks: An Application to Speech Phoneme Classification. In: Proc. of Interna-
tional Conference on Artificial Intelligence and Pattern Recognition; 2007. p. 286–293.
64. Werbos PJ. Backpropagation through time: what it does and how to do it. Proceedings of the IEEE.
1990; 78(10):1550–1560. https://doi.org/10.1109/5.58337
65. Hochreiter S. The vanishing gradient problem during learning recurrent neural nets and problem solu-
tions. Int J Uncertain Fuzziness Knowl-Based Syst. 1998; 6(2):107–116. https://doi.org/10.1142/
S0218488598000094
66. Hochreiter S, Schmidhuber J. Long short-term memory. Neural computation. 1997; 9(8):1735–1780.
https://doi.org/10.1162/neco.1997.9.8.1735
67. Graves A, Schmidhuber J. Framewise phoneme classification with bidirectional LSTM and other neu-
ral network architectures. Neural Networks. 2005; 18:602–10. https://doi.org/10.1016/j.neunet.2005.
06.042
68. Schuster M, Paliwal K. Bidirectional recurrent neural networks. Signal Processing, IEEE Transactions
on. 1997; 45:2673–2681. https://doi.org/10.1109/78.650093
69. Fan Y, Qian Y, Xie FL, Soong FK. TTS synthesis with bidirectional LSTM based recurrent neural net-
works. In: Fifteenth annual conference of the international speech communication association; 2014.
p. 1964– 1968.
70. Graves A, Jaitly N, Mohamed Ar. Hybrid speech recognition with deep bidirectional LSTM. In: 2013
IEEE workshop on automatic speech recognition and understanding. IEEE; 2013. p. 273–278.
71. Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, et al. Attention is all you need.
arXiv preprint arXiv:170603762. 2017.
72. Wolf T, Chaumond J, Debut L, Sanh V, Delangue C, Moi A, et al. Transformers: State-of-the-art natu-
ral language processing. In: Proceedings of the 2020 Conference on Empirical Methods in Natural
Language Processing: System Demonstrations; 2020. p. 38–45.
73. Devlin J, Chang MW, Lee K, Toutanova K. Bert: Pre-training of deep bidirectional transformers for lan-
guage understanding. arXiv preprint arXiv:181004805. 2018.
74. Su Y, Xiang H, Xie H, Yu Y, Dong S, Yang Z, et al. Application of BERT to Enable Gene Classification
Based on Clinical Evidence. BioMed Research International. 2020: https://doi.org/10.1155/2020/
5491963 PMID: 33083472
75. Dembczyński K, Waegeman W, Cheng W, Hüllermeier E. On label dependence and loss minimization
in multi-label classification. Machine Learning. 2012; 88(1-2):5–45. https://doi.org/10.1007/s10994-
012-5285-8
76. Zhang Z, Sabuncu MR. Generalized cross entropy loss for training deep neural networks with noisy
labels. arXiv preprint arXiv:180507836. 2018;.
77. Hamers L, et al. Similarity measures in scientometric research: The Jaccard index versus Salton’s
cosine formula. Information Processing and Management. 1989; 25(3):315–18. https://doi.org/10.
1016/0306-4573(89)90048-4
78. Fürnkranz J, Hüllermeier E, Mencı́a EL, Brinker K. Multilabel classification via calibrated label ranking.
Machine learning. 2008; 73(2):133–153. https://doi.org/10.1007/s10994-008-5064-8
79. Jeni LA, Cohn JF, De La Torre F. Facing imbalanced data–recommendations for the use of perfor-
mance metrics. In: 2013 Humaine association conference on affective computing and intelligent inter-
action. IEEE; 2013. p. 245–251.
80. Lewis DD, Schapire RE, Callan JP, Papka R. Training algorithms for linear text classifiers. In: Proceed-
ings of the 19th annual international ACM SIGIR conference on Research and development in infor-
mation retrieval; 1996. p. 298–306.
81. Lancet T. India under COVID-19 lockdown. Lancet (London, England). 2020; 395(10233):1315.
https://doi.org/10.1016/S0140-6736(20)30938-7
82. Total Coronavirus Cases in India. Worldometer. Date last accessed: 26th, July, 2021: https://www.
worldometers.info/coronavirus/country/india/
83. Unique identification authority of India; 2020. Date last accessed: 26th, July, 2021: https://uidai.gov.in/
images/state-wise-aadhaar-saturation.pdf.
84. Lamsal R. Design and analysis of a large-scale COVID-19 tweets dataset. Applied Intelligence. 2020;
p. 1–15.
85. Jack RE, Garrod OG, Schyns PG. Dynamic facial expressions of emotion transmit an evolving hierar-
chy of signals over time. Current biology. 2014; 24(2):187–192. https://doi.org/10.1016/j.cub.2013.11.
064
86. Trampe D, Quoidbach J, Taquet M. Emotions in everyday life. PloS one. 2015; 10(12):e0145450.
https://doi.org/10.1371/journal.pone.0145450
87. Cowen AS, Keltner D. Self-report captures 27 distinct categories of emotion bridged by continuous
gradients. Proceedings of the National Academy of Sciences. 2017; 114(38):E7900–E7909. https://
doi.org/10.1073/pnas.1702247114
88. Wintre MG, Vallance DD. A developmental sequence in the comprehension of emotions: Intensity,
multiple emotions, and valence. Developmental psychology. 1994; 30(4):509. https://doi.org/10.1037/
0012-1649.30.4.509
89. Kemper TD. How many emotions are there? Wedding the social and the autonomic components.
American journal of Sociology. 1987; 93(2):263–289. https://doi.org/10.1086/228745
90. Ekman P. An argument for basic emotions. Cognition & emotion. 1992; 6(3-4):169–200. https://doi.
org/10.1080/02699939208411068
91. Lucas RE, Diener E, Larsen RJ. Measuring positive emotions; 2003.
92. Friedman S. Comedy and distinction: The cultural currency of a ‘Good’sense of humour. Routledge;
2014.
93. Rossato L, Chiaro D. Audiences and translated humour: An empirical study. Translation, humour and
the media. 2010; p. 121–137.
94. Kuipers G. The politics of humour in the public sphere: Cartoons, power and modernity in the first
transnational humour scandal. European Journal of Cultural Studies. 2011; 14(1):63–80. https://doi.
org/10.1177/1367549410370072
95. Chey J. Youmo and the Chinese sense of humour. Humour in Chinese life and letters: Classical and
traditional approaches. 2011; p. 1–30.
96. Clasquin M. Real Buddhas don’t laugh: Attitudes towards humour and laughter in ancient India and
China. Social Identities. 2001; 7(1):97–116. https://doi.org/10.1080/13504630120043549
97. Elst K. Humour in Hinduism. Humour and Religion: Challenges and Ambiguities London: Bloomsbury
Academic. 2011; p. 35–53.
98. Chandra R, Jain K, Deo RV, Cripps S. Langevin-gradient parallel tempering for Bayesian neural learn-
ing. Neurocomputing. 2019; 359:315–326. https://doi.org/10.1016/j.neucom.2019.05.082
99. Chandra R, Jain M, Maharana M, Krivitsky PN. Revisiting Bayesian Autoencoders with MCMC. arXiv
preprint arXiv:210405915. 2021.
100. Chandra R, Bhagat A, Maharana M, Krivitsky PN. Bayesian graph convolutional neural networks via
tempered MCMC. arXiv preprint arXiv:210408438. 2021.