2020 NLPDeepLearning
2020 NLPDeepLearning
2020 NLPDeepLearning
AVANÇOS RECENTES NO
PROCESSAMENTO DE
LINGUAGEM NATURAL COM DEEP
LEARNING
08/out/2020
Summary
2
Introduction
NLP Periods
Symbolic based
Corpus-based
Neural-based
Conclusions
3
Introduction
What is NLP?
4
Homonymy, polysemy, …
Jaguar is the luxury vehicle brand of Land Rover.
1950s-1990s
Georgetown-IBM experiment
11
1966
Cyc Project
13
1984-
WordNet
14
155,327 words
organized in
175,979 synsets for
a total of 207,016
word-sense pairs
1985-
WordNet – graph fragment
15
chicken
Is_a poultry Purpose supply Typ_obj
clean Is_a Quesp
smooth Typ_obj keep
Is_a hen duck
Is_a
Typ_obj Purpose meat
preen Typ_subj Caused_by
Is_a egg
Means quack
Not_is_a plant
chatter Typ_subj animal
Is_a Is_a
Is_a Is_a creature
make bird Is_a
Typ_obj sound
gaggle Part feather
Is_a Is_a
Classifier goose wing Is_a limb
peck Is_a
number Typ_subj Is_a
claw
Is_a Means Is_a
beak Part Part
hawk Is_a
Typ_obj
strike Typ_subj
fly
leg
turtle catch
Is_a Typ_subj Is_a
bill arm
face Location mouth Is_a opening
16
Corpus-based NLP (1990s-2010s)
Corpus-based NLP
17
1990s-2010s
Corpus-based NLP (aka ML-based)
18
1990s-2010s
Corpus-based NLP (aka ML-based)
19
Text Mining
1990s-2010s
20
Neural-based NLP (2010s-present)
21
Conception, gestation, …, birth!
22
1950s
Distributional Hypothesis
24
1950s
Vector Space Model (for Information Retrieval)
25
1986
Latent semantic analysis (LSA)
28
1988
Latent semantic analysis (LSA)
29
1988
Latent semantic analysis (LSA)
30
1988
Distributed representation – an example
31
Conception, gestation
Distributional hypothesis
Vector Space model
LSA
Distributed representations
Now, for the Deep Learning based NLP
birth…
Neural-based NLP (aka Deep Learning based)
35
Skip-gram NN
architecture
The amount of neurons in the hidden layer (a hyperparameter) determines de size of the embedding.
word2vec
43
word2vec
44
Credits: http://jalammar.github.io/illustrated-word2vec/
Embedding models
46
Word2Vec
GloVe Currently, the distributional hypothesis through vector
embeddings models generated by ANNs is used
SkipThoughts pervasively in NLP.
Paragraph2Vec
Doc2Vec
FastText
Encoder-Decoder models (aka seq2seq models)
47
Encoder
Decoder
https://pytorch.org/tutorials/intermediate/seq2seq_translation_tutorial.html
“Classical” Encoder-Decoder model
48
ATTENTION
2018-2020
54
Conclusions
Take away notes
55
https://code.google.com/archive/p/word2vec/
https://github.com/google-research/bert
Neural Nets need a Vapnik!
57
Headlines:
Enraged Cow Injures Farmer With Ax
Hospitals Are Sued by 7 Foot Doctors
Ban on Nude Dancing on Governor’s Desk
Iraqi Head Seeks Arms
Local HS Dropouts Cut in Half
Juvenile Court to Try Shooting Defendant
Stolen Painting Found by Tree
Humans use their underlying understanding of the world as context
Source: CS188
TODO: Common Sense Knowledge
59
Geoff Hinton
These slides are available at
http://eic.cefet-rj.br/˜ebezerra/
An example:
"The quick brown fox jumps over the lazy
dog."
Another example:
Unigram model
Bigram model