2010
2010
IEEE
IEEE
Fourth
International
International
Conference
Conference
on Semantic
on Semantic
Computing
Computing
Reference Resolution Supporting Lexical
Disambiguation
Marjorie McShane, Stephen Beale and Sergei Nirenburg
Department of Computer Science and Electrical Engineering
University of Maryland Baltimore County
Baltimore, Maryland, USA
{marge,sbeale,sergei}@umbc.edu
underspecification and ambiguity that parallel our methods of
treating residual lexical ambiguity.
Abstract— This paper describes ongoing work in carrying out the
semantic analysis of texts and reference resolution in a control
structure that permits each process to inform the other, rather
than in a more traditional, unidirectional fashion (semantics
followed by reference resolution). We concentrate on situations in
which a polysemous predicate cannot be lexically disambiguated
until the meaning of one of its arguments has been specified, and
that can only be accomplished with the help of reference
resolution procedures. As a sidebar, we briefly introduce our
“feature value bundling” approach to configuring reference
resolution engines without the need for large annotated corpora.
In the OntoSem environment, the text analyzer takes as
input raw text and carries out its tokenization and
morphological, syntactic, semantic and pragmatic analysis to
yield text meaning representations. Text analysis in OntoSem
relies on: the OntoSem language-independent ontology of over
8,000 concepts, each of which is described by a large number
of properties whose values can be locally defined or inherited;
the OntoSem lexicon of English of about 35,000 senses that
contains linked syntactic and semantic zones, the latter of
which uses ontological concepts to describe word meaning;
agent memory, also called the fact repository; the OntoSem
text analyzers; and the text meaning representation language
itself, which is the unambiguous metalanguage for representing
text meaning in all resources and in the automatically generated
text meaning representations.
Keywords-NLP; semantics; reference resolution
I.
INTRODUCTION
This paper presents our ongoing work on interleaving the
processing of semantic analysis and reference resolution such
that each can inform the other in the most beneficial way to
support the reasoning of human-like intelligent agents. The
paper begins with some definitions, background notes about the
OntoSem environment, and our basic approach to lexical
disambiguation and reference resolution. We then present the
algorithm for resolving our selected class of phenomena,
followed by examples of its usage in the system. We conclude
with future directions of work.
Our basic approach to lexical disambiguation is to use
mutual constraints of predicates and their arguments in a
bidirectional way. For example, the unambiguous meanings of
the subjects and direct objects in (1) and (2) permit the analyzer
to automatically understand that the highly ambiguous verb
have in (1) means the event INGEST, whereas in (2) it is used as
a light verb that, in conjunction with the direct object migraine,
means the event MIGRAINE.
We define semantic analysis as the interpretation of text
meaning as rendered using an unambiguous metalanguage – in
our case, the text meaning representation language of the
OntoSem environment [1]. We define reference resolution as
anchoring the meaning of each referring expression (RE) in the
mental model of the intelligent agent processing the text, such
that new information about entities can supplement or amend
old information (if any), leading to memory population not
unlike what a human would carry out [2]. Our treatment of
reference, therefore, goes beyond the typical coreference task
of NLP in several ways: we treat all referring expressions, not
just the subset of pronouns typically treated in the well-known
pronominal coreference task [3]; we go beyond establishing
textual coreference to anchoring meaning in an agent’s
memory; we seek to achieve full understanding of coreference
relations across texts, as realized by anchoring coreferential
instances of entities and events to the same anchors in memory;
and we are developing methods of treating residual referential
1. The woman had a burger.
2. The woman had a migraine.
What makes this disambiguation possible is the fact that the
lexicon includes different senses of the word have that include
mutually exclusive semantic constraints: one expects the
THEME to be an INGESTIBLE whereas another expects the
THEME to be a DISEASE or SYMPTOM (there are many more
sense of have as well, many of them constructions and idioms).
Similarly, an unambiguous predicate head can imply the
meaning of its arguments: e.g. in (3), even though the system
cannot know what the invented word trala means, it knows that
it must be some SURFACE-OF-OBJECT because the ontological
concept TILE-EVENT (the only meaning of ‘tile’ in our lexicon)
includes the specification that the THEME of TILE-EVENT is
SURFACE-OF-OBJECT.
3. The workmen tiled the trala.
978-0-7695-4154-9/10 $26.00 © 2010 IEEE
DOI 10.1109/ICSC.2010.103
56
a confident semantic unification with one or more meanings
of the selecting predicate
And if the underspecified referring expression can be
resolved confidently using non-semantic methods
(i.e., high-confidence, “surfacy” feature bundles)
Then resolve the referring expression and use its
meaning to help disambiguate the predicate
… ; we do not discuss additional conditions in this paper
For further details on semantic analysis in OntoSem, see [4],
Our basic approach to reference resolution is what we call
“feature value bundling” [5]. We have compiled a large
inventory of features of REs that are relevant for predicting
which of the candidate sponsors for a referring expression is
the actual sponsor. These features range from “surfacy” (e.g.,
the respective gender and number of the RE and the candidate),
to syntactic (e.g., the respective syntactic functions of the RE
and the candidate), to semantic (e.g., the meaning of the events
selecting the candidate and the RE as their case-roles) to
pragmatic (e.g., the location of speaker changes in a dialog).
We then manually create combinations of coreferencepromoting feature values – what we call “feature value
bundles” – that we believe will have significant power to
predict the correct sponsor. We then vet those hypotheses using
a corpus and assign each bundle a score indicating the
confidence of prediction attested by the corpus. This feature
value bundling methodology was originally created to support
research and development on the more difficult and less studied
referring expressions, like it, this and that, for which no
adequate annotated corpora exist; however, we are finding it
useful for a wide range of referring expressions. As in many
approaches to reference resolution, our system keeps a running
list of REs that might function as sponsors for later REs;
however, unlike most systems, ours stores the candidates both
as strings and as instances of ontological concepts. As such, our
heuristics can refer to both the form and the meaning of
referring expressions.
As promised above, we will use an example-driven method of
illustrating the algorithm since we believe that examples can be
the best way to succinctly and informally convey how a system
operates. We will first walk through one example, then present
a number of other examples with only minimal commentary.
At a first glance, one might not even detect the lexical or
referential ambiguity in an example like (4) since we as people
resolve ambiguity so effortlessly. However, for a system
attempting to disambiguate every aspect of an input, the
challenges of resolving polysemous save and underspecified it
are formidable.
(4) A dialog box will open and ask if you want to open
the file, save it or cancel.
Lexically, save has at least 3 meanings: ‘rescue from harm’,
‘store in a digital file’ and ‘store for the future’. It can refer to
‘a dialog box’, ‘the file’, or even the propositions ‘a dialog box
will open’, ‘ask if you want to open the file’, or ‘you want to
open the file’. (Reference to propositions, realized as spans of
text, has received relatively little attention in NLP but is a
prominent phenomenon in language use; see, e.g., [2] and [5].)
The nature of the work reported here is linguistic analysis
aimed at near-term implementation for the natural language
support of cognitively human-like intelligent agents
functioning in working applications. Although there are many
phenomena related to the interaction between lexical
disambiguation and reference resolution, here we will
concentrate on just one: situations in which a polysemous
predicate cannot be lexically disambiguated until the meaning
of one of its arguments has been specified, and that can only be
accomplished with the help of reference resolution procedures.
We will use an example-based methodology of description and
minimize the use of formalism in order to emphasize that the
approach is not system-specific but, rather, is likely to be
required by any system seeking to carry out both semantic
analysis and reference resolution.
II.
Table 1 shows the three verbal senses of save that are
recorded in our lexicon. The middle column presents a very
abbreviated version of their semantic descriptions that includes
only those aspects of meaning central to an understanding of
this disambiguation task. For example, the description of savev1 indicates that this word sense has the meaning of the
concept RESCUE and that its THEME should be some kind of
ANIMAL – more specifically, any lexical item mapped to the
concept ANIMAL or any of its descendants. The third column
provides an example of usage, since the meaning of ontological
concepts cannot properly be understood without consulting the
concept’s description in the ontology. (Although concept
names look like and are typically similar to the meaning of
English words, they are not English words.)
THE ALGORITHM AND EXAMPLES
The processing algorithm we will be discussing is as follows:
Table 1: Senses of save
If a predicate (verb) cannot be confidently disambiguated using
bidirectional constraints with its arguments
And if one or more of its arguments is an underspecified
referring expression (e.g., it)
And if exactly one resolution of the pronoun (based on
coreference with candidate sponsors) leads to a confident
semantic unification with exactly one of the meanings
of its selecting predicate
Then establish the given coreference and use the unifying
predicate analysis
Else
If more than one resolution of the pronoun leads to
v1
v2
RESCUE (THEME ANIMAL)
v3
STORE-FOR-FUTURE (THEME OBJECT)
SAVE-COMPUTER-DATA (THEME
COMPUTER-DATA, COMPUTER-FILE)
save a bear cub
save a file
save a seashell
When the analyzer encounters (4), all three meanings of save
are available. Since one of the arguments of save is an
underspecified referring expression (it), the analyzer will
attempt every available resolution of it – using the meanings of
the candidate sponsors in the candidate list – and see if any of
them makes a strong suggestion about what save means.
57
The analyzer will begin by coreferring it with dialog box,
which is semantically analyzed as COMPUTER-DIALOG-BOX.
This meaning does not meet the narrow constraints on the
THEME of save-v1 or save-v2 because COMPUTER-DIALOG-BOX
is not a descendant of ANIMAL, COMPUTER-DATA or COMPUTERFILE. This meaning does meet the broad constraint on the
THEME of save-v3, making this sense selection a viable option;
however, it is not a confident option because there is a great
ontological distance between the very specific concept
COMPUTER-DIALOG-BOX and the very general constraint
OBJECT. Next the analyzer will corefer it with the file, which
was disambiguated in its own clause as COMPUTER-FILE.
COMPUTER-FILE is a direct match of a selectional constraint for
the THEME of sense 2 and offers a high-confidence coreference
link that will be selected over the low-confidence link offered
by sense 3. The analyzer will not, in the case, consider the text
span propositions to be viable candidates because propositions
typically refer to EVENTs and none of our senses of save
expects an EVENT as its THEME. In sum, for example (4)
semantic analysis is sufficient to both resolve the meaning of
the referring expression and choose a meaning of the selecting
verb.
The remaining examples are presented using the same
formalism as above.
(5) Combat stress is a natural result of the heavy mental
and emotional work required when facing danger in
tough conditions. Like physical fatigue and stress,
handling combat stress depends on the level of your
fitness/training. It can come on quickly or slowly, and it
gets better with rest and replenishment.
Table 2. Senses of get better
v1
v2
-
His sax playing got
better.
His cough got better.
Analysis: If it is coreferential with combat stress (COMBATwhich is a type of DISEASE), then HEAL (get better-v2)
has a perfect filler for its THEME case-role.
STRESS,
A corroborating strongly predictive feature bundle:
- C is a matching pronoun
- C is the most recent candidate that matches in
gender/number/animacy
- C and RE have matching syntactic functions (both are
subjects)
- C and RE are in a VP conjunction structure with
‘and’
- long chain of coreference (C is part of a 3-member
chain even before coreferring with RE)
However, imagine that the bidirectional semantic correlation
of the head and its argument could not confidently suggest
exactly one resolution for example (4). In that case, the system
would first attempt to resolve the pronoun using “surfacy”
heuristics combined in the feature bundling strategy; if
successful, it would use the pronoun’s meaning to
unidirectionally impose a meaning on its selecting predicate.
As it turns out, our example matches a feature value bundle that
that has been attested to have very high predictive power of
coreference:
-
represented as an increase in the
value of evaluative modality; its
THEME is any EVENT
HEAL (THEME ANIMAL, DISEASE)
(6) When it comes to the causes of autism, here are the
facts: we know it runs strongly in families, although
it is not strictly inherited like muscular dystrophy or
hemophilia.
C (the candidate) is the most recent candidate that
matches the RE in gender/number/animacy
C and RE have matching syntactic functions
C and RE are in a VP conjunction structure
C and RE have matching case-roles (both are THEMEs
under any semantic interpretation of the predicate)
Table 3. Senses of inherit
v1
INHERIT-GOODS (THEME FINANCIAL-
inherit a fortune
OBJECT)
This feature value bundle would confidently select the file
(COMPUTER-FILE) as the resolution of the pronoun it. Once the
meaning of it was established, it would be matched to the
constraint on the theme of SAVE-COMPUTER-DATA in sense 2,
and that meaning would be unidirectionally imposed on the
predicate save.
v2
INHERIT-GENETICALLY (THEME GENE,
DISEASE, CHARACTERISTIC, etc.)
inherit cystic
fibrosis
Analysis: If it is coreferential with autism (AUTISM, which is a
type of DISEASE), then INHERIT-GENETICALLY (inherit-v2) has
a perfect filler for its THEME case-role.
At this point, one might ask, Why not always use the
computationally less expensive feature-bundling strategy first,
before resorting to more expensive semantic analysis? One
could, but we choose not to because (a) in our environment all
texts are processed semantically anyway, so unless we do some
selective processing of sentences in a corpus, we will always
have an antecedent list containing both strings and concepts;
(b) we believe that semantic evidence, when available, is
stronger and more certain than any other kind of evidence; and
(c) in some contexts there will be no available feature bundles
that can confidently predict the resolution of the pronoun in
isolation.
A corroborating strongly predictive feature bundle:
- C is a matching pronoun
- C is the most recent one that matches in
gender/number/animacy
- C and RE have matching syntactic functions (both are
subjects)
- C and RE are in a main/subordinate relationship
- chain of coreference (C is part of a 2-member chain
even before coreferring with RE)
(7) Primitive Medicine is timeless. It is as old as the
Paleolithic cave-dwellers. It is as new as today. Early
58
people. As such, we do not draw hard lines between
traditionally divided realms like syntactic analysis, word sense
disambiguation and reference resolution. Due to feasibility
constraints, we cannot, it is true, allow heuristics from all
modules of text processing to fire at once: e.g., some part-ofspeech decisions are made before syntactic analysis is
launched, less probable syntactic parses are removed before
semantic analysis occurs, etc. However, we attempt to postpone
difficult cases of upstream decision-making until semantics can
act as an arbiter. The same is true of reference resolution. It
would be infeasible to attempt to resolve reference without the
benefit of any semantic analysis decisions having been made;
however, this does not mean that semantic analysis must be
completed, with no outstanding options, before reference
resolution is attempted. A real key to achieving outstanding
text analysis, we believe, is to be able to automatically evaluate
confidence in each stage of analysis, and leave low-confidence
decisions open until later stages of processing can register a
vote. At the time of writing, we are working on developing
such confidence-assigning engines for each stage of processing.
evidences of its practice can be traced back 10,000
years. Yet it is being practiced in some part of the
world at this very hour…
Table 4. Senses of practice
v1
v2
v3
PRACTICE (THEME EVENT)
PLAY-MUSICAL-INSTRUMENT
(THEME MUSICAL-INSTRUMENT)
HAS-RELIGION (RANGE
practice soccer
practice the trumpet
practice Catholicism
RELIGION)
v4
WORK-ACTIVITY (THEME FIELDOF-STUDY)
practice allopathic
medicine
Analysis: If it is coreferential with primitive medicine (FIELDOF-MEDICINE, which is a type of FIELD-OF-STUDY), then
WORK-ACTIVITY (practice-v4) has a perfect filler for its THEME
case-role.
A corroborating strongly predictive feature bundle:
- long chain of coreference (C is part of a 4-member
chain even before coreferring with RE)
- most members of the chain have matching syntactic
function (subject)
We have been using our text processing capabilities in realworld applications – most recently, in dialog systems in the
medical domain [6]-[7]. The applications in question – a
medical education system called Maryland Virtual Patient and
a CLinician’s ADvisor called CLAD – are prime examples of
applications for which high-quality, deep text understanding
are needed, the processing of difficult phenomena cannot be
postponed, and the returns of developing methods for
effectively treating difficult phenomena should be great. In
short, the work reported here is not being carried out in a
conceptual bubble – it is being incorporated into working,
forward-looking systems.
(8) [Lord Illingworth]: Then why does he write to me?
[Mrs. Arbuthnot]: What do you mean? [Lord
Illingworth]: What letter is this? [Mrs. Arbuthnot]:
That—is nothing. Give it to me. [Lord Illingworth]: It is
addressed to me. [Mrs. Arbuthnot]: You are not to open
it.
Table 5. Senses of open
v1
OPEN (THEME BAG, WINDOW,
open a letter
ENVELOPE, LETTER, etc.)
v2
v3
cause to BE-AVAILABLE
begin + EVENT
REFERENCES
open a road
open a conference
(with a speech)
[1]
Analysis: If it is coreferential with the chain of coreferred
elements meaning LETTER, then OPEN (open-v1) has a perfect
filler for its THEME case-role.
[2]
[3]
A corroborating strongly predictive feature bundle:
- C is a matching pronoun
- C is the most recent one that matches in
gender/number/animacy
- long chain of coreference (C is part of a 4-member
chain even before coreferring with RE)
[4]
[5]
[6]
III.
DISCUSSION
One might say that a main thread in the overall program of
research and development in the OntoSem environment is that
all aspects of natural language processing are connected and
are aimed at populating an agent’s memory so that it can carry
out sophisticated reasoning with the results mimicking those of
[7]
59
Nirenburg, S., Raskin, V., 2004. Ontological Semantics, The
MIT Press, Cambridge, Mass.
McShane, Marjorie, 2009. Reference resolution challenges for
an intelligent agent: The need for knowledge. IEEE Intelligent
Systems, vol. 24, no. 4, pp. 47-58.
Hirshman, L., Chinchor, N., 1998. MUC-7 coreference task
definition. Version 3.0. Proceedings of the Seventh Message
Understanding Conference (MUC-7). Applications International
Corporation.
Beale, S., 1997. Using branch-and-bound with constraint
satisfaction in optimization problems. Proc. AAAI-97.
Providence, Rhode Island.
McShane, Marjorie. 2009, Advances in difficult aspects of
reference resolution: Working Notes. ILIT Working Paper #0109,
Nov.
18,
2009
(62
pp.)
Available
at
http://ilit.umbc.edu/MargePub/SGER-As-Paper.pdf.
Nirenburg, S., McShane, M., Beale, S., 2008. A simulated
physiological/cognitive “double agent”. In: Workshop on
Naturally Inspired Cognitive Architectures at AAAI 2008 Fall
Symposium.
McShane, M., Nirenburg, S., Beale, S., 2008. Two kinds of
paraphrase in modeling embodied cognitive agents. In:
Workshop on Biologically Inspired Cognitive Architectures,
AAAI 2008 Fall Symposium.