Professional Documents
Culture Documents
1 Introduction
In general, a Question Answering (QA) system is a type of information retrieval
system triggered by an input query. In the particular context of this paper, QA
systems should produce direct answers/facts (rather than a ranked list of relevant
documents) for a question expressed in natural language. Accordingly, such QA
systems are regarded as the next generation search engines, since they require
capabilities for query analyzes, recognition of most relevant facts, and deducing
new facts and answers including the use of background or domain knowledge
where necessary.
Automated QA systems are broadly classified into two categories: open-
domain and restricted (or closed) domain systems. Restricted-domain QA deals
2 Qin et al.
and
p → P erson(young) isv M ary
which answer the questions “How old is Mary?” and “Who is young?” re-
spectively. In Zadeh [14], three basic operations on generalized constraints are
proposed (conjunction, projection, and propagation). If fuzzy sets are prede-
fined, these operations can be used to calculate the precisiated implications of
propositions.
A protoform (short for “prototypical form”) is a key concept that facilitates
deduction and reasoning with PNL expressions. Simple NL expressions with a
single “to be” verb phrase (in all conjugated forms, tenses, and modalities) can
be abstracted into respective “X is A” protoform, such as “Mary is young” or
“Price of a golf ball is two dollars”. Similarly, “John is a little older than Mary”
and “Price of a golf club is much more than the price of a golf ball” can be
abstracted to a “Y is X + B” protoform, and “Some balls are blue” can be
abstracted to a “Q1 As are Bs” protoform. Furthermore, if the first two phrases
are recognized as “age(Mary) is young” and “price(golf ball) is two dollars”, they
4 Qin et al.
Each of the forms above can be extended with respect to different modalities
(probabilistic, usuality, bimodal interpolation, fuzzy graph interpolation, etc.).
The deduction engine is designed to operate offline (i.e. without query input) as
well as to be query-driven [14].
3 Deduction Module
The deduction module is a major part of an overall PNL-based QA system.
A corpus of text documents are processed by the information extraction (IE)
module to identify and tag each sentence as a PNL protoform (thus far X is
A, Y is X + B, QAs are Bs, f (X) is A), causal fact (a causes b), if-then
fact, procedure, or simply ‘fact’ (none of the above). Similar techniques are
applied to the input question to identify the query-type (what, where, when,
how, how much/quantity). Details regarding the precisiation process and its
semi-automated implementation in the IE module are described in [1], and the
general system design and process flow of the QA system is discussed in [8].
Modules and processes related to the deduction engine are shown in Figure 1
below. The deduction engine has access to all corpus knowledge which were
precisiated by the IE module, and represented as a collection of fact-types.
If a sentence is a PNL protoform, it will be processed according to PNL rea-
soning. Else, phrase-based deductive reasoning is applied. In the presence of an
input query, our system analyzes the query key phrases and query-type, selects a
subset of relevant facts, and (where possible) combines with background knowl-
edge to generate the most relevant answer set according to a ranking system.
If deductive reasoning and concept matching (between query and facts) yield
results with low ranking values, a search engine (e.g., Lucene [6]) is invoked for
standard keyword-based search, which provides a minimal performance level.
Fig. 1. The general structure of deduction engine for the PNL-based QA system.
module analyzes existing facts to see if new facts can be generated based on
PNL protoforms or phrase-based deductive reasoning. Thus far, standard logic
applied to phrases include: negation, transitivity, and chained reasoning.
– Negation: If we have If-Then rules or causal rules like “If A then B” or
“A causes B” (A → B), we can conclude that “if NOT B then NOT A”
(¬B → ¬A)
– Transitive Reasoning: If we have causal relations A → B and B → C, we
then can conclude that A → C.
– Chaining: chaining is a multi-step transitive reasoning. Given a set of causal
relations: A → B, B → C · · · E → F , we can conclude that A → F .
Semantic interpretation of chained reasoning results may sometimes appear
to be unclear or illogical, hence the deduction module supports an option to
show (explain) all intermediate results.
In addition to standard reasoning above, we also combine X is A facts with
related simple facts. For example, given the facts: (1) Carnivores eat meat. (2)
Lions are carnivores. The second fact is in X is A form but the first one is simply
a fact. The two noun subjects (or subject phrases) are treated as equivalent, and
via substitution of ‘carnivores’ by ‘lions’ in the first sentence, the deduction
module will conclude that “lions eat meat”.
As explained in [8] an application manager may provide background knowl-
edge in the forms of simple facts or abstract facts. During offline reasoning,
the QA system can also generate new knowledge by combing the facts from
a corpus content and background knowledge. For example, in a domain about
telecommunication, the background abstract facts: “wireless *yy is less secure
than wired *yy” and “wireless *zz simplifies physical installation” may be pro-
vided. The notation ‘∗yy’ and ‘∗zz’ are reserved variables, and the deduction
module searches for matches of phrase pattern surrounding the reserved words
6 Qin et al.
in the corpus text. For example, if another sentence contains information about
‘wireless router’, then new sentences “wireless router is less secure than wired
router” and “wireless router simplifies physical installation” are generated al-
though they do not explicitly appear in the corpus text.
We note that such knowledge augmentation may also generate some mean-
ingless sentences - if a phrase “wireless chat session” is found elsewhere, gen-
erated sentence “wireless chat session simplifies physical installation” does not
make sense. However, the generated knowledge are passively stored, and when
a user inputs a query, the online processing helps to filter sensible results. It
is more likely that sentences more relevant to the query will be ranked higher.
For instance, if asked, “what type of routers simplify physical installation?” The
correct answer “wireless router simplifies physical installation” will be ranked
higher than the phrase about wireless chat session.
Offline reasoning improves performance and execution time during online
use. The module operates at the phrase-level; each sentence in the corpus is
segmented into subject phrase, verb phrase, and object phrase, such that the
phrase components can be identified with the X, Y or A, B, C ... components
of the PNL or non-PNL protoforms above. Accurate recognition of matching
components (e.g. X in one subject phrase with X in another object phrase)
depends on the quality of the concept matching module for two input strings.
This module removes stop words and checks synonyms of remaining keywords
(stemmed & non-stemmed) using WordNet [7] to compare phrases at the concept
(rather than string) level. Given a query, the answer generation process flows as
described below.
our QA system will actually add the fuzzy sets for “low” plus “a lot more” and
produce the approximate defuzzified value. If the fuzzy sets are not defined, the
answer remains in the NL expression as “Horsepower of sports model is low
PLUS a lot more” (i.e. a lot more than “low”) which is just a linguistic interpre-
tation of Y is A+B form. In short, pre-defined fuzzy concepts enable computing
more precise responses in the PNL paradigm; otherwise the system will return
a linguistic interpretation of PNL protoforms.
As explained in [1] some “f(X)” protoforms (e.g. cost(a flat screen monitor))
can be identified from phrases like “A flat screen monitor is expensive”, and the
deduction module supports associated computations. Given the following facts:
– A mouse is cheap.
– A computer monitor is two hundred US dollars.
– A keyboard is fifty US dollars.
If asked “what is the cost of a computer monitor plus a keyboard?” the system
returns:
If we ask the cost for a mouse plus a keyboard, the answer will be ‘fifty US
dollars’ PLUS ‘cheap”. The defuzzification value will be returned if ‘cheap’ is
predefined by a fuzzy set. Otherwise, only the linguistic answers as above.
If the relevant facts (to a query) are not PNL protoforms, they could be one of
the following (as tagged by the IE module): causal fact, if-then fact, procedure, or
just “fact”. These facts are processed through the phrase-based deduction pipe,
and the deduction engine (i) generates primary rankings of facts based on the
degree of concept match with the query, (ii) attempts deductive reasoning based
on new constraints supplied in the query, and (iii) generates secondary ranking
based on the query type (if applicable). Deductive reasoning with new con-
straints include negation, transitive and chained reasoning as explained above,
to produce an answer. In many cases, multiple facts are determined to be rel-
evant to a query, as they include the same concepts/key phrases appearing in
the query. Thus, during the secondary ranking process, ranks of causal or if-then
facts are incremented for ‘why’-type question, ranks of procedure facts would be
incremented for ‘how’-type questions, and facts containing quantity terms (nu-
meric or fuzzy terms such as “few” or “most”) are incremented for how-much
type questions. Our system displays N top ranked answers, where N may be
specified by the application manager.
8 Qin et al.
and the query “how many telephones can be connected to Beta?”. Typical
keyword based systems will rank (1) or (2) as the highest answer, but our system
detects that (3) is a X is A form and “controller” and ”the heart of Beta”
is equivalent. Therefore, a new generated fact: “Up to 16 telephones can be
connected to the heart of Beta” is the highest ranked answer. In some cases
where the relevant subset is sparse or the query is ill-formed, and deduction
process can not be applied. If concept match results yield low rankings then the
highest scores of a standard search engine are returned.
4 More Examples
Additional details and output samples below summarize key functionalities of
the deduction module. Given example facts:
and key concepts defined by fuzzy sets: ‘about 150’ is defined by a triangu-
lar fuzzy set centered on 150 with width of the fuzzy set ± 10% (i.e. ± 15 for
this example). The term ‘a lot more’ is defined by a percentage (e.g. 100%) on
the range of the reference fuzzy set ‘about 150’. Since the two facts are PNL
protoforms X is A and Y is X + B respectively, the deduction engine produces
the result Y is A + B, and the following fuzzy addition is performed:
T riF uzzy(150, 15) + 150 × 100% = T riF uzzy(300, 15)
where T riF uzzy(a, b) represents a triangular fuzzy set whose center is at a and
the width is b. The final answer is returned as “about 300” where 300 is the
defuzzified value of the composite fuzzy set. QA system built on standard search
technology cannot provide this type of response. In practice, all key concepts in
a corpus are not likely to be pre-defined with fuzzy set representation, but for a
particular sub-section(s) where the application manager expects many questions
and desires to provide detailed answers, PNL-based computing offers value-added
performance.
Examples of phrase-based deductive reasoning are provided below. For ex-
ample, given two facts:
Deduction Engine Design for PNL-based Question Answering System 9
(i) If over 50 percent of customers complain about new product WRTG54, all types of
WRTG routers will be recalled.
(ii) If all types of WRTG routers are recalled, we will start new sale on TRG100 type
of router.
If the query is: ‘Why did you start new sale of TRG100?’ The system returns
two answers:
(1) If all types of WRTG routers are recalled , we will start new sale on TRG100
type of router.
(2) If over 50 percent of customers complain about new product WRTG54 then we will
start new sale on TRG100 type of router.
If the query is: ‘Why would you not start new sale of TRG100?’. The answers are:
(3) NOT (we will start on a new sale on TRG100 type of router) IMPLIES NOT
(all types of WRTG routers are recalled).
(4) NOT (we will start on a new sale on TRG100 type of router) IMPLIES NOT (over
50 percent customers complain about new product WRTG54.
The second set of answers (3 and 4) result from the transitive implications of
facts (i) and (ii).
5 Summary
mains under development, and further development and testing are planned
using TREC’s QA data sources [9]. Further research topics also include concept
matching, ranking algorithms, different knowledge representation schemes such
as RDF and OWL formats [10], and integration of additional NLP tools such as
entity extractors for analyzing corpus knowledge and query phrases.
Acknowledgements
Qin and Beg are British Telecommunications (BT) Research Fellows in BISC
Group. This research was funded by BT/BISC Fellowship.
References
1. M. M. S. Beg, M. Thint, Z. Qin, Precisiating natural language for a question an-
swering system, submitted to the 11th World Multi Conf. on Systemics, Cybernetics,
and Informatics, 2007.
2. F. Benamara, Cooperative question answering in restricted domains: the WEB-
COOP experiment, Proceedings ACL 2004 Workshop on Question Answering in Re-
stricted Domains, 2004.
3. W. Ceusters W, B. Smith, M. Van Mol, Using ontology in query answering systems:
scenarios, requirements and challenges. In Bernardi R, Moortgat M (Eds.) Proceed-
ings of the 2nd CoLogNET-ElsNET Symposium, Amsterdam, pp.5-15, 2003.
4. H. Chung, Y.-I. Song, K.-S. Han, D.-S. Yoon, J.-Y. Lee, H.-C. Rim and S. -H. Kim,
A practical QA system in restricted domains, 2004.
5. H. Doan-Nguyen, L. Kosseim, The problem of precision in restricted-domain
question-answering: some proposed methods of improvements, Proceedings ACL 2004
Workshop on Question Answering in Restricted Domains, 2004.
6. O. Gospodnetic, E. Hatcher, Lucene in Action, Manning, 2004.
7. G. Miller, Wordnet: a lexical database, Communications of the ACM, 38(11), 1995,
pp. 39-41.
8. M. Thint, M. M. S. Beg and Z. Qin, PNL-enhanced restricted domain question
answering system, submitted to IEEE-FUZZ, London, UK.
9. Available: http://trec.nist.gov/data/qa.html
10. Available: http://www.w3.org.org/TR/owl-ref
11. O. Tsur, M. de Rijke, and K. Sima’an, BioGrapher: biography questions as a re-
stricted domain question answering task, Proceedings ACL 2004 Workshop on Ques-
tion Answering in Restricted Domains, 2004.
12. L. A. Zadeh, A new direction in AI - toward a computational theory of perceptions,
A.I. Magazine, Spring 2001.
13. L. A. Zadeh, From computing with numbers to computing with words − from
manipulation of measurements to manipulation of perceptions, Int. J. Appl. Math.
Comput. Sci., Vol. 12/3: pp. 307-324, 2001.
14. L. A. Zadeh, Toward a generalized theory of uncertainty (GTU) − an outline,
Information Sciences, Vol. 172: pp. 1-40, 2005