Professional Documents
Culture Documents
Sewon Min1,2
, Julian Michael1, Hannaneh Hajishirzi1,3
, Luke Zettlemoyer1,2
1
University of Washington 2 Facebook AI Research 3 Allen Institute for Artificial Intelligence
{sewon,julianjm,hannaneh,lsz}@cs.washington.edu
Abstract
Ambiguity is inherent to open-domain ques-
tion answering; especially when exploring
arXiv:2004.10645v2 [cs.CL] 5 Oct 2020
Entity references How many sacks does clay matthews have in his career?
Q: How many sacks does Clay Matthews Jr. have in his career? / A: 69.5
(23%)
Q: How many sacks does Clay Matthews III have in his career? / A: 91.5
Answer types Who sings the song what a beautiful name it is?
Q: Which group sings the song what a beautiful name it is? / A: Hillsong Live
(16%)
Q: Who is the lead singer of the song what a beautiful name it is? / A: Brooke Ligertwood
Time- When does the new family guy season come out?
dependency Q: When does family guy season 16 come out? / A: October 1, 2017
Q: When does family guy season 15 come out? / A: September 25, 2016
(13%)
Q: When does family guy season 14 come out? / A: September 27, 2015
Multiple Who was british pm and viceroy during quit india movement?
sub-questions Q: Who was british viceroy during quit India movement? / A: Victor Hope
(3%) Q: Who was british pm during quit India movement? / A: Winston Churchill
Table 1: Breakdown of the types of ambiguity in 100 randomly sampled items from the A MBIG NQ development
data. Each example may fall into multiple categories.
navigate, and read multiple Wikipedia pages to an open-domain question, along with disam-
find as many answers as possible. The high preva- biguated questions to differentiate them.
lence of ambiguity makes the task difficult even
2. We construct A MBIG NQ, a dataset with
for human experts; it is inherently difficult to know
14,042 annotations on NQ- OPEN questions
if you have found every possible interpretation of
containing diverse types of ambiguity.
a question. Nonetheless, we are able to collect
high quality data covering high levels of ambigu- 3. We introduce the first baseline models that
ity (2.1 distinct answers per question on average) produce multiple answers to open-domain
with high estimated agreement (89.0 F1) on valid questions, with experiments showing their ef-
answers. The types of ambiguity are diverse and fectiveness in learning from our data while
sometimes subtle (Table 1), including ambiguous highlighting avenues for future work.
entity or event references, or ambiguity over the an-
swer type; many are only apparent after examining 2 Related Work
one or more Wikipedia pages. Open-domain Question Answering requires a
To establish initial performance levels on this system to answer any factoid question based on
data, we present a set of strong baseline methods. evidence provided by a large corpus such as
We extend a state-of-the-art QA model (Karpukhin Wikipedia (Voorhees et al., 1999; Chen et al., 2017).
et al., 2020) with three new components: (1) Existing benchmarks use questions of various
set-based question answering with a sequence-to- types, from open-ended information-seeking (Be-
sequence model, (2) a question disambiguation rant et al., 2013; Kwiatkowski et al., 2019; Clark
model, and (3) a modification to democratic co- et al., 2019) to more specialized trivia/quiz (Joshi
training (Zhou and Goldman, 2004) which lever- et al., 2017; Dunn et al., 2017). To the best of our
ages the partial supervision available in the full knowledge, all existing formulations assume each
NQ- OPEN dataset. We also do an ablation study question has a single clear answer.
and qualitative analysis, which suggest there is sig- Our work is built upon an open-domain ver-
nificant room for future work on this task. sion of NATURAL Q UESTIONS (Kwiatkowski et al.,
To summarize, our contributions are threefold. 2019), denoted NQ- OPEN, composed of questions
posed by real users of Google search, each with
1. We introduce A MBIG QA, a new task which an answer drawn from Wikipedia. NQ- OPEN
requires identifying all plausible answers to has promoted several recent advances in open-
domain question answering (Lee et al., 2019; Asai q whose answer is unambiguously yi . We consider
et al., 2020; Min et al., 2019a,b; Guu et al., 2020; two subtasks.
Karpukhin et al., 2020). Nonetheless, Kwiatkowski
Multiple Answer Prediction. Given a question
et al. (2019) report that the answers to such ques-
q, output a set of semantically distinct and equally
tions are often debatable, and the average agree-
plausible answers y1 , . . . , yn , where n is unknown.
ment rate on NQ- OPEN test data is 49.2%,1 in large
part due to ambiguous questions. In this work, Question Disambiguation. Given q and a set of
we embrace this ambiguity as inherent to informa- answers y1 , . . . , yn , generate disambiguated ques-
tion seeking open-domain QA, and present the first tions x1 , . . . , xn , where each xi is a minimal edit
methods for returning sets of answers paired with of q which makes it unambiguous so that yi is a
different interpretations of the question. correct answer and all yj for all j 6= i are incorrect.
Clarification Questions have been used to study When n = 1, this task is trivial, as x1 = q.
question ambiguity in other settings. Research on We choose to represent ambiguity with a set of
community Q&A (Braslavski et al., 2017; Rao and disambiguated questions because it is well-defined,
Daumé III, 2018, 2019) studies finding underspec- immediately human-interpretable, and allows for
ification in the question, but it does not find the straightforward annotation of a wide range of am-
answer to the original question. In recent work, Xu biguities without complex guidelines.
et al. (2019) study clarification of questions that
3.2 Evaluation Metrics
are intentionally annotated with pre-specified en-
tity reference ambiguities. Aliannejadi et al. (2019) To evaluate model performance, we present several
and Zamani et al. (2020) use clarification questions ways to compare a model prediction with m
to refine intents of simple query logs without im- question-answer pairs (x1 , y1 ), . . . , (xm , ym )
mediately apparent information needs (e.g., single with a gold reference set with n pairs
keywords like dinosaur2 ). (x̄1 , Ȳ1 ), . . . , (x̄n , Ȳn ). Since there may be
In contrast, we study open-domain factoid ques- more than one way to refer to a single answer
tions asked by real users: these present clear infor- (e.g., Michael Jordan and Michael Jeffrey Jordan)
mation needs, but carry diverse naturally occurring each gold answer Ȳi is a set of acceptable answer
ambiguities (see Table 1). Furthermore, instead of strings, where all Ȳi are disjoint.
prolonging the user’s information-seeking session We assign each predicted question-answer pair
with clarification questions, our task formulation (xi , yi ) a correctness score based on a string simi-
provides a complete and immediate solution with larity function f valued in [0, 1].
unambiguous rewrites of the original question.
ci = max I[yi ∈ Ȳj ]f (xi , x̄j ).
1≤j≤n
Question Rewriting is a novel, well-defined task
which we propose for differentiating distinct an- Intuitively, ci considers (1) the correctness of the
swers. To the best of our knowledge, it has not answer and (2) the similarity f (xi , x̄j ) between the
been studied for resolving ambiguity; we are only predicted and reference question. We calculate F1
aware of Elgohary et al. (2019) which use question treating the ci as measures of correctness:
rewriting to convert conversational questions into
P P
self-contained questions. i ci ci
precf = , recf = i ,
m n
3 Task: A MBIG QA 2 × precf × recf
F1f = .
3.1 A MBIG QA Setup precf + recf
Figure 1 depicts the A MBIG QA task. The input We consider three choices of Ff . F1ans is the
is a prompt question q, and the output is a list F1 score on answers only, where f always yields 1.
of n question-answer pairs (x1 , y1 ), . . . , (xn , yn ), This may be used without the question disambigua-
where each yi is an equally plausible answer to q, tion step. F1BLEU accounts for string similarity
and each xi is a minimally edited modification of between questions, calculating f with BLEU (Pa-
1 pineni et al., 2002). F1EDIT-F1 uses E DIT-F1 as
The NQ- OPEN test data has 5-way annotations; we com-
pute their pairwise agreement based on string match. f , where E DIT-F1 is a new measure that repre-
2
The average query length in Zamani et al. (2020) is 2.6. sents each disambiguated question by its added and
deleted unigrams compared to the prompt ques- Split # data
# QAs %
tion, and computes the F1 score between them. 1 2 3 4+
For example, consider the prompt question “Who Train 10,036 53 24 14 10
made the play the crucible?”, the reference “Who Dev 2,002 49 23 14 13
wrote the play the crucible?” and the prediction Test 2,004 44 24 16 16
“Who made the play the crucible in 2012?”. The
Table 2: Data statistics. For the number of QA pairs (#
gold edits3 here are * -made , +wrote + while QAs), the minimum is taken when there are more than
the predicted edits are * +in , +2012 +. Their 1 accepted annotations.
E DIT-F1 is thus zero, even though the questions
are similar. Unlike BLEU which we use to directly
measure similarity to the gold question, this metric instruct workers to remove the time-dependence by
only gives credit for getting the key semantic dif- rewriting the prompt question for up to three most
ferences correct between the original question and recent events before Jan 1, 2018, e.g., “When does
the clarification. family guy season 16 come out?” (see Table 1).
30 30
%
%
20 20
10 10
0 0
1 2 3 4+ 1 2 3 4+
# of viewed Wikipedia pages # of search queries
(c) Word cloud of the edits made in ques-
(a) Number of unique Wikipedia (b) Number of search queries written tions; and indicate added and deleted
pages visited by crowdworkers.† by crowdworkers. unigrams, respectively.
Figure 2: Data Analysis on the development data. † This is actually an underestimate; we could not track when
annotators viewed pages by following hyperlinks for technical reasons.
indicates a high rate of ambiguity in NQ- OPEN, can easily disambiguate entity or event references
even though previous work has studied it with the or remove time dependence. Wh-word changes are
assumption that each question has a single answer. also common, especially for specifying the answer
We also find a discrepancy between development type (e.g., from who to which group; see Table 1).
and test; this is likely due to the way in which NQ- The distribution of edits is fairly long-tailed, with
OPEN is constructed, which over-samples difficult the 100 most frequent edits covering 36% of the
questions in the test set (see Appendix B for de- total, and the top 1,000 covering 69%.
tails). The training set contains relatively fewer
ambiguous examples (47%), presumably because 5 Model
using only one worker per training example yielded
slightly lower recall. To set initial performance levels on A MBIG NQ,
we present a baseline A MBIG QA model combin-
Types of ambiguity. Table 1 shows a breakdown ing ideas from recent advances in open-domain
of the types of ambiguity in A MBIG NQ. They are QA (Karpukhin et al., 2020) and generation (Lewis
diverse, including ambiguity in entity references, et al., 2020). Given a prompt question q, our model
event references, properties, and answer types, with predicts answers y1 ..yn , and generates correspond-
a relatively uniform distribution between them. In ing questions x1 ..xn conditioning on q, the answers
comparison to Xu et al. (2019), who intention- y1 ..yn , and the evidence passages. A novel co-
ally elicit questions with ambiguous entity refer- training step also allows the model to leverage the
ences, our analysis shows that unintended ambi- partial supervision available in NQ- OPEN.
guity comes from diverse sources. In many cases,
ambiguity is not apparent from the prompt question Multiple Answer Prediction. Here we describe
alone, but only after researching the question on S PAN S EQ G EN, our model for multiple answer pre-
Wikipedia, as evidenced by differences in model diction. Following Karpukhin et al. (2020), a state-
performance (Section 6.2). of-the-art model on NQ- OPEN, S PAN S EQ G EN first
retrieves 100 passages with a BERT-based (Devlin
Annotator behavior. Figures 2a and 2b show the
et al., 2019) dual encoder, and reranks them us-
number of unique Wikipedia pages and the number
ing a BERT-based cross encoder. Then, instead of
of search queries used by workers during annota-
predicting an answer span from the top 1 passage
tion. More often than not, workers used multiple
as Karpukhin et al. (2020) does, S PAN S EQ G EN
queries and navigated multiple Wikipedia pages,
uses another sequence-to-sequence model based on
showing how our setup captures ambiguity in the
BART (Lewis et al., 2020). Specifically, it condi-
retrieval step of open-domain question answering,
tions on the concatenation of q and the top passages
which is missed in approaches that assume a pre-
in order up to 1024 tokens, and sequentially gener-
specified evidence document.
ates distinct answers token-by-token, separated by
Distribution of edits. Figure 2c shows unigram [SEP]. We pretrain S PAN S EQ G EN on NQ- OPEN
edits made to questions in the development data, and finetune it on A MBIG NQ.
where we remove stopwords except wh-words and We develop S PAN S EQ G EN primarily because
group numeric values by the number of digits. Karpukhin et al. (2020) is designed for generating
Adding numerals such as years is common, as they a single answer, but S PAN S EQ G EN also boosts the
Algorithm 1 Democratic co-training with weak 6 Experiments
supervision (Section 5).
We describe the baseline models used in our ex-
1: // Each question in Dfull has an answer list annotated
2: // Each question in Dpartial has one answer annotated periments, followed by results and ablations. Im-
3: D̂full ← Dfull plementation details and hyperparameters of all
4: for iter ∈ {1..N } do models are provided in Appendix D.
5: // Train C sequence-to-sequence QA models
6: for i ∈ {1..C} do
6.1 Baselines
7: φi ← train(D̂full )
8: D̂L ←Dfull D ISAMBIG - FIRST. This baseline disambiguates
9: for (q j , y j ) ∈Dpartial do the prompt question without any context from plau-
10: // Get predictions by using yj as prefix
11: Yˆj ← {ŷ | ŷ 6= y j , and
sible answers or reference passages. Specifically,
12: |{i | ŷ ∈ φi (q j |y j ), 1 ≤ i ≤ C}| > C2 it implements the following pipeline: (1) Feed the
13: } prompt question q into a BERT-based binary clas-
14: if |Ŷ j | > 0 then sifier to determine whether it is ambiguous. (2) If
15: // Add it as a multiple answer case
16: D̂full ← D̂L ∪ {(q j , {y j } ∪ Yˆj )}
q is ambiguous, pass it into a BART-based model
17: else if ∀i = 1..C, |φi (xj ) − {y j }| = 0 then which generates a sequence of disambiguated ques-
18: // Add it as a single answer case tions x1 ..xn (n > 1), separated by [SEP]; other-
19: D̂full ← D̂L ∪ {(q j , {y j })} wise, consider only x1 = q. (3) Feed each xi into
a state-of-the-art model on NQ- OPEN (Karpukhin
et al., 2020) to produce its answer yi .
performance on NQ- OPEN (41.5→42.2 on the test
data). We include ablations on different approaches Thresholding + QD. We also include a model
and models in Section 6.2. based on Karpukhin et al. (2020), with thresholding
for multiple answer prediction and our question dis-
Question Disambiguation. We design a ques- ambiguation (QD) model. Karpukhin et al. (2020)
tion disambiguation (QD) model based on BART. outputs a likelihood score for each span; we obtain
The model generates each question xi (i = 1..n) y1 ..yn by taking valid spans with likelihood larger
conditioning on the concatenation of q, the target than a hyperparameter γ. The model is trained to
answer yi , other answers y1 ..yi−1 , yi+1 ..yn , and maximize the marginal likelihood of any span in the
the top passages as used by S PAN S EQ G EN. We gold answer set Ȳ1 ..Ȳn . As with S PAN S EQ G EN,
pretrain on NQ- OPEN to generate questions given we pretrain on NQ- OPEN and finetune on A M -
an answer and passage, and then finetune it on the BIG NQ. We then produce disambiguated questions
full task data in A MBIG NQ. We include ablations using our BART-based QD model (Section 5).
on different variants of the model in Section 6.2.
6.2 Results
Co-training with weak supervision. Given the Table 3 reports the performance of our baselines;
prevalence of unlabelled ambiguity in NQ- OPEN, example model outputs are provided in Table 5.
we introduce a method that treats the NQ- OPEN
annotations as weak supervision and learns to dis- Main results. We first find that D ISAMBIG -
cover potential ambiguity in the data. We modify a FIRST is significantly worse than other models. In
democratic co-training algorithm (Zhou and Gold- particular, classification accuracy on whether the
man, 2004) as described in Algorithm 1. We itera- prompt question is ambiguous is 67%, close to the
tively grow the training set D̂full from A MBIG NQ majority baseline (60%). When the model does
(Dfull ) with silver data from NQ- OPEN (Dpartial ) identify an ambiguous question, its rewrites often
predicted by a majority of a set C of S PAN S EQ G EN look reasonable on the surface, but do not match
models trained on D̂full . The key step is injecting the facts. For instance, in example 1 of Table 5, it
the known answer y j from NQ- OPEN as a prefix asks about filming in 2017 and during season 1 for
to S PAN S EQ G EN’s output during prediction. In Snow White and the Huntsman, which was actually
each step, if a majority of C predict an additional a film released in 2012. This shows that reading
answer, we assume we have found a false negative evidence documents is crucial for identifying and
and add the result to the training set D̂full . If all characterizing ambiguities.
models predict no additional answer, we add the While S PAN S EQ G EN outperforms Karpukhin
example to D̂full with y j as a single answer. et al. (2020) with thresholding, the difference is
F1ans (all) F1ans (multi) F1BLEU F1EDIT-F1
Model
dev test dev test dev test dev test
D ISAMBIG - FIRST 28.1 24.8 21.9 18.8 4.2 4.0 2.7 2.2
Thresholding + QD 37.1 32.3 28.4 24.8 13.4 11.3 6.6 5.5
S PAN S EQ G EN + QD 39.7 33.5 29.3 24.5 13.4 11.4 7.2 5.8
S PAN S EQ G EN† + QD 41.2 35.2 29.8 24.5 13.6 10.6 7.4 5.7
S PAN S EQ G EN† (Co-training) + QD 42.3 35.9 31.7 26.0 14.3 11.5 8.0 6.3
Table 3: Results on A MBIG NQ. The multi measure only considers examples with multiple question-answer pairs.
†
indicates ensemble. See Appendix B for details on the discrepancy between development and test.
Table 4: Ablations on question disambiguation (development data, multiple answers only). QD model refers to the
question disambiguation model described in Section 5. For multiple answer prediction, we use S PAN S EQ G EN†
with co-training (Full task) or the gold answers (Gold answers given).
not as great as we expected. This suggests two in NQ- OPEN; detailed discussion is in Appendix B.
things. First, thresholding may be a surprisingly
effective baseline for outputting multiple answers, Effect of co-training. The last two rows of Ta-
even though the answers must compete with each ble 3 reports the effect of our co-training method.
other for probability mass in order to surpass the As co-training requires multiple trained models,
threshold γ. Second, maximizing likelihood in a we compare with a naive ensemble. While we see
sequence-to-sequence model like S PAN S EQ G EN gains from ensembling alone, an ensemble trained
may not produce well-calibrated results. For in- with the co-training method achieves the best per-
stance, the model seems to suffer due to variation formance on all metrics. This result demonstrates
in the length of the output sequence, outputting the potential of jointly using A MBIG NQ and partial
shorter sequences on average (3.0 tokens) than gold supervision from NQ- OPEN.
(6.7).5 This leads to low recall when there are mul- Ablations on question disambiguation. Ta-
tiple answers; our best model achieves a precision ble 4 reports results of an ablation experiment on
of 49.6 and recall of 25.3 for its F1ans of 31.7 on question disambiguation (QD). Among our abla-
such questions. tions, we include models without the prompt ques-
Overall, S PAN S EQ G EN achieves reasonable tion or untargeted answers as input, and a naive
F1ans scores. F1ans on examples with multiple baseline that always outputs the prompt question.
question-answer pairs (multi) are lower, indicat- We report the metrics both in the scenarios of the
ing that predicting all plausible answers is more full task and the gold answers given, to see the
challenging than predicting a single answer, as ex- performance dependent on and independent from
pected. S PAN S EQ G EN also obtains the best per- multiple answer prediction, respectively.6
formance in F1BLEU and F1EDIT-F1 , although their Simply copying the prompt question gives high
absolute values are low in general; we discuss this F1BLEU , which is natural since the questions were
in our question disambiguation ablations below. disambiguated using minimal edits. This justifies
There is a substantial difference in performance using F1EDIT-F1 to evaluate semantic differences
between development and test overall, likely due to from the prompt question. In addition, we find that
distributional differences in the original questions 6
Note that a high F1ans and low F1EDIT-F1 may not indi-
cate bad question disambiguation. For instance, if a model
5
This problem has also been reported in other conditional correctly predicts one out of two answers and does not per-
generation tasks (Sountsov and Sarawagi, 2016; Stahlberg and form any edits to the question, it obtains high F1ans and zero
Byrne, 2019); we leave it for future work. F1EDIT-F1 , despite the error being in answer prediction.
Prompt question #1: Where was snow white and the huntsman filmed?
Reference:
Q: Where were beach scenes for snow white and huntsman predominantly filmed? / A: Marloes Sands Beach
Q: Where was principal photography for snow white and huntsman filmed? / A: United Kingdom
Q: Where was castle in snow white and huntsman filmed? / A: Gateholm island
Prediction of D ISAMBIG - FIRST: (F1ans =0.40, F1EDIT-F1 =0.00)
Q: Where was snow white and the huntsman filmed in 2017? / A: Marloes Sands Beach
Q: Where was snow white and the huntsman filmed during the filming of Season 1 of the TV series? / A: Marloes Sands Beach
Prediction of S PAN S EQ G EN: (F1ans =0.80, F1EDIT-F1 =0.69)
Q: Where was snow white and huntsman principal photography filmed / A: United Kingdom
Q: Where were beach scenes for snow white and huntsman mostly filmed / A: Marloes Sands Beach
Prompt question #2: When was the city of new york founded?
Reference:
Q: When was city of new york founded by dutch and initially called new amsterdam? / A: 1624
Q: When was city of new york under english control and renamed to new york? / A: 1664
Prediction of S PAN S EQ G EN: (F1ans =1.00, F1EDIT-F1 =0.67)
Q: When was city of new york city founded with dutch protection? / A: 1624
Q: When was city of new york city founded and renamed with english name? / A: 1664
Table 5: Model predictions on samples from the development data. (#1) D ISAMBIG - FIRST generates questions
that look reasonable on the surface but don’t match the facts. S PAN S EQ G EN produces the reasonable answers and
questions, although not perfect. (#2) S PAN S EQ G EN produces correct answers and questions.
Performance is low overall, even given the gold Since A MBIG NQ provides an evaluation set with
answers, highlighting the challenge of the task. We explicit sets of multiple answers, we can also test
think there are two major reasons. First, maxi- if models trained on partial supervision only (NQ-
mizing the likelihood of the output sequence can OPEN ) are capable of producing full answer sets.
miss the importance of edits to the prompt question, In fact, the problem of ambiguity already exists
leading the QD model to miss the information that in previous QA tasks, and a single labeled answer
is most important to differentiate one answer from can be viewed as a sample from a multi-modal dis-
the others. Second, there is a lack of annotated tribution of answers. This setting is important for
data, especially for question disambiguation which modeling in domains where single-answer datasets
does not benefit from weakly supervised learning are available but full annotations like in A MBIG NQ
with NQ- OPEN; future work can explore how to are not. To this end, we present a zero-shot setting
maximize the use of supervision from other avail- where a system predicts multiple distinct answers
able data. It is also worth noting that the metric without using A MBIG NQ training data. We in-
may miss edits that are semantically correct, but clude four NQ- OPEN models including ours, con-
phrased differently (see Table 5, example 2). sisting of diverse approaches and model architec-
tures, as baselines. These models, when trained line model for producing multiple answers to open-
on NQ- OPEN, may be made to predict multiple an- domain questions, with experiments showing its
swers via thresholding as described in Section 6.1.7 effectiveness in learning from our data while high-
Table 7 reports zero-shot performance. Although lighting possible areas for improvement.
S PAN S EQ G EN outperforms Karpukhin et al. (2020) Future research developing on AmbigQA mod-
in the standard setting, it is worse in zero-shot F1ans els may include explicitly modeling ambiguity over
(multi), potentially because thresholding exacer- events and entities or in the retrieval step, as well as
bates the problems that S PAN S EQ G EN has with improving performance on the difficult problems
long sequences (Section 6.2). of answer recall and question disambiguation. Fur-
thermore, future work may build on the AmbigQA
6.4 Error Analysis task with more open-ended approaches such as (1)
Table 6 reports an analysis of predictions by applying the approach to QA over structured data
S PAN S EQ G EN with co-training, based on 50 ran- (such as ambiguous questions that require return-
dom samples from the development data; examples ing tables), (2) handling questions with no answer
can be found in the Appendix (Table 10). When or ill-formed questions that require inferring and
there are multiple reference answers, the model satisfying more complex ambiguous information
rarely gets all correct answers, although often gen- needs, and (3) more carefully evaluating usefulness
erates a subset of them. In 15 out of 20 partially to end users.
correct cases, the model produces only one answer,
Acknowledgments
consistent with the under-generation we found in
Section 6.2. In four out of those 15 cases, the model This research was supported by ONR N00014-18-
prediction is arguably the most likely answer,8 but 1-2826, DARPA N66001-19-2-403, the NSF (IIS-
in the other 11 cases, it hard to argue for one an- 1252835, IIS-1562364), an Allen Distinguished
swer over the other(s). It is also worth noting that Investigator Award, and the Sloan Fellowship. We
accuracy on examples with a single answer is quite thank Mandar Joshi, H2Lab members and the
high, being correct in 13 out of 20 cases. This anonymous reviewers for their helpful comments
estimated accuracy on unambiguous questions is and suggestions.
higher than state-of-the-art levels on NQ- OPEN (42
EM), suggesting that NQ- OPEN may substantially
underestimate performance due to the prevalence References
of unmarked ambiguity. Together with our experi- Mohammad Aliannejadi, Hamed Zamani, Fabio
mental results, this seems to indicate that recall of Crestani, and W Bruce Croft. 2019. Asking clarify-
ing questions in open-domain information-seeking
multiple answers is one of the primary challenges conversations. In SIGIR.
in AmbigQA.
Akari Asai, Kazuma Hashimoto, Hannaneh Hajishirzi,
7 Conclusion & Future Work Richard Socher, and Caiming Xiong. 2020. Learn-
ing to retrieve reasoning paths over wikipedia graph
for question answering. In ICLR.
We introduced A MBIG QA, a new task that involves
providing multiple possible answers to a potentially Jonathan Berant, Andrew Chou, Roy Frostig, and Percy
ambiguous open-domain question, and providing Liang. 2013. Semantic parsing on Freebase from
question-answer pairs. In EMNLP.
a disambiguated question corresponding to each
answer. We constructed A MBIG NQ, a dataset with Pavel Braslavski, Denis Savenkov, Eugene Agichtein,
14,042 annotations on NQ- OPEN questions. Our and Alina Dubatovka. 2017. What do you mean ex-
analysis shows the dataset contains diverse types actly? Analyzing clarification questions in CQA. In
Proceedings of the 2017 Conference on Conference
of ambiguity, often not visible from the prompt Human Information Interaction and Retrieval.
question alone. We also introduced a first base-
Danqi Chen, Adam Fisch, Jason Weston, and Antoine
7
We allow using development data to tune the threshold γ, Bordes. 2017. Reading Wikipedia to answer open-
although this arguably makes our setting not zero-shot in the domain questions. In ACL.
strictest sense.
8
For example, a question “Who did <title-of-the- Christopher Clark, Kenton Lee, Ming-Wei Chang,
song>” is ambiguous, but a well-known performer of the Tom Kwiatkowski, Michael Collins, and Kristina
song may be argued to be a more likely answer than its little- Toutanova. 2019. BoolQ: Exploring the surprising
known songwriter. difficulty of natural yes/no questions. In NAACL.
Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Myle Ott, Sergey Edunov, Alexei Baevski, Angela
Kristina Toutanova. 2019. BERT: Pre-training of Fan, Sam Gross, Nathan Ng, David Grangier, and
deep bidirectional transformers for language under- Michael Auli. 2019. fairseq: A fast, extensi-
standing. In NAACL. ble toolkit for sequence modeling. arXiv preprint
arXiv:1904.01038.
Matthew Dunn, Levent Sagun, Mike Higgins, Ugur
Guney, Volkan Cirik, and Kyunghyun Cho. 2017. Kishore Papineni, Salim Roukos, Todd Ward, and Wei-
SearchQA: A new q&a dataset augmented with Jing Zhu. 2002. BLEU: a method for automatic eval-
context from a search engine. arXiv preprint uation of machine translation. In ACL.
arXiv:1704.05179.
Adam Paszke, Sam Gross, Soumith Chintala, Gregory
Ahmed Elgohary, Denis Peskov, and Jordan L. Boyd- Chanan, Edward Yang, Zachary DeVito, Zeming
Graber. 2019. Can you unpack that? learning to Lin, Alban Desmaison, Luca Antiga, and Adam
rewrite questions-in-context. In EMNLP. Lerer. 2017. Automatic differentiation in PyTorch.
Sudha Rao and Hal Daumé III. 2018. Learning to ask
Kelvin Guu, Kenton Lee, Zora Tung, Panupong Pa-
good questions: Ranking clarification questions us-
supat, and Ming-Wei Chang. 2020. REALM:
ing neural expected value of perfect information. In
Retrieval-augmented language model pre-training.
ACL.
In ICML.
Sudha Rao and Hal Daumé III. 2019. Answer-based
Mandar Joshi, Eunsol Choi, Daniel S Weld, and Luke adversarial training for generating clarification ques-
Zettlemoyer. 2017. TriviaQA: A large scale dis- tions. In NAACL.
tantly supervised challenge dataset for reading com-
prehension. In ACL. Pavel Sountsov and Sunita Sarawagi. 2016. Length
bias in encoder decoder models and a case for global
Vladimir Karpukhin, Barlas Oguz, Sewon Min, Ledell conditioning. In EMNLP.
Wu, Sergey Edunov, Danqi Chen, and Wen-tau Yih.
2020. Dense passage retrieval for open-domain Felix Stahlberg and Bill Byrne. 2019. On nmt search
question answering. In EMNLP. errors and model errors: Cat got your tongue? In
EMNLP.
Tom Kwiatkowski, Jennimaria Palomaki, Olivia Red-
field, Michael Collins, Ankur Parikh, Chris Al- Ellen M Voorhees et al. 1999. The TREC-8 question
berti, Danielle Epstein, Illia Polosukhin, Jacob De- answering track report. In Trec.
vlin, Kenton Lee, Kristina Toutanova, Llion Jones,
Matthew Kelcey, Ming-Wei Change, Andrew M. Thomas Wolf, Lysandre Debut, Victor Sanh, Julien
Dai, Jakob Uszkoreit, Quoc Le, and Slav Petrov. Chaumond, Clement Delangue, Anthony Moi, Pier-
2019. Natural Questions: a benchmark for question ric Cistac, Tim Rault, R’emi Louf, Morgan Funtow-
answering research. TACL. icz, and Jamie Brew. 2019. HuggingFace’s Trans-
formers: State-of-the-art natural language process-
Kenton Lee, Ming-Wei Chang, and Kristina Toutanova. ing. arXiv preprint arXiv:1910.03771.
2019. Latent retrieval for weakly supervised open Jingjing Xu, Yuechen Wang, Duyu Tang, Nan Duan,
domain question answering. In ACL. Pengcheng Yang, Qi Zeng, Ming Zhou, and SUN
Xu. 2019. Asking clarification questions in
Mike Lewis, Yinhan Liu, Naman Goyal, Mar- knowledge-based question answering. In EMNLP.
jan Ghazvininejad, Abdelrahman Mohamed, Omer
Levy, Ves Stoyanov, and Luke Zettlemoyer. Hamed Zamani, Gord Lueck, Everest Chen, Rodolfo
2020. BART: Denoising sequence-to-sequence pre- Quispe, Flint Luu, and Nick Craswell. 2020. Mim-
training for natural language generation, translation, ics: A large-scale data collection for search clarifica-
and comprehension. In ACL. tion. In ACM International on Conference on Infor-
mation and Knowledge Management.
Julian Michael, Gabriel Stanovsky, Luheng He, Ido
Dagan, and Luke Zettlemoyer. 2018. Crowdsourc- Y. Zhou and S. Goldman. 2004. Democratic co-
ing question-answer meaning representations. In learning. In IEEE International Conference on
NAACL. Tools with Artificial Intelligence.
(b) Interface in the generation stage when the workers click and read one of Wikipedia pages from the search results.
(c) Interface in the validation stage when the workers are given annotations from two generation workers and click the
Wikipedia page that the generation workers have read.
Table 9: Breakdown of cases that NQ- OPEN answer is not included in A MBIG NQ answers.
†
en.wikipedia.org/wiki/List_of_National_Inventors_Hall_of_Fame_inductees
‡
en.wikipedia.org/wiki/History_of_computing_hardware_(1960s%E2%80%93present)
Reference has multiple answers; Multiple answer prediction is correct (2%)
Prompt question: Who was england’s prime minister during ww1?
Reference: H. H. Asquith (beginning of WW1), David Lloyd George (end of WW1)
Prediction: (F1ans =1.00) H. H. Asquith, David Lloyd George
Reference has multiple answers; Multiple answer prediction is partially correct (40%)
Prompt question: Who played kelly on the drew carey show?
NQ- OPEN answer: Cynthia Watros
Reference: Cynthia Watros (as Kellie N.), Jenny McCarthy (as M. Kelly), Brett Butler (as G. Kelly), Anna Gunn (as Kelly W.)
Prediction: (F1ans =0.40): Brett Butler
Reference has multiple answers; Multiple answer prediction is incorrect (14%)
Prompt question: Who plays the white queen in alice through the looking glass?
Reference: Amelia Crouch (young White Queen), Anne Hathaway (adult White Queen)
Prediction: (F1ans =0.00): Helena Bonham Carter†
Reference has one answer; over-generated predictions (2%)
Prompt question: How many times csk reached final in ipl?
Reference: eight
Prediction: (F1ans =66.7): eight, seven‡
Reference has one answer; correct single answer prediction (26%)
Prompt question: When did the 5th circuit became the 11th circuit?
Reference: October 1, 1981
Prediction: (F1ans =100.0): October 1, 1981
Reference has one answer; incorrect single answer prediction (12%)
Prompt question: Who is considered the home team for super bowl 52?
Reference: New England Patriots
Prediction: (F1ans =0.0): Atlanta Falcons
Reference is incorrect (4%)
Prompt question: Who has won the most trophies man utd or liverpool?
Reference: Man utd (trophies), Liverpool (FIFA and UEFA Cups)
Prediction: (F1ans =66.7): Manchester United
Table 10: Analysis of multiple answer predictions made by S PAN S EQ G EN with co-training, on 50 samples from
the development data. Rewrites are omitted but differentiation of multiple answers is denoted as a keyword in
italic.
†
Helena Bonham Carter played Red Queen.
‡
In fact, the model may have found time-dependency, because the eighth event happened only in 2019.