Baleen: Robust Multi-Hop Reasoning at Scale via Condensed Retrieval
Abstract
Multi-hop reasoning (i.e., reasoning across two or more documents) is a key ingredient for NLP models that leverage large corpora to exhibit broad knowledge. To retrieve evidence passages, multi-hop models must contend with a fast-growing search space across the hops, represent complex queries that combine multiple information needs, and resolve ambiguity about the best order in which to hop between training passages. We tackle these problems via Baleen, a system that improves the accuracy of multi-hop retrieval while learning robustly from weak training signals in the many-hop setting. To tame the search space, we propose condensed retrieval, a pipeline that summarizes the retrieved passages after each hop into a single compact context. To model complex queries, we introduce a focused late interaction retriever that allows different parts of the same query representation to match disparate relevant passages. Lastly, to infer the hopping dependencies among unordered training passages, we devise latent hop ordering, a weak-supervision strategy in which the trained retriever itself selects the sequence of hops. We evaluate Baleen on retrieval for two-hop question answering and many-hop claim verification, establishing state-of-the-art performance.
1 Introduction
In open-domain reasoning, a model is tasked with retrieving evidence from a large corpus to answer questions, verify claims, or otherwise exhibit broad knowledge. In practice, such open-domain tasks often further require multi-hop reasoning, where the evidence must be extracted from two or more documents. To do this effectively, a model must learn to use partial evidence it retrieves to bridge its way to additional documents leading to an answer. In this vein, HotPotQA [31] contains complex questions answerable by retrieving two passages from Wikipedia, while HoVer [13] contains claims that can only be verified (or disproven) by combining facts from up to four Wikipedia passages.
Table 1 illustrates this using the claim “The MVP of [a] game Red Flaherty umpired was elected to the Baseball Hall of Fame” from the HoVer validation set. To verify this claim, a HoVer model must identify facts spread across three Wikipedia pages: Red Flaherty umpired in the World Series in 1955, 1958, 1965, and 1970. The MVP of the 1965 World Series was Sandy Koufax. Koufax was later elected to the Baseball Hall of Fame.
This three-hop claim illustrates three major challenges in multi-hop retrieval. First, multi-hop queries encompass multiple information needs; the claim above referenced facts from three disparate passages. Second, retrieval errors in each hop propagate to subsequent hops. This can happen if the model directly retrieves information about the Baseball Hall of Fame, confuses Red Flaherty with, say, Robert Flaherty, or singles out the MVP of, say, the 1958 World Series, which Flaherty also umpired. Third, due to the dependency between hops, retrievers must learn an effective sequence of hops, where previously-retrieved clues lead to other relevant passages. These inter-passage dependencies can be non-obvious for many-hop problems with three or more passages, and are often left unlabeled, as it can be expensive to annotate one (or every) sequence in which facts could be retrieved.
The MVP of [a] game Red Flaherty umpired was elected to the Baseball Hall of Fame. | |
---|---|
The MVP of [a] game Red Flaherty umpired was elected to the Baseball Hall of Fame. Red Flaherty: He umpired in World Series 1955, 1958, 1965, and 1970. | |
The MVP of [a] game Red Flaherty umpired was elected to the Baseball Hall of Fame. Red Flaherty: He umpired in World Series 1955, 1958, 1965, and 1970. 1965 World Series: It is remembered for MVP Sandy Koufax. | |
The MVP of [a] game Red Flaherty umpired was elected to the Baseball Hall of Fame. Red Flaherty: He umpired in World Series 1955, 1958, 1965, and 1970. 1965 World Series: It is remembered for MVP Sandy Koufax. Sandy Koufax: He was elected to the Baseball Hall of Fame. |
These challenges call for highly expressive query representations, robustness to retrieval errors, and scalability to many hops over massive document collections. Existing systems fall short on one or more of these criteria. For instance, many state-of-the-art systems rely on bag-of-words [23] or single-vector dot-product [29] retrievers, whose capacity to model an open-domain question is limited [17], let alone complex multi-hop queries. Furthermore, existing systems embed trade-offs when it comes to “hopping”: they employ brittle greedy search, which limits recall per hop; or they employ beam search over an exponential space, which reduces scalability to many hops; or they assume explicit links that connect every passage with related entities, which ties them to link-structured corpora. Lastly, to order the hops, many systems use fragile supervision heuristics (e.g., finding passages whose page titles appear in other passages) tailored for particular datasets like HotPotQA.
We tackle these problems with Baleen,111https://github.com/stanford-futuredata/Baleen
In marine biology, “baleen” refers to the filter-feeding system that baleen whales use to capture small organisms, filtering out seawater. So too does our system seek to capture relevant facts from a sea of documents.
a scalable multi-hop reasoning system that improves accuracy and robustness.
We introduce a condensed retrieval architecture, where the retrieved facts from each hop are summarized into a short context that becomes a part of the query for subsequent hops, if any (Table 1). Unlike beam search, condensed retrieval allows effective scaling to many hops, and we find that it complements greedy search (i.e., taking the best passage per hop) in improving recall considerably. We then tackle the complexity of queries by proposing a focused late interaction passage retriever (FLIPR), a robust learned search model that allow different parts of the same query representation to match disparate relevant passages. FLIPR inherits the scalability of the vanilla late interaction paradigm of ColBERT [16] but uniquely allows the same query to exhibit tailored matching patterns against each target passage. Lastly, we devise latent hop ordering, a weak-supervision strategy that uses the retriever itself to select effective hop paths.
We first test Baleen on the two-hop HotPotQA benchmark, finding evidence of saturation in retrieval: we achieve 96.3% answer recall in the top-20 retrieved passages, up from 89.4% for existing work. We then test Baleen’s ability to scale accurately to more hops, reporting our main results using the recent many-hop HoVer task. We build a strong many-hop baseline model that combines recent results from the open-domain question answering [17] and multi-hop [23] literatures. This baseline combines ColBERT retrieval and standard greedy search with an ELECTRA [5] re-ranker, and generalizes typical text-based heuristics to order the hops for training. After verifying its strong results on HotPotQA, we show that it outperforms the official TF-IDF + BERT baseline of HoVer by over 30 points in retrieval accuracy. Against this strong baseline itself, Baleen improves passage retrieval accuracy by another 17 points, raising it to over 90% at passages. Baleen also improves the evidence extraction F1 score dramatically, outperforming even the oracle retrieval + BERT results by Jiang et al. [13]. Our ablations (§5.4) show that Baleen’s FLIPR retriever, condenser architecture, and latent supervision are essential to its strong performance.
2 Background and related work
Open-domain reasoning
There is significant recent interest in NLP models that can solve tasks by retrieving evidence from a large corpus. The most popular such task is arguably open-domain question answering (OpenQA; Chen et al. [2]), which extends the well-studied machine reading comprehension (MRC) problem over supplied question–passage pairs to answering questions given only a large text corpus like Wikipedia. Other open-domain tasks span claim verification (e.g., FEVER; Thorne et al. [24]), question generation (e.g., with Jeopardy as in Lewis et al. [19]), and open dialogue (e.g., Wizard of Wikipedia; Dinan et al. [8]), among others. Many of these datasets are compiled in the recently-introduced KILT benchmark [21] for knowledge-intensive tasks. Among models for these tasks, the most relevant to our work are OpenQA models that include learned retrieval components. These include ORQA [18], REALM [10], and DPR [15]. Lewis et al. [19] introduced RAG, which by virtue of a seq2seq architecture can tackle OpenQA as well as other open-domain problems, including claim verification and question generation.
Multi-hop open-domain reasoning
Most of the open-domain tasks from §2 can be solved by finding one relevant passage in the corpus, often by design. In contrast, a number of recent works explore multi-hop reasoning over multiple passages. These include QAngaroo [27] and 2WikiMultiHopQA [11], among others. While thus “multi-hop”, these tasks supply the relevant passages for each example (possibly with small set of “distractor” candidates), and thus do not require retrieval from a large corpus. To our knowledge, HotPotQA was the first large-scale open-domain multi-hop task, particularly in its retrieval-oriented “fullwiki” setting. HotPotQA catalyzed much follow-up research. However, it is limited to only two-hop questions, many of which may not require strong multi-hop reasoning capabilities (see Chen and Durrett [3]; Wang et al. [26]), softening the retrieval challenges discussed in §1. Very recently, Jiang et al. [13] introduced the HoVer many-hop verification dataset, which contains two-, three-, and four-hop examples. HoVer contains just over 18,000 training examples, about 5 smaller than HotPotQA, adding to the challenge posed by HoVer.
Multi-hop open-domain models
To conduct the “hops”, many prior multi-hop systems (e.g., Asai et al. [1]; Zhao et al. [32]) assume explicit links that connect every passage with related entities. We argue that this risks tying systems to link-structured knowledge bases (like Wikipedia) or producing brittle architectures tailored for datasets constructed by following hyperlinks (like HotPotQA). Recently, Xiong et al. [29] and Qi et al. [23] introduce MDR and IRRR, state-of-the-art systems that assume no explicit link structure. Instead, they use an iterative retrieval paradigm—akin to that introduced by Das et al. [6] and Feldman and El-Yaniv [9]—that retrieves passages relevant to the question, reads these passages, and then formulates a new query for another hop if necessary. We adopt this iterative formulation and tackle three major challenges for multi-hop retrieval.
ColBERT: late interaction paradigm
Most learned-retrieval systems for OpenQA (§2) encode every query and passage into a single dense vector. Khattab and Zaharia [16] argue that such single-vector representations are not sufficiently expressive for retrieval in many scenarios and introduce late interaction, a paradigm that represents every query and every document at a finer granularity: it uses a vector for each constituent token. Within this paradigm, they propose ColBERT, a state-of-the-art retriever wherein a BERT encoder embeds the query into a matrix of vectors (given tokens) and encodes every passage in the corpus as matrix of vectors (for tokens per passage). The passage representations are query-independent and thus are computed offline and indexed for fast retrieval.
During retrieval with query , ColBERT assigns a score to a passage by finding the maximum-similarity (MaxSim) score between each vector in ’s representation and all the vectors of and then summing these MaxSim scores. This MaxSim-decomposed interaction enables scaling to massive collections with millions of passages, as the token-level passage vectors can all be indexed for fast nearest-neighbor search. Khattab and Zaharia [16] find this fine-grained matching to outperform single-vector representations, a finding extended to open-domain question answering by Khattab et al. [17]. In this work, we propose FLIPR (§3.1), a retriever with an improved late interaction mechanism for complex multi-hop queries. Whereas existing retrievers, whether traditional (e.g., TF-IDF) or neural [18; 15; 16], seek passages that match all of the query, multi-hop queries can be long and noisy and need to match disparate contexts. FLIPR handles this explicitly with focused late interaction.
Relevance-guided supervision
Khattab et al. [17] propose an iterative strategy for weakly-supervising retrievers called relevance-guided supervision (RGS). RGS assumes that no labeled passages are supplied for training the retriever. Instead, every training question is associated with a short answer string whose presence in a passage is taken as a weak signal for relevance. RGS starts with an off-the-shelf retriever and uses it to collect the top- passages for each training question, dividing these passages into positive and negative examples based on inclusion of the short answer string. These examples are used to train a stronger retriever, which is then used to repeat this process one or two times. We draw inspiration from RGS when designing latent hop ordering: unlike RGS, we do have gold-labeled passages for training but we crucially have multiple hops, whose order is not given. Lastly, our supervision for the first hop shares inspiration with GoldEn [22]. GoldEn uses a bag-of-words model—and thus effectively term overlap—to identify the “more easily retrieved” passage among just two for the two-hop task HotPotQA.

3 Baleen
We now introduce Baleen, which uses an iterative retrieval paradigm to find relevant facts in successive hops. On HotPotQA (§4) and HoVer (§5), we use Baleen with and , respectively. As illustrated in Table 1, the input to Baleen is a textual query like a question to answer or a claim to verify. The goal of hop is to take in query and output an updated query containing the initial input query and pertinent facts extracted from the hops.
The Baleen architecture is depicted in Figure 1. In every hop, FLIPR (§3.1) uses to retrieve passages from the full corpus. These passages are fed into a two-stage condenser (§3.3), which reads these passages and extracts the most relevant facts, which we model as individual sentences. The facts are collected into a single sequence and added to for the subsequent hop, if any. Once all hops are complete, Baleen’s task-specific reader processes , which now contains the query and all condensed facts, to solve the downstream task. If desired, the top- passages from each hop can also be collected and fed to a downstream model after retrieval is complete. Baleen’s retriever, condenser, and reader are implemented as Transformer encoders [25] and are trained individually. We use BERT-base [7] for FLIPR and, like Qi et al. [23] and Xiong et al. [29], use ELECTRA-large [5] for the other components. For supervision, we introduce the latent hop ordering scheme (§3.2).
3.1 FLIPR: focused late interaction
The FLIPR encoders and interaction mechanism are shown in Figure 2. Our query encoder reads and outputs a vector representation of every token in the input. Each query embedding interacts with all passage embeddings via a MaxSim operator, permitting us to inherit the efficiency and scalability of ColBERT [16]. While vanilla ColBERT sums all the MaxSim outputs indiscriminately, FLIPR considers only the strongest-matching query embeddings for evaluating each passage: it sums only the top- partial scores from scores, with .
We refer to this top- filter as a “focused” interaction. Intuitively, it allows the same query to match multiple relevant passages that are contextually unrelated, by aligning—during training and inference—a different subset of the query embeddings with different relevant passages. These important subsets of the query are not always clear in advance. For instance, in Table 1 should find the page about the 1965 World Series. This requires ignoring many distractions (e.g., World Series 1970 and Baseball Hall of Fame) that seem important in isolation while focusing on other phrases, namely a prominent “MVP” and the World Series 1965, a distinction evident only after finding the document and attempting to match it against the query.
FLIPR applies this mechanism over the representation of the query and the facts separately, keeping the top- and top- from and embeddings, respectively. More formally, the FLIPR score of a passage given a query is the summation over the scores corresponding to the original query and to the facts. Define , that is, the maximum similarity of the query embedding against the passage embeddings. The partial score is computed as
(1) |
where is an operator that keeps the largest values in a set. We define similarly, using the fact matches instead of .
We confirm the gains of FLIPR against the state-of-the-art ColBERT retriever in §5.4. By decomposing retrieval into token-level maximum similarities and pre-computing document representations, FLIPR maintains the scalability of the late interaction paradigm of ColBERT, reducing the candidate generation stage of retrieval into highly-efficient approximate -nearest neighbor searches. See Appendix §B.1 for the implementation details of FLIPR.

3.2 Supervision: latent hop ordering
For every training example, our datasets supply unordered gold passages. However, dependencies often exist in retrieval as Table 1 illustrates: a system needs to retrieve and read the page of Red Flaherty to deduce that it needs information about 1965 World Series. Such dependencies can be complex for 3- and 4-hop examples, as in HoVer, especially as multiple passages may be appropriate to retrieve together in the same hop.
We propose a generic mechanism for identifying the best gold passage(s) to learn to retrieve next. The key insight is that among the gold passages, a weakly-supervised retriever with strong inductive biases would reliably “prefer” those passages it can more naturally retrieve, given the hops so far. Thus, given , we train the retriever with every remaining gold passage and, among those, label as positives (for ) only those ranked highly by the model. We subsequently use these weakly-labeled positives to train another retriever for all the hops.
Latent hop ordering is summarized in Algorithm 1, which assumes we have already trained a single-hop (i.e., first-hop) retriever in the manner of relevance-guided supervision (see §2). To start, we use to retrieve the top- (e.g., ) passages for each training question (Line 2). We then divide these passages into positives and negatives (Line 3). Positives are the highly-ranked gold passages (i.e., the intersection of gold passages and the top- passages with, e.g., ) and negatives are the non-gold passages. Subsequently, we expand first-hop queries with the oracle facts from to obtain the queries for the second hop (Line 4).
We train a second-hop retriever using the aforementioned queries and negatives (Line 6). As we do not have second-hop positives, as a form of weak supervision, we train the retriever with all gold passages (per query) besides those already in . Once trained, we use to discover the second-hop positives , to collect negatives , and to expand the queries (Lines 7 and 8). We repeat this procedure for the third hop onward. Once this iterative procedure is complete, we have bootstrapped positives (and negatives) corresponding to every retrieval hop for each query. We combine these sets to train a single multi-hop retriever (Lines 10 and 11), which takes a query and learns to discriminate the positives in from the negatives in for every . In §5.4, we indeed find that this generic latent procedure outperforms a typical hand-crafted heuristic that relies on trying to match titles against the text.
3.3 Condenser: per-hop fact extraction
After each hop, the condenser proceeds in two stages. In the first stage, it reads and each of the top- passages retrieved by FLIPR. The input passages are divided into their constituent sentences and every sentence is prepended with a special token. The output embedding corresponding to these sentence markers are scored via a linear layer, assigning a score to each sentence in every retrieved passage. We train the first-stage condenser with a cross-entropy loss over the sentences of two passages, a positive and a negative.
Across the passages, the top- facts are identified and concatenated into the second-stage condenser model. As above, we include special tokens between the sentences, and the second-stage condenser assigns a score to every fact while attending jointly over all of the top- facts, allowing a direct comparison within the encoder. Every fact whose score is positive is selected to proceed, and the other facts are discarded. We train the second-hop condenser over a set of 7–9 facts, some positive and others (sampled) negative, using a linear combination of cross-entropy loss for each positive fact (against all negatives) and binary cross-entropy loss for each individual fact.
In practice, this condensed architecture offers novel tradeoffs against a re-ranking pipeline. On one hand, condensed retrieval scales better to more hops, as it represents the facts from long passages using only a few sentences. This may provide stronger interpretability, as the inputs to the retriever (e.g., the example in Table 1) become more focused. On the other hand, re-ranking requires only passage-level labels, which can be cheaper to collect, and provides the retriever with broader context, which can help resolve ambiguous references. Empirically, we find the two approaches to be complementary. In §5.4, we show that condensing is competitive with re-ranking (despite much shorter context; Appendix §C) and that a single retriever that combines both pipelines together substantially outperforms a standard rerank-only pipeline.
3.4 Reader: task-specific processing
After all hops are conducted, a reader is used to extract answers or verify the claim. We use a standard Transformer encoder, in particular ELECTRA-large [5], which we feed the final output of our multi-hop retrieval and condensing pipeline. We train for question answering similar to Khattab et al. [17] or claim verification similar to Jiang et al. [13].
4 Diagnosing retrieval saturation on HotPotQA
Model | Retriever | Architect. | Supervis. | P-EM | P-R@20 | Ans-R@20 |
---|---|---|---|---|---|---|
MDR | DPR | Beam | Heuristic | 81.2 | 82.9 | 89.4 |
IRRR |
ELECTRA
/BM25 |
Rerank | Heuristic | 84.1 | - | - |
ColBERT-Hop (ablation) | ColBERT | Rerank | Heuristic | 85.2 | 90.3 | 94.7 |
Baleen | FLIPR | Condense | Latent | 86.7 | 93.3 | 96.3 |
Before beginning our primary evaluation on the many-hop HoVer benchmark, we first investigate the capacity of HotPotQA in reflecting the challenges on multi-hop retrieval (§1). We use HotPotQA’s “fullwiki” setting, where the input to the model is a question whose answer requires retrieving two passages from Wikipedia. Similar to related work [29], we report passage-level exact-match (EM) and Recall@ (R@) on the development set. These are the percentage of questions for which the model correctly identifies the two gold passages and for which the model’s top- retrieved passages contain both gold passages, respectively.222For MDR, we use the authors released retrieval to compute P-R@20. The authors originally report 80.2, but we do not penalize the MDR retriever for finding both gold passages split across different pairs. We also report Answer-Recall@, the percentage of questions whose short answer string is found in the top- passages (excluding yes/no questions). Here, the top-20 passages from Baleen are simply the union of the top-10 passages retrieved during the first hop and the top-10 passages from the second hop, without duplicates.
We compare against the state-of-the-art models MDR and IRRR. We additionally compare with ColBERT-Hop, a strong baseline that ablates Baleen’s retriever, condenser, and supervision: ColBERT-Hop uses the state-of-the-art neural retriever ColBERT (see §2), employs a typical re-ranker architecture, and uses heuristic supervision. The results are shown in Table 2. Baleen demonstrates strong results, finding both gold passages in over 93% of all questions and finding passages that contain the short answer string for over 96% of all questions in just the top =20 passages. We argue that this effectively reduces the open-retrieval task to a narrow distractor-setting one, in which reader models are supplied with only a few passages (ones that are (nearly) guaranteed to contain the answer) and are then expected to extract the answer from just those passages.
In other words, such high recall largely shifts the competition in downstream quality to the reader, since it almost always has access to the answer in the top- set and is thus responsible for the majority of downstream failures. While reading the retrieved passages may still pose additional, important challenges beyond retrieval that remain open, it should be just one of the many challenging aspects of multi-hop reasoning at scale. This motivates our focus on the many-hop HoVer task.
5 Main evaluation on HoVer
We now conduct our primary evaluation, applying Baleen on the many-hop claim verification dataset HoVer, following the evaluation protocol of Jiang et al. [13]. First, we study the retrieval performance of Baleen in §5.2. Subsequently (§5.3), we investigate its sentence extraction and claim verification performance, which constitutes a test of overall end-to-end system effectiveness. We compare a four-hop Baleen system against the official baseline and the strong ColBERT-Hop baseline from §4.333At the time of writing, MDR’s open-source implementation only supports two-hop retrieval and the IRRR implementation is not publicly available. Moreover, both Xiong et al. [29] and Qi et al. [23] explore training only with two gold passages, though IRRR includes evaluation on a yet-unreleased 3-hop test set. See Appendix §B for implementation details and hyperparameters.
5.1 Task description
The input to the model is a claim, which is often one or a few sentences long. The model is to attempt to verify this claim, outputting “Supported” if the corpus contains enough evidence to verify the claim, or otherwise “Unsupported” if the corpus contradicts—or otherwise fails to support—this claim. Alongside this binary label, the model must extract facts (i.e., sentences) that justify its prediction. In particular, HoVer contains two-, three-, and four-hop claims that require facts from 2–4 different Wikipedia pages, and the model must find and extract one or more supporting facts from each of these pages for every claim. Models are not given the number of hops per claim.
5.2 Evaluating retrieval effectiveness
Retrieval@100 | Passage EM | Passage F1 | ||||||||||
Model / # Hops | All | 2 | 3 | 4 | All | 2 | 3 | 4 | All | 2 | 3 | 4 |
TF-IDF + BERT⋆ | 44.6 | 80.0 | 39.2 | 15.6 | 12.5 | 34.0 | 5.8 | 1.0 | 60.2 | 69.9 | 58.2 | 53.4 |
ColBERT-Hop | 74.8 | 95.8 | 77.9 | 47.6 | - | - | - | - | - | - | - | - |
Baleen | 92.2 | 97.7 | 93.1 | 85.1 | 63.6 | 75.8 | 62.5 | 52.6 | 89.2 | 90.2 | 89.9 | 86.8 |
Oracle + BERT⋆ | - | - | - | - | 34.0 | 50.9 | 28.1 | 26.2 | 80.6 | 81.7 | 79.1 | 82.2 |
Human⋆ | - | - | - | - | 77.0 | 85.0 | 82.4 | 65.8 | 93.5 | 92.5 | 95.3 | 91.4 |
Table 3 reports the retrieval and passage extraction quality of the systems. We report three metrics, each sub-divided by the number of hops of each claim. The first is Retrieval@100, which is the percentage of claims for which the system retrieves all of the relevant passages within the top-100 results. To compare with Jiang et al. [13], we report the Retrieval@100 results only for supported claims. All other metrics are reported for all claims. The second metric is Passage EM, which is the percentage of claims for which the system can provide the exact set of relevant passages. The third metric is Passage F1, which uses the standard F1 measure to offer a relaxed version of EM.
At the top of the table, the TF-IDF baseline retrieves top-100 results in one round of retrieval. At the bottom of the table, we include the results reported by Jiang et al. [13] for their BERT ranking model when supplied with oracle retrieval, and also their reported human performance. We show Baleen’s performance after four hops, where Baleen’s FLIPR retrieves 25 passages in each hop. Note that we exclude passages retrieved in hop from the retrieval in further hops . As the table shows, Baleen achieves strong results, outperforming the official baseline model by 47.6 points in Retrieval@100, 51.1 points in Passage EM, and 29.0 points in Passage F1. In fact, Baleen’s performance also consistently exceeds “Oracle + BERT” at passage extraction.
5.3 Evaluating end-to-end effectiveness
Sentence EM | Sentence F1 | Verification Accuracy | |||||||
Model / # of Hops | All | 2 | 3 | 4 | All | 2 | 3 | 4 | |
TF-IDF + BERT⋆ | 4.8/4.5 | 13.6 | 1.9 | 0.2 | 50.6/49.5 | 57.2 | 49.8 | 45.0 | 73.7 |
Baleen 1-hop | 19.7 | 40.9 | 15.4 | 4.3 | 72.3 | 77.5 | 72.4 | 66.4 | - |
Baleen 2-hop | 37.0 | 46.9 | 35.7 | 28.4 | 80.8 | 81.2 | 81.8 | 78.7 | - |
Baleen 3-hop | 38.9 | 47.1 | 37.0 | 33.2 | 81.4 | 81.2 | 82.3 | 80.0 | - |
Baleen 4-hop | 39.2/39.8 | 47.3 | 37.7 | 33.3 | 81.5/80.4 | 81.2 | 82.5 | 80.0 | 84.5/84.9 |
Oracle + BERT⋆ | 19.9 | 25.0 | 18.4 | 17.1 | 71.9 | 68.3 | 71.5 | 76.4 | 81.2 |
Human⋆ | 56.0 | 75.0 | 73.5 | 42.1 | 88.7 | 86.5 | 93.1 | 87.3 | 88.0 |
Next, we evaluate the performance of fact/sentence extraction and thereby the end-to-end claim verification of Baleen. The results are in Table 4, where Sentence EM and Sentence F1 are defined similar to the corresponding metrics in Table 3 but focus on sentence-level extraction. Moreover, Verification Accuracy is the percentage of claims that are labelled correctly as “Supported” or “Unsupported”. In Table 4, we see that Baleen’s sentence-level results mirror its document-level results, with large performance gains across the board against the baseline model. Baleen also outperforms the oracle-retrieval BERT model of Jiang et al. [13], emphasizing the strong retrieval and condensing performance of Baleen. In the leaderboard test-set evaluation (scores expected to be posted soon), Baleen increases the overall “HoVer Score” metric from 15.3 for the baseline to 57.5.
5.4 Ablation studies
Lastly, Table 5 reports various settings and ablations of Baleen, corresponding to its contributions. Here, models A and F are simply the four-hop Baleen and ColBERT-Hop in the previous experiments, respectively. Models B and C test the effect of different architectural decisions on Baleen. In particular, model B replaces our condensed retrieval architecture with a simpler re-ranking architecture (Appendix §B.2), which reflects a common design choice for multi-hop systems. In this case, FLIPR is given the top-ranked passage as context in each hop, helping expose the value of extracting facts in condensed retrieval, which exhibits much shorter contexts as we show in Appendix §C. Second, model C trains a FLIPR retriever that uses both a condenser pipeline and a reranking pipeline independently, and then retains the overall top-100 unique passages retrieved (Appendix §B.2). The results suggest that condensing offers complimentary gains, and the hybrid model attains the highest scores.
Moreover, models D and E test modifications to the retriever, while using the condenser of model A. Specifically, model D replaces our FLIPR retriever with a simpler ColBERT retriever, while retaining all other components of our architecture. This allows us to contrast our proposed focused late interaction with the “vanilla” late interaction of ColBERT. Further, model E replaces our weak-supervision strategy for inferring effective hop orders with a hand-crafted rule (Appendix §B.3) that deliberately exploits a construction bias in HotPotQA and HoVer, namely that passage titles can offer a strong signal for hop ordering.
Model | All | 2 | 3 | 4 | |||
Architecture Ablations | |||||||
A | Baleen (main architecture) | 92.2 | 97.7 | 93.1 | 85.1 | ||
B | Baleen | 91.2 | 98.7 | 93.0 | 80.0 | ||
C | Baleen | 94.5 | 99.2 | 94.4 | 90.0 | ||
|
|||||||
D | w/o FLIPR (uses ColBERT retrieval modeling) | 87.4 | 97.3 | 89.5 | 73.4 | ||
E | w/o LHO (uses a recursive title/passage-overlap heuristic for ordering) | 84.8 | 97.1 | 88.3 | 65.6 | ||
|
|||||||
F |
|
74.8 | 95.8 | 77.9 | 47.6 |
6 Discussion
Research limitations
This work investigates multi-hop reasoning at a large scale by focusing on the challenges presented by retrieval in this setting. While reading the retrieved (or supplied) passages to solve a downstream task is another important aspect, one that has received more attention in the literature, we do not attempt to advance the state of the art approaches for multi-hop reader models. The majority of multi-hop retrieval work to date has focused on two-hop retrieval. We leverage the recently-released HoVer dataset to scale our investigation to four retrieval hops, and Baleen allows for an arbitrary number of hops in principle, but more evaluation is needed before we can claim generality to an arbitrary number of hops. We propose a condenser architecture that summarizes contexts extractively. This architectural decision allows us to scale easily to many hops as it condenses the hop information into a short context, but doing so relies on the availability on sentence-level training information (as in HotPotQA and HoVer) and extracting sentences might in principle lose important context. We expect future work to be able to infer the sentence-level labels in a latent manner and we are excited about work like decontextualization [4] that tries to retain important context for standalone sentences. Lastly, considering that learning neural retrievers for large-scale many-hop tasks is a recently-emerging topic, we had to design our own strong baseline to strengthen our HoVer evaluation, after confirming that our baseline was very strong on HotPotQA.
Environmental and societal impact
By turning to retrieval from text corpora (see §2) as a mechanism to find and use knowledge, we seek to build scalable and efficient models that can reason and solve downstream tasks with concrete, checkable provenance from text sources and without growing the models’ number of trainable parameters dramatically. We use Wikipedia, which has favorable licensing (generally under CC BY-SA 3.0), and publicly-released datasets HoVer and HotPotQA (CC BY-SA 4.0 licenses). HotPotQA and HoVer are crowd-sourced, and we expect models trained on them to reflect the biases of the underlying data. Moreover, automated systems for answering questions or verifying claims can have misleading outputs or may even be abused. That said, we believe that focusing on retrieval and extractive models as a way to help users find information can offer net gains to society, especially in contrast with large generative models that may hallucinate or make statements ungrounded in a human-written corpus.
7 Conclusion
In this paper, we propose Baleen, a system for multi-hop reasoning that tackles the complexity of multi-hop queries with the focused late interaction mechanism for retrieval and mitigates the exponential search space problem by employing an aggressive condensed retrieval pipeline in every hop, which consolidates the pertinent information retrieved throughout the hops into a relatively short context while preserving high recall. Moreover, Baleen deals with the difficulty of ordering passages for many-hop supervision via latent hop ordering. By employing a strong retriever, incorporating a condenser, and avoiding brittle heuristics, Baleen can robustly learn from limited training signal. We evaluated Baleen’s retrieval on HotPotQA and the recent many-hop HoVer claim verification dataset and found that it greatly outperforms the baseline models.
Acknowledgments and Disclosure of Funding
We would like to thank Ashwin Paranjape, Megha Srivastava, and Ethan A. Chi for valuable discussions and feedback. This research was supported in part by affiliate members and other supporters of the Stanford DAWN project—Ant Financial, Facebook, Google, and VMware—as well as Cisco, SAP, a Stanford HAI Cloud Credit grant from AWS, and the NSF under CAREER grant CNS-1651570. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of the National Science Foundation.
References
- Asai et al. [2020] Akari Asai, Kazuma Hashimoto, Hannaneh Hajishirzi, Richard Socher, and Caiming Xiong. Learning to retrieve reasoning paths over wikipedia graph for question answering. In 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020. OpenReview.net, 2020. URL https://openreview.net/forum?id=SJgVHkrYDH.
- Chen et al. [2017] Danqi Chen, Adam Fisch, Jason Weston, and Antoine Bordes. Reading Wikipedia to answer open-domain questions. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1870–1879, Vancouver, Canada, 2017. Association for Computational Linguistics. doi: 10.18653/v1/P17-1171. URL https://aclanthology.org/P17-1171.
- Chen and Durrett [2019] Jifan Chen and Greg Durrett. Understanding dataset design choices for multi-hop reasoning. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4026–4032, Minneapolis, Minnesota, 2019. Association for Computational Linguistics. doi: 10.18653/v1/N19-1405. URL https://aclanthology.org/N19-1405.
- Choi et al. [2021] Eunsol Choi, Jennimaria Palomaki, Matthew Lamm, Tom Kwiatkowski, Dipanjan Das, and Michael Collins. Decontextualization: Making sentences stand-alone. Transactions of the Association for Computational Linguistics, 9:447–461, 2021.
- Clark et al. [2020] Kevin Clark, Minh-Thang Luong, Quoc V. Le, and Christopher D. Manning. ELECTRA: pre-training text encoders as discriminators rather than generators. In 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020. OpenReview.net, 2020. URL https://openreview.net/forum?id=r1xMH1BtvB.
- Das et al. [2019] Rajarshi Das, Shehzaad Dhuliawala, Manzil Zaheer, and Andrew McCallum. Multi-step retriever-reader interaction for scalable open-domain question answering. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019. OpenReview.net, 2019. URL https://openreview.net/forum?id=HkfPSh05K7.
- Devlin et al. [2019] Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4171–4186, Minneapolis, Minnesota, 2019. Association for Computational Linguistics. doi: 10.18653/v1/N19-1423. URL https://aclanthology.org/N19-1423.
- Dinan et al. [2019] Emily Dinan, Stephen Roller, Kurt Shuster, Angela Fan, Michael Auli, and Jason Weston. Wizard of wikipedia: Knowledge-powered conversational agents. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019. OpenReview.net, 2019. URL https://openreview.net/forum?id=r1l73iRqKm.
- Feldman and El-Yaniv [2019] Yair Feldman and Ran El-Yaniv. Multi-hop paragraph retrieval for open-domain question answering. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 2296–2309, Florence, Italy, 2019. Association for Computational Linguistics. doi: 10.18653/v1/P19-1222. URL https://aclanthology.org/P19-1222.
- Guu et al. [2020] Kelvin Guu, Kenton Lee, Zora Tung, Panupong Pasupat, and Ming-Wei Chang. Realm: Retrieval-augmented language model pre-training. arXiv preprint arXiv:2002.08909, 2020. URL https://arxiv.org/abs/2002.08909.
- Ho et al. [2020] Xanh Ho, Anh-Khoa Duong Nguyen, Saku Sugawara, and Akiko Aizawa. Constructing a multi-hop QA dataset for comprehensive evaluation of reasoning steps. In Proceedings of the 28th International Conference on Computational Linguistics, pages 6609–6625, Barcelona, Spain (Online), 2020. International Committee on Computational Linguistics. doi: 10.18653/v1/2020.coling-main.580. URL https://aclanthology.org/2020.coling-main.580.
- Izacard et al. [2020] Gautier Izacard, Fabio Petroni, Lucas Hosseini, Nicola De Cao, Sebastian Riedel, and Edouard Grave. A memory efficient baseline for open domain question answering. arXiv preprint arXiv:2012.15156, 2020. URL https://arxiv.org/abs/2012.15156.
- Jiang et al. [2020] Yichen Jiang, Shikha Bordia, Zheng Zhong, Charles Dognin, Maneesh Singh, and Mohit Bansal. HoVer: A dataset for many-hop fact extraction and claim verification. In Findings of the Association for Computational Linguistics: EMNLP 2020, pages 3441–3460, Online, 2020. Association for Computational Linguistics. doi: 10.18653/v1/2020.findings-emnlp.309. URL https://aclanthology.org/2020.findings-emnlp.309.
- Johnson et al. [2019] Jeff Johnson, Matthijs Douze, and Hervé Jégou. Billion-scale similarity search with gpus. IEEE Transactions on Big Data, 2019.
- Karpukhin et al. [2020] Vladimir Karpukhin, Barlas Oguz, Sewon Min, Patrick Lewis, Ledell Wu, Sergey Edunov, Danqi Chen, and Wen-tau Yih. Dense passage retrieval for open-domain question answering. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 6769–6781, Online, 2020. Association for Computational Linguistics. doi: 10.18653/v1/2020.emnlp-main.550. URL https://aclanthology.org/2020.emnlp-main.550.
- Khattab and Zaharia [2020] Omar Khattab and Matei Zaharia. ColBERT: Efficient and effective passage search via contextualized late interaction over BERT. In Jimmy Huang, Yi Chang, Xueqi Cheng, Jaap Kamps, Vanessa Murdock, Ji-Rong Wen, and Yiqun Liu, editors, Proceedings of the 43rd International ACM SIGIR conference on research and development in Information Retrieval, SIGIR 2020, Virtual Event, China, July 25-30, 2020, pages 39–48. ACM, 2020. doi: 10.1145/3397271.3401075. URL https://doi.org/10.1145/3397271.3401075.
- Khattab et al. [2021] Omar Khattab, Christopher Potts, and Matei Zaharia. Relevance-guided supervision for openqa with ColBERT. Transactions of the Association for Computational Linguistics, 9:929–944, 2021.
- Lee et al. [2019] Kenton Lee, Ming-Wei Chang, and Kristina Toutanova. Latent retrieval for weakly supervised open domain question answering. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 6086–6096, Florence, Italy, 2019. Association for Computational Linguistics. doi: 10.18653/v1/P19-1612. URL https://aclanthology.org/P19-1612.
- Lewis et al. [2020] Patrick S. H. Lewis, Ethan Perez, Aleksandra Piktus, Fabio Petroni, Vladimir Karpukhin, Naman Goyal, Heinrich Küttler, Mike Lewis, Wen-tau Yih, Tim Rocktäschel, Sebastian Riedel, and Douwe Kiela. Retrieval-augmented generation for knowledge-intensive NLP tasks. In Hugo Larochelle, Marc’Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan-Tien Lin, editors, Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual, 2020. URL https://proceedings.neurips.cc/paper/2020/hash/6b493230205f780e1bc26945df7481e5-Abstract.html.
- Nguyen et al. [2016] Tri Nguyen, Mir Rosenberg, Xia Song, Jianfeng Gao, Saurabh Tiwary, Rangan Majumder, and Li Deng. MS MARCO: A human-generated MAchine reading COmprehension dataset. arXiv preprint arXiv:1611.09268, 2016. URL https://arxiv.org/abs/1611.09268.
- Petroni et al. [2021] Fabio Petroni, Aleksandra Piktus, Angela Fan, Patrick Lewis, Majid Yazdani, Nicola De Cao, James Thorne, Yacine Jernite, Vladimir Karpukhin, Jean Maillard, Vassilis Plachouras, Tim Rocktäschel, and Sebastian Riedel. KILT: a benchmark for knowledge intensive language tasks. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 2523–2544, Online, 2021. Association for Computational Linguistics. doi: 10.18653/v1/2021.naacl-main.200. URL https://aclanthology.org/2021.naacl-main.200.
- Qi et al. [2019] Peng Qi, Xiaowen Lin, Leo Mehr, Zijian Wang, and Christopher D. Manning. Answering complex open-domain questions through iterative query generation. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 2590–2602, Hong Kong, China, 2019. Association for Computational Linguistics. doi: 10.18653/v1/D19-1261. URL https://aclanthology.org/D19-1261.
- Qi et al. [2020] Peng Qi, Haejun Lee, Oghenetegiri Sido, Christopher D Manning, et al. Retrieve, rerank, read, then iterate: Answering open-domain questions of arbitrary complexity from text. arXiv preprint arXiv:2010.12527, 2020. URL https://arxiv.org/abs/2010.12527.
- Thorne et al. [2018] James Thorne, Andreas Vlachos, Christos Christodoulopoulos, and Arpit Mittal. FEVER: a large-scale dataset for fact extraction and VERification. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), pages 809–819, New Orleans, Louisiana, 2018. Association for Computational Linguistics. doi: 10.18653/v1/N18-1074. URL https://aclanthology.org/N18-1074.
- Vaswani et al. [2017] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin. Attention is all you need. In Isabelle Guyon, Ulrike von Luxburg, Samy Bengio, Hanna M. Wallach, Rob Fergus, S. V. N. Vishwanathan, and Roman Garnett, editors, Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, December 4-9, 2017, Long Beach, CA, USA, pages 5998–6008, 2017. URL https://proceedings.neurips.cc/paper/2017/hash/3f5ee243547dee91fbd053c1c4a845aa-Abstract.html.
- Wang et al. [2019] Haoyu Wang, Mo Yu, Xiaoxiao Guo, Rajarshi Das, Wenhan Xiong, and Tian Gao. Do multi-hop readers dream of reasoning chains? In Proceedings of the 2nd Workshop on Machine Reading for Question Answering, pages 91–97, Hong Kong, China, 2019. Association for Computational Linguistics. doi: 10.18653/v1/D19-5813. URL https://aclanthology.org/D19-5813.
- Welbl et al. [2018] Johannes Welbl, Pontus Stenetorp, and Sebastian Riedel. Constructing datasets for multi-hop reading comprehension across documents. Transactions of the Association for Computational Linguistics, 6:287–302, 2018. doi: 10.1162/tacl_a_00021. URL https://aclanthology.org/Q18-1021.
- Wolf et al. [2020] Thomas Wolf, Lysandre Debut, Victor Sanh, Julien Chaumond, Clement Delangue, Anthony Moi, Pierric Cistac, Tim Rault, Remi Louf, Morgan Funtowicz, Joe Davison, Sam Shleifer, Patrick von Platen, Clara Ma, Yacine Jernite, Julien Plu, Canwen Xu, Teven Le Scao, Sylvain Gugger, Mariama Drame, Quentin Lhoest, and Alexander Rush. Transformers: State-of-the-art natural language processing. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pages 38–45, Online, 2020. Association for Computational Linguistics. doi: 10.18653/v1/2020.emnlp-demos.6. URL https://aclanthology.org/2020.emnlp-demos.6.
- Xiong et al. [2020] Wenhan Xiong, Xiang Lorraine Li, Srini Iyer, Jingfei Du, Patrick Lewis, William Yang Wang, Yashar Mehdad, Wen-tau Yih, Sebastian Riedel, Douwe Kiela, et al. Answering complex open-domain questions with multi-hop dense retrieval. arXiv preprint arXiv:2009.12756, 2020. URL https://arxiv.org/abs/2009.12756.
- Yamada et al. [2021] Ikuya Yamada, Akari Asai, and Hannaneh Hajishirzi. Efficient passage retrieval with hashing for open-domain question answering. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 2: Short Papers), pages 979–986, Online, 2021. Association for Computational Linguistics. doi: 10.18653/v1/2021.acl-short.123. URL https://aclanthology.org/2021.acl-short.123.
- Yang et al. [2018] Zhilin Yang, Peng Qi, Saizheng Zhang, Yoshua Bengio, William Cohen, Ruslan Salakhutdinov, and Christopher D. Manning. HotpotQA: A dataset for diverse, explainable multi-hop question answering. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pages 2369–2380, Brussels, Belgium, 2018. Association for Computational Linguistics. doi: 10.18653/v1/D18-1259. URL https://aclanthology.org/D18-1259.
- Zhao et al. [2020] Chen Zhao, Chenyan Xiong, Corby Rosset, Xia Song, Paul N. Bennett, and Saurabh Tiwary. Transformer-xh: Multi-evidence reasoning with extra hop attention. In 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020. OpenReview.net, 2020. URL https://openreview.net/forum?id=r1eIiCNYwS.
Appendices for Baleen
Appendix A Data Details
Multi-Hop Dataset | Train | Dev | Test |
---|---|---|---|
HotPotQA | 90,447 | 7,405 | 7,405 |
HoVer | 18,171 | 4,000 | 4,000 |
As our retrieval corpus for both HotPotQA and HoVer, we use the Wikipedia dump released by Yang et al. [31] from Oct 2017,444https://hotpotqa.github.io/wiki-readme.html which is the Wikipedia dump used officially for both datasets. For each Wikipedia page, this corpus contains only the first paragraph and the passages are already divided into individual sentences. It contains approximately 5M passages (1.5 GiB uncompressed). We use the official data splits for both datasets, described in Table 6.
Appendix B Baleen implementation & hyperparameters
We implement Baleen using Python 3.7 and PyTorch 1.6 and rely extensively on the HuggingFace Transformers library [28].555https://github.com/huggingface/transformers We train and test with automatic mixed precision that is built into PyTorch.
B.1 FLIPR retriever
Our implementation of FLIPR is an extension of the ColBERT [16] open-source code,666https://github.com/stanford-futuredata/ColBERT where we primarily modify the retrieval modeling components (e.g., adding focused late interaction and including condensed fact tokens in the query encoder).
For FLIPR, we fine-tune a BERT-base model (110M parameters). For each round of training, we initialize the model parameters from a ColBERT model previously trained on the MS MARCO Passage Ranking task [20]. To train the single-hop retriever used to initiate the supervision procedure of §3.2, we follow the training strategy of Khattab et al. [17]. In particular, we use this out-of-domain ColBERT model to create training triples, and then we train our retriever (in this case, FLIPR for first-hop) with these triples. Once we have this first-hop model, the rest of the procedure follows Algorithm 1 for latent hop ordering.
Hyperparameter | HoVer | HotPotQA |
Embedding Dimension | 128 | 128 |
Maximum Passage Length | 256 | 256 |
Maximum Query Length: query/overall | 64/512 | 64/512 |
Focused Interaction Filter: / | 32/8 | 32/8 |
Learning rate | ||
Batch size (triples) | 48 | 48 |
Training steps (round #1; per hop) | 10k, 5k, 5k, 5k | 20k, 20k |
Training steps (round #2) | 10k | 40k |
Training Negative Sampling Depth (for each hop) | 1000 | 1000 |
Training Context (Fact) Extraction Depth (from each hop) | 5 | 5 |
Training Positive Sampling Depth (round #1; per hop) | 20, all, all, all | 20, all |
Training Positive Sampling Depth (round #2; per hop) | 10, 10, 10, all | 10, all |
FAISS centroids (probed) | 8192 (16) | 8192 (16) |
FAISS results per vector: training/inference | 256/512 | 256/512 |
Inference Top- Passages Per Hop | 25, 25, 25, 25 | 10, 40 |
Table 7 describes our hyperparameters for FLIPR. We manually explored a limited space of hyperparameters in preliminary experiments, tuning validation-set Retrieval@ Accuracy, with =100 for HoVer and =20 for HotPotQA, while also being cognizant of downstream Psg-EM and Sent-EM. We expect that a larger tuning budget would lead to further gains, which is consistent with the fact that our Psg-EM and Sent-EM on the held-out leaderboard test set are 1.0 and 0.6 points higher than the public validation set with which we developed our methods. Unless otherwise stated, the hyperparameters apply to both FLIPR and ColBERT.
We adopt the default learning rate from ColBERT, namely . We set the embedding dimension to the default and use a batch size of 48 triples. We truncate passages to 256 tokens. For the query encoder, we truncate queries to 64 tokens and allow up to 512 tokens in total, particularly for the condensed facts or reranker context passages from the previous hops. We adopt the ColBERT infrastructure for end-to-end retrieval, where candidate generation is done using nearest-neighbor search with FAISS [14].777https://github.com/facebookresearch/faiss/ See Khattab and Zaharia [16] for use with late interaction models. Since multi-hop queries are much longer than standard IR or OpenQA queries, we conduct candidate generation for FLIPR and ColBERT using the embeddings corresponding to the query tokens (defined in §3.1), and then apply (focused) late interaction on both the query and facts embeddings.
The settings for LHO are also summarized in Table 7, namely the depth (i.e., number of passages) used for sampling negative passages and positive passages and for extracting multi-hop context facts.
B.2 Condensers and rerankers
For the two-stage condenser, we train two ELECTRA-large models (335M parameters each), one per stage. We simply use a [MASK] token to separate the facts, although we expect that any other special-token choice would work similarly provided enough training data is available.
We train the first-stage condenser with triples. We use a cross-entropy loss over the individual sentences of both passages in each example, where the model has to select the positive sentence out of for each positive . We average the cross-entropy loss per example, then across examples per batch. We train the second-hop condenser over a set of 7–9 facts, some positive and others (sampled) negative, using a linear combination of cross-entropy loss for each positive fact (against all negatives) and binary cross-entropy loss for each individual fact.
Hyperparameter | HoVer | HotPotQA |
Learning rate | ||
Batch size | 64 | 64 |
Maximum Sequence Length | 512 | 512 |
Warmup Steps | 1000 | 1000 |
Training steps (stage #1) | 5k | 10k |
Training steps (stage #2) | 5k | 10k |
Negative Sampling Depth (stage #1; per hop) | 20, 20, 20, 20 | 10, 30 |
Negative Sampling Depth (stage #2; for each hop) | 10 | 10 |
Context Sampling Depth (from each hop) | 5 | 5 |
Positive Sampling Depth (stage #1; per hop) | 10, 10, 10, all | 10, all |
Positive Sampling Depth (stage #2; for each hop) | 10 | 10 |
Facts fed to stage # 2: training (inference) | 7–9 (9) | 7–9 (9) |
Table 8 describes our hyperparameters for the condensers.
Claim verification
For HoVer, we train an ELECTRA-large model for claim verification. The input contains the query and the condensed facts and the output is binary, namely supported or unsupported. We use batches of 16 examples and train for 20,000 steps with 2–6 facts per input, but otherwise adopt the same hyperparameters as the second-stage condensers.
Reranker
We similarly use ELECTRA-large for the rerankers. The input contains the query and one reranker-selected passage for each of the previous hops as well as one passage to consider for the current hop. We adopt the same positive and negative sampling as the first-stage condenser, as well as other hyperparameters, but we allow twice the training budget when tuning the number of steps since there is only one stage for reranking. During training and inference, a primary difference between the reranking and condensing architectures is that condensing takes zero or more new facts per hop, whereas reranking always adds exactly one per hop as additional context, namely the top-ranked passage (positive) for inference (training).
Hybrid condenser/reranker implementation
We train a single retriever on top of the first-round retrievers for the condenser and reranker Baleen architectures. During inference, we run two independent pipelines, one with the condenser and the other with the reranker. After both retrieval pipelines have completed, we merge the overall top-100 results by taking the top-13 and top-12 per hop from both pipelines without duplicates, for a total of 254 = 100 unique passages.
B.3 Heuristic ordering: using overlap between passages and titles
To ablate our latent hop ordering (LHO) approach, we consider an algorithm that generalizes Xiong et al. [29]’s heuristic for ordering the two HotPotQA passages for training. This heuristic applies to any number of hops and attempts to capitalize on our understanding of the construction of HotPotQA and HoVer: each hop’s passage is referred to directly by at least one other passage, often by title.
Where short answer strings are available (i.e., in HotPotQA), if only one passage contains that answer string, we treat it as the final hop’s passage and consider the order among the remaining passage(s). We then assign a score to all remaining passages, which reflects whether the full or partial title of the passage appears inside the claim. We take the highest-scoring passage(s) as the target of the current hop, expand the claim with their full text, and repeat this process with any remaining passages.
In case all passage titles are completely unmatched, which is relatively uncommon, we recurse by attempting to pick each possible candidate passage for the current hop, and keep the choice that (at the end of the recursion) results in the largest overlap scores for the remaining hops. For the ablations that use this heuristic, we train in two rounds like with LHO. However, the first round with this heuristic does not require multiple short training runs, one per hop, since all the hops are determined in advance. Instead, we allow the same total training budget for the first round (as well as the second) over all hops.
We note that an additional signal for the order of hops is to recover (from Wikipedia) the hyperlinks between the passages and use them to build a directed graph. However, this violates our assumption that we are only given the text of the passages and would make very strong assumptions about the data, ones specific to the construction of HotPotQA and HoVer.
B.4 Resources used
We conducted our experiments primarily using internal cluster resources. We use four 12GB Titan V GPUs for retrievers and four 32GB V100 GPUs for condensers, rerankers, and readers. Training FLIPR on the four-hop HoVer dataset requires five (4+1) short training runs, for a total time of approximately five hours. Similarly, we encode and index the corpus five times in total (four intermediate and one final time) less than six hours in total. Retrieving positives and negatives for training from the index four times consumes a total of less than three hours. All four hops of retrieval on the validation set with the final FLIPR model take a total of a little over one hour. Training both condenser stages for 5k steps each and training the claim verification reader for 20k steps takes a total of less than eight hours. We use python scripts for pre- and post-processing (e.g., for LHO) and run the condensers during evaluation, which generally add only limited overhead to the running time of our experiments.
Our FLIPR retriever adopts a fine-grained late interaction paradigm like ColBERT (see §2), so our memory footprint is relatively large, as it involves storing a small 256-byte (2-byte 128 dimensions) vector per token. The uncompressed index is about 83 GiBs. We note that the authors of ColBERT Khattab and Zaharia [16] have recently released a quantized implementation that can reduce the storage per vector 4–8 fold, and reducing the storage space of dense retrieval methods through compression and quantization while preserving accuracy is an active area of research [12; 30], with recent encouraging results.
Appendix C The effect of condensing on the context lengths
We compare our condenser architecture of Baleen to a reranker after four hops on HoVer. We find that the average context per query is 91 words for Baleen’s condenser architecture versus 325 words for the reranking ablation, on average. This 3.6 improvement for Baleen’s condenser suggests that for tasks with even more hops, a condenser approach would be less likely to overwhelm typical maximum sequence lengths of existing Transformer architectures.
By design, the condenser architecture scales better to a large number of hops than existing work. For instance, MDR uses beam search, and would require exploring a very large set of paths (e.g., ) from the third hop onwards. As another example, IRRR concatenates long passages together as it conducts more hops. In contrast, the condenser architecture only needs to concatenate short facts together. While this efficiency gap grows when more than two hops are necessary, condensing maintains a comparable cost to MDR and IRRR in the simple two-hop (HotPotQA) setting, where Baleen invokes an ELECTRA-large model only tens of times per query, namely for the top-10 passages from the first hop, the top-40 passages from the second hop, and twice in the second stage of condensing.