This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Enhanced Facet Generation with LLM Editing

Abstract

In information retrieval, facet identification of a user query is an important task. If a search service can recognize the facets of a user’s query, it has the potential to offer users a much broader range of search results. Previous studies can enhance facet prediction by leveraging retrieved documents and related queries obtained through a search engine. However, there are challenges in extending it to other applications when a search engine operates as part of the model. First, search engines are constantly updated. Therefore, additional information may change during training and test, which may reduce performance. The second challenge is that public search engines cannot search for internal documents. Therefore, a separate search system needs to be built to incorporate documents from private domains within the company. We propose two strategies that focus on a framework that can predict facets by taking only queries as input without a search engine. The first strategy is multi-task learning to predict SERP. By leveraging SERP as a target instead of a source, the proposed model deeply understands queries without relying on external modules. The second strategy is to enhance the facets by combining Large Language Model (LLM) and the small model. Overall performance improves when small model and LLM are combined rather than facet generation individually.

Keywords: facet generation, search clarification, sub-intent mining, large language model

\NAT@set@cites

Enhanced Facet Generation with LLM Editing


Joosung Lee, Jinhong Kim
NAVER
{rung.joo, kim.jinhong}@navercorp.com

Abstract content

1.   Introduction

Search clarification has been an area of interest in information retrieval for a long time (Radlinski and Craswell, 2017; Vtyurina et al., 2017). Users send queries with various sub-intents to the search system and expect various search results. These sub-intents are called facets. For instance, facets could include "warcraft game", "warcraft movie", "warcraft book", "warcraft history" and more when a user searches for "warcraft". Previous studies (Hashemi et al., 2021; Samarinas et al., 2022) introduce facet generation task as generating facets from the query. If a search system can predict the query facets in advance, it can provide more diverse and higher-quality search results.

The previous studies (Hashemi et al., 2021; Samarinas et al., 2022; Hashemi et al., 2022; Liu, 2023; Zhao et al., 2023) demonstrated that models can improve their performance in generating various query facets by leveraging Search Engine Result Page (SERP). The most commonly used information in SERP is the snippet of the retrieved document. Constructing input with both query and document snippets provides the model with richer information, leading to improved performance in facet prediction. However, there are several challenges to commercialize these methods. First, public search engines like Bing or Google are continuously updated. Search algorithms change over time, and user documents are continually updated. The external researchers cannot grasp the principles and changes of the private search algorithm. Therefore, if the search engine is used as part of the model, SERP changes between training and testing, leading to a drop in performance. The second challenge is that public search engines only search for public documents. If the systems need to create a query facet for an in-house service, the facet distribution to target will be different. However, there is a significant cost involved in constructing a separate search engine to leverage in-house documents. Finally, external communication is essential for SERP. Therefore, the previous methods are difficult for customers who want on-premise services to consider.

We focus on a framework that operates independently of the search engine by using only a query as input during testing. We propose two strategies to predict query facets without SERP. The first is multi-task learning, which uses SERP only in the training and not in the test. The approach of simply concatenating documents as input for training is not efficient during the test (refer to Section 4). Therefore, we consider SERP as the target to improve the performance of our model. The second is editing facets using LLM. Recently, LLM has made remarkable progress since InstructGPT (Ouyang et al., 2022), achieving high performance in a variety of tasks. However, simply instructing an LLM to generate query facets can result in inaccurate facet generation. Because LLM does not know the distribution of the dataset, it is difficult to predict the facet that fits the target. We improve performance by editing the facets predicted by the fine-tuned small model with LLM. This is the effect of allowing LLM to generate accurate facets by informing LLM of the distribution of the dataset through a small model learned with the training dataset. In other words, LLM editing is more effective than end-to-end generation because a fine-tuned small model generates intermediate results to the target facet. We also demonstrate that LLM editing works effectively on the previous models as well.

2.   Related Work

Previous studies leverage SERP to enhance facet generation performance.  Hashemi et al., 2021 proposes the NMIR framework, which learns multiple intent representations by taking query and retrieved documents as input.  Samarinas et al., 2022 introduces five methods based on query and retrieved documents. FG (Facet Generation) generates a facet by inputting the retrieved document. SL (Sequence Labeling) determines whether a token is a facet through sequence labeling in the retrieved document. EFC (Extreme Facet Classification) considers terms that frequently appear in facets as classes and trains a classifier to find facet terms in documents.  Hashemi et al., 2022 proposes a permutation-invariant approach based on NMIR.  Liu, 2023 proposes a method that leverages related queries in addition to documents to enhance the performance.  Zhao et al., 2023 finds better facets for queries by combining external structured information in addition to documents. SR (Structured Relation) uses hypernyms through external knowledge (Concept Graph Wang et al. (2015) and WebIsA Seitner et al. (2016)) and list structure through HTML as input.

Search Clarification is closely related to interactive search systems (Sekulić et al., 2021; Aliannejadi et al., 2021). This is because the search system can clarify the user’s intent and provide more accurate services. Therefore, for ambiguous queries, the search system asks a clarifying question (Aliannejadi et al., 2019).  Zamani et al., 2020a identifies the taxonomy of clarification and generates clarifying questions.  Rao and Daumé III, 2018 builds a neural network model for the task of ranking clarification questions.

Facet identification of a query is also related to learning the query representation or expansion. Traditionally, query representations were constructed from term frequencies in search logs Salton et al. (1975).  Rocchio Jr, 1971; Lavrenko and Croft, 2017 introduce query representation using query expansion and relevance feedback.  Mikolov et al., 2013 learns the relationships between adjacent words in the corpus and represents words as embeddings. Words and queries can be expressed through pre-trained language models learned with large corpora such as BERT Devlin et al. (2018) and GPT2 Radford et al. (2019). Recently, Wang et al., 2023; Jagerman et al., 2023 introduce query expansion through LLM.

3.   Task Definition

This paper focuses on generating facets based on only queries. In the training, Ttrain={(q1,D1,R1,F1),,(qN,DN,RN,FN)}T_{train}=\{(q_{1},D_{1},R_{1},F_{1}),...,(q_{N},D_{N},R_{N},F_{N})\} where qiq_{i} represents a query, Di={di1,,dim}D_{i}=\{d_{i1},...,d_{im}\} consists of snippets from mm retrieved documents, Ri={ri1,,rit}R_{i}=\{r_{i1},...,r_{it}\} contains tt related queries, obtained from query logs, and Fi={fi1,,fik}F_{i}=\{f_{i1},...,f_{ik}\} represents kk target facets. In the test, Ttest={(q1,F1),(qM,FM)}T_{test}=\{(q_{1},F_{1}),...(q_{M},F_{M})\} where we generate FiF_{i} from qiq_{i}. In contrast to the training, DD and RR cannot be used during the test.

Training / Test Term Overlap (F1) Exact Match (F1) Set BLEU-mean Set BERTScore (F1)
QD / QD 0.2914 0.0732 0.3265 0.8794
QD / Q 0.1299 0.004 0.2367 0.8472
Table 1: Performance changes according to the input of training and test. Q means the query, D means the snippet of the document. D and Q are concatenated to form the input.

4.   Method

The previous methods are frameworks that learn a model by using the SERP for the query as input. Similar to the previous method, the model in Table 1 is fine-tuned to generate facets based on BART-base (Lewis et al., 2019) by receiving queries and documents. The performance of the fine-tuned model changes depending on the input configuration of training and testing. The performance will significantly decrease if SERP used for training is not used in testing. As a result, the most ideal scenario is a situation where the input configuration in training and test is the same.

Our method assumes a scenario where SERP is not available for test. Therefore, we utilize multi-task learning by placing information in the target rather than the input of the model. Additionally, we leverage vast knowledge by combining LLM and small models. The term "small model" denotes a model that can be trained on a single GPU. In our experiments, this corresponds to BART-base. The term "LLM" denotes a model of size 7B or larger and can be used in various tasks through pretraining and instruction-tuning.

4.1.   Multi-task Learning

The input is constructed by prepending special tokens to the query, which allows us to control the target to be generated. Special tokens include [facet], [document], and [related], and are used to generate facets, snippets of documents, and related queries, respectively. The input is as follows:

is=concat([s],query)i_{s}=\text{concat}([s],\text{query}) (1)

where s{facet,document,related}s\in\{facet,document,related\}. The target output is composed of each sentence separated by "," as follows:

os=``s1,s2,"o_{s}=``s_{1},s_{2},..." (2)

where sis_{i} is the target sentence corresponding to ss. The loss is calculated as cross entropy as follows:

Ls=1Ni=1NCE(f(is),os)L_{s}=\frac{1}{N}\sum^{N}_{i=1}CE(f(i_{s}),o_{s}) (3)

The sum of losses based on the targets used in multi-task learning is the final loss. The trained model can generate not only facets but also documents or related queries. The model’s additional capabilities assist in generating accurate facets from queries.

4.2.   LLM Editing

Multi-task learning improves the performance of small models without relying on the search engine. However, the fine-tuned model still has a limitation in not being able to leverage rich external information during the test. Therefore, we propose a strategy to mitigate this drawback by leveraging an LLM with extensive knowledge from a large corpus.

LLM editing is a technique that refines facets generated by a fine-tuned small model. When LLM is instructed to generate target facets corresponding to a query, it relies on general generation capabilities. This generative ability comes from a massive pretraining corpus and instruction tuning. Therefore, it is difficult to convey the distribution of the desired target facets in the dataset to LLM simply through few-shot demonstrations of query and facet pairs. On the other hand, the fine-tuned small model knows the distribution of facets to be generated because it has been learned from the training dataset. Therefore, we provide the facets predicted by the small model to LLM to regenerate the improved facets. With the assistance of a small model, LLM can perform modified facet identification from a state close to the target facets, making the task easier. In other words, it is a method of leveraging the distribution of the training dataset, which is the knowledge of a fine-tuned small model.

### User:
The predicted facets for ‘{example query1}’ are ‘{predicted facets1}’.
But the correct facets are ‘{label facets1}’.
The predicted facets for ‘{example query2}’ are ‘{predicted facets2}’.
But the correct facets are ‘{label facets2}’.
As in the example above, modify the predicted facets.
The predicted facets for ‘{input query}’ are ‘{predicted facets}’.
What are the correct facets?
### Assistant:
The correct facets for ‘{input query}’ are
Table 2: Prompt given to LLM. {predicted facets} are the output of the small model.
Editing Prompt.

Table 2 shows the prompt for LLM editing the results of the small model. We inform LLM of two-shot demonstrations (predicted facets => label facets). If LLM does not combine small models, we instruct LLM to generate facets via few-shot or zero-shot. In E(zero), LLM is instructed to generate facets without information about the dataset distribution. In E(few), LLM can obtain limited information via standard prompting through few-shot demonstrations. In Appendix A, prompt configuration is introduced in more detail.

Model Term Overlap (F1) Exact Match (F1) Set BLEU-mean Set BERTScore (F1)
F 0.2374 0.0284 0.2898 0.8657
E(zero) 0.0817 0.0134 0.1512 0.8586
E(few) 0.2101 0.0424 0.3511 0.8803
FR+M 0.2519 0.0337 0.2963 0.8687
FR+M+E 0.2385 0.0509 0.3766 0.8807
FD+M 0.2508 0.0338 0.2992 0.871
FD+M+E 0.2381 0.0518 0.3772 0.8812
Table 3: Our model’s performance to multi-task learning and LLM editing. F, D, and R indicate that the model was trained to generate facet, document, and related queries, respectively. +M represents multi-task learning and +E represents LLM editing combined into the small model.

5.   Experiments

We followed previous studies (Hashemi et al., 2021; Samarinas et al., 2022; Liu, 2023; Hashemi et al., 2022; Zhao et al., 2023) and used BART-base as a small model. ChatGPT (OpenAI, 2022) or GPT4 are private LLMs and have cost issues. Additionally, OpenAI’s models are constantly updated, making it difficult to reproduce our results, so we use open-source LLMs with public parameters. At the time of our experiments, we used UP 30B (Upstage, 2023), which ranks high on the LLM leaderboard Beeching et al. (2023).

5.1.   Dataset

MIMICS dataset (Zamani et al., 2020) 111https://github.com/microsoft/MIMICS is widely used in search clarification or facet generation. MIMICS is collected from the Bing search engine and consists of three subsets. Following previous studies, MIMICS-Click is used as a training dataset and MIMICS-Manual is used as a test dataset. SERP was used as public data 222http://ciir.cs.umass.edu/downloads/mimics-serp/MIMICS-BingAPI-results.zip.

5.2.   Evaluation Metric

5.2.1.   Automatic Evaluation

We follow the automatic metric proposed in Hashemi et al., 2021. Term Overlap indicates that the terms of generated facets and ground truth facets overlap. Exact Match indicates whether the generated facets are identical to the ground truth facets. Set BLEU-mean represents the average of the 1-gram, 2-gram, 3-gram, and 4-gram scores of each facet sentence. Set BERTScore (Zhang* et al., 2020) calculates the similarity of each facet sentence using RoBERTa-large (Liu et al., 2019). For intuitive analysis, we use a single score for metrics such as each F1 or average score. Evaluation scripts are provided in  Samarinas et al., 2022.

Facets refer to the sub-intents desired by the user who
searched the query.
The following are facets about "{query}".
Which facets set is better? (without explanation)
A: {predicted facets by model A}
B: {predicted facets by model B}
Table 4: Assessment prompt instructed to LLMs

5.2.2.   LLM-based Evaluation

It is difficult to select the best model only by automatic evaluation. Since automatic evaluation has four metrics, a good model depends on the metric. Previous studies Chiang and Lee (2023); Liu et al. (2023) introduce that the LLM evaluator works as a good evaluator in various NLG tasks. LLM evaluators show a high correlation with human evaluators and show more reliable results than traditional metrics (e.g. BLEU, ROUGE, METEOR). Additionally, LLM assessments are highly reproducible and unaffected by previous test samples. LLM evaluator can be utilized in various ways, such as through a win-lose method or by computing scores.  Kocmi and Federmann, 2023 introduces evaluating the LLM evaluator using a scoring method in translation tasks, but it has the disadvantage that the score distribution is biased to one side. Since we only need to determine the superiority between two compared models, we utilize the LLM evaluator in a win-lose method. Inspired by these results, we attempt to evaluate using gemini-pro (Team et al., 2023) and GPT4 (OpenAI, 2023), which are known to have the best performance as LLM.

Table 4 shows the model assessment prompt. We provided the predicted facets of models A and B to the LLM and asked which one was better. Therefore, the model responds with either A or B. However, LLM often has a different response format because it is a generative model. To minimize the risk of not being able to verify the correct answer due to such parsing, we set it to temperature=0.1, top_p=1. Nonetheless, samples responding in a different format are excluded from the evaluation.

5.3.   Result and Discussion

Table 3 shows the experimental results of our strategies. F model is fine-tuned only for facet generation. E models are the result of instructing LLM to generate facets by providing a query and few(two)- or zero-shot demonstrations without a small model. +M is multi-task learning, where the model is trained to generate related queries (R) or snippets of documents (D) in addition to facet generation. +E indicates that LLM editing was performed on the results of the small model. Some examples and statistics of the generated facets are introduced in Appendix B.

Model Training / Test Term Overlap (F1) Exact Match (F1) Set BLEU-mean Set BERTScore (F1)
FG Q / Q 0.0664 0.0306 0.0751 0.8446
FD+M (ours) QD / Q 0.2508 0.0338 0.2992 0.871
FD+M+E (ours) 0.2381 0.0518 0.3772 0.8812
FG QD / QD 0.2919 0.0707 0.3544 0.8785
FG+E 0.2702 0.0694 0.4092 0.8844
SL 0.1914 0.0515 0.2483 0.8748
SL+E 0.1895 0.0542 0.2618 0.8769
EFC 0.0515 0.0289 0.0544 0.423
EFC+E 0.0678 0.028 0.0863 0.8626
SR QDS / QDS 0.274 0.0888 0.4297 0.8903
SR+E 0.2626 0.849 0.4302 0.8896
Table 5: Performance comparison between our model and other models. S in QDS indicates structured information. Bold indicates the highest performance in each test type.

Multi-task learning improves performance even when there is only a query in the test. Both FR+M and FD+M outperform F in all four metrics. The enhanced capability of the small model to infer not only the facets but also retrieved documents or related queries leads to a better understanding of the query. When considering all automatic metrics, FD+M is slightly superior to FR+M. As a result, we confirmed that a snippet of the document is more effective for small models than a related query.

LLM editing enhances the facets generated by the small model. FD+M+E slightly reduces performance in Term Overlap compared to FD+M, but improves performance in the other three metrics. Overall, FD+M+E is better than FD+M. FD+M+E outperforms E(few) in all aspects, which proves that the facets generated by the small model contribute to the target facet distribution. E(few) receives a distribution of the dataset via in-context learning through few-shot demonstrations, but it is very limited information. E(zero) has lower performance because it does not know the distribution of the dataset at all. It is important to note that facets generated by LLM without prior information are difficult to match the target distribution. In other words, the method of combining the small model and LLM proves to be more effective than simply fine-tuning and standard prompting. In the Appendix C, we show that LLM editing is effective regardless of LLM size.

5.3.1.   Comparison with Previous Methods

Table 5 shows a comparison of the performance of our model with the previous methods. The descriptions of comparative models are in Section 2. FD+M+E demonstrates the second-best performance in Set BLEU-mean and Set BERTScore among previous models that did not combine LLM editing. This means that FD+M+E generates more semantically sufficient facets than FG, which is the best in QD test type without SERP. SR improves performance with structured information (hypernyms and HTML) in addition to document snippets but is more dependent on SERP. From the results of SR, we expect that leveraging structured information in multi-task learning will lead to improved performance in the future.

5.3.2.   Result of LLM-based Evaluation

Table 6 shows comparison results between FD+M+E and other models. We selected FD+M, FG(QD/QD) and SR as comparison models. The numbers in the cell are the percentages that FD+M+E won in competition with other models. For example, GPT4 determines that FD+M+E is better than SR for 63.86% of the test data. Both LLMs determine that FD+M+E performs better than the other three models. These results prove that FD+M+E is a more effective method than previous methods using SERP. In other words, FD+M+E is better than the previous SoTA in terms of LLM-based evaluation perspective. In particular, the higher win rate compared to FD+M indicates that LLM editing is an important factor. Since LLM evaluators are known to be more relevant to human evaluators than traditional metrics, our method is considered to have achieved state-of-the-art performance without leveraging SERP.

5.3.3.   Combine LLM Editing with Previous Methods

We applied LLM editing to previous models that used SERP as input. Table 5 shows that the effect of LLM editing varies depending on the performance of the models. Similar to Table 3, LLM editing tends to improve overall Set BLUE-mean and Set BERTScore performance. However, LLM editing tends to decrease Term Overlap performance, and changes in Exact Match performance vary depending on the model. LLM editing substantially enhances the performance of models such as EFC, which demonstrate inferior performance overall. We demonstrate that LLM editing is an effective technique for regenerating semantic facets regardless of the small model.

FD+M+E vs Comparison model
Comparison model Gemini-pro GPT4
FD+M 72.33 90.65
FG (QD/QD) 59.05 78.59
SR 74.08 63.86
Table 6: Win ratio of FD+M+E that competed with the other models. Gemini-pro and GPT4 are used as LLM evaluators.

6.   Conclusion

The proposed method generates facets using only queries, which eliminates the dependency on search engines. To address the limitation of not being able to utilize SERP as input, we propose two strategies: multi-task learning and LLM editing. Multi-task learning helps the small model better understand the query. LLM receives prior information from a small model and generates improved facets. Even without SERP, FD+M+E shows similar performance to FG in automatic evaluation and achieves the best performance in LLM-based evaluation. LLM editing is a way to effectively combine small models and LLMs in various NLP tasks, rather than using them separately. Therefore, our method can be extended to various NLP tasks.

7.   Bibliographical References

\c@NAT@ctr

  • Aliannejadi et al. (2021) Mohammad Aliannejadi, Julia Kiseleva, Aleksandr Chuklin, Jeff Dalton, and Mikhail Burtsev. 2021. Building and evaluating open-domain dialogue corpora with clarifying questions. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 4473–4484, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
  • Aliannejadi et al. (2019) Mohammad Aliannejadi, Hamed Zamani, Fabio Crestani, and W. Bruce Croft. 2019. Asking clarifying questions in open-domain information-seeking conversations. In Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR’19, page 475–484, New York, NY, USA. Association for Computing Machinery.
  • Beeching et al. (2023) Edward Beeching, Clémentine Fourrier, Nathan Habib, Sheon Han, Nathan Lambert, Nazneen Rajani, Omar Sanseviero, Lewis Tunstall, and Thomas Wolf. 2023. Open llm leaderboard. https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard.
  • Chia et al. (2023) Yew Ken Chia, Guizhen Chen, Luu Anh Tuan, Soujanya Poria, and Lidong Bing. 2023. Contrastive chain-of-thought prompting. arXiv preprint arXiv:2311.09277.
  • Chiang and Lee (2023) Cheng-Han Chiang and Hung-yi Lee. 2023. Can large language models be an alternative to human evaluations? In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 15607–15631, Toronto, Canada. Association for Computational Linguistics.
  • Devlin et al. (2018) Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805.
  • Hashemi et al. (2021) Helia Hashemi, Hamed Zamani, and W. Bruce Croft. 2021. Learning multiple intent representations for search queries. In Proceedings of the 30th ACM International Conference on Information & Knowledge Management, CIKM ’21, page 669–679, New York, NY, USA. Association for Computing Machinery.
  • Hashemi et al. (2022) Helia Hashemi, Hamed Zamani, and W. Bruce Croft. 2022. Stochastic optimization of text set generation for learning multiple query intent representations. In Proceedings of the 31st ACM International Conference on Information & Knowledge Management, CIKM ’22, page 4003–4008, New York, NY, USA. Association for Computing Machinery.
  • Hsieh et al. (2023) Cheng-Yu Hsieh, Chun-Liang Li, Chih-kuan Yeh, Hootan Nakhost, Yasuhisa Fujii, Alex Ratner, Ranjay Krishna, Chen-Yu Lee, and Tomas Pfister. 2023. Distilling step-by-step! outperforming larger language models with less training data and smaller model sizes. In Findings of the Association for Computational Linguistics: ACL 2023, pages 8003–8017, Toronto, Canada. Association for Computational Linguistics.
  • HyperbeeAI (2023) HyperbeeAI. 2023. Hyperbeeai/tulpar-7b-v0. https://huggingface.co/HyperbeeAI/Tulpar-7b-v0.
  • Jagerman et al. (2023) Rolf Jagerman, Honglei Zhuang, Zhen Qin, Xuanhui Wang, and Michael Bendersky. 2023. Query expansion by prompting large language models. arXiv preprint arXiv:2305.03653.
  • Kocmi and Federmann (2023) Tom Kocmi and Christian Federmann. 2023. Large language models are state-of-the-art evaluators of translation quality. In Proceedings of the 24th Annual Conference of the European Association for Machine Translation, pages 193–203, Tampere, Finland. European Association for Machine Translation.
  • Lavrenko and Croft (2017) Victor Lavrenko and W Bruce Croft. 2017. Relevance-based language models. In ACM SIGIR Forum, volume 51, pages 260–267. ACM New York, NY, USA.
  • Lee et al. (2023) Ariel N. Lee, Cole J. Hunter, Nataniel Ruiz, Bleys Goodson, Wing Lian, Guan Wang, Eugene Pentland, Austin Cook, Chanvichet Vong, and "Teknium". 2023. Openorcaplatypus: Llama2-13b model instruct-tuned on filtered openorcav1 gpt-4 dataset and merged with divergent stem and logic dataset model. https://huggingface.co/Open-Orca/OpenOrca-Platypus2-13B.
  • Lewis et al. (2019) Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed, Omer Levy, Veselin Stoyanov, and Luke Zettlemoyer. 2019. BART: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. CoRR, abs/1910.13461.
  • Liu (2023) Xinyu Liu. 2023. Query sub-intent mining by incorporating search results with query logs for information retrieval. In 2023 IEEE 8th International Conference on Big Data Analytics (ICBDA), pages 180–186.
  • Liu et al. (2023) Yang Liu, Dan Iter, Yichong Xu, Shuohang Wang, Ruochen Xu, and Chenguang Zhu. 2023. G-eval: NLG evaluation using gpt-4 with better human alignment. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 2511–2522, Singapore. Association for Computational Linguistics.
  • Liu et al. (2019) Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. Roberta: A robustly optimized BERT pretraining approach. CoRR, abs/1907.11692.
  • Mikolov et al. (2013) Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg S Corrado, and Jeff Dean. 2013. Distributed representations of words and phrases and their compositionality. Advances in neural information processing systems, 26.
  • OpenAI (2022) OpenAI. 2022. Chatgpt: Optimizing language models for dialogue.
  • OpenAI (2023) OpenAI. 2023. Gpt-4 technical report.
  • Ouyang et al. (2022) Long Ouyang, Jeffrey Wu, Xu Jiang, Diogo Almeida, Carroll Wainwright, Pamela Mishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Ray, et al. 2022. Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems, 35:27730–27744.
  • Radford et al. (2019) Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, Ilya Sutskever, et al. 2019. Language models are unsupervised multitask learners. OpenAI blog, 1(8):9.
  • Radlinski and Craswell (2017) Filip Radlinski and Nick Craswell. 2017. A theoretical framework for conversational search. In Proceedings of the 2017 Conference on Conference Human Information Interaction and Retrieval, CHIIR ’17, page 117–126, New York, NY, USA. Association for Computing Machinery.
  • Rao and Daumé III (2018) Sudha Rao and Hal Daumé III. 2018. Learning to ask good questions: Ranking clarification questions using neural expected value of perfect information. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 2737–2746, Melbourne, Australia. Association for Computational Linguistics.
  • Rocchio Jr (1971) Joseph John Rocchio Jr. 1971. Relevance feedback in information retrieval. The SMART retrieval system: experiments in automatic document processing.
  • Salton et al. (1975) Gerard Salton, Anita Wong, and Chung-Shu Yang. 1975. A vector space model for automatic indexing. Communications of the ACM, 18(11):613–620.
  • Samarinas et al. (2022) Chris Samarinas, Arkin Dharawat, and Hamed Zamani. 2022. Revisiting open domain query facet extraction and generation. In Proceedings of the 2022 ACM SIGIR International Conference on Theory of Information Retrieval, ICTIR ’22, page 43–50, New York, NY, USA. Association for Computing Machinery.
  • Seitner et al. (2016) Julian Seitner, Christian Bizer, Kai Eckert, Stefano Faralli, Robert Meusel, Heiko Paulheim, and Simone Paolo Ponzetto. 2016. A large DataBase of hypernymy relations extracted from the web. In Proceedings of the Tenth International Conference on Language Resources and Evaluation (LREC’16), pages 360–367, Portorož, Slovenia. European Language Resources Association (ELRA).
  • Sekulić et al. (2021) Ivan Sekulić, Mohammad Aliannejadi, and Fabio Crestani. 2021. Towards facet-driven generation of clarifying questions for conversational search. In Proceedings of the 2021 ACM SIGIR International Conference on Theory of Information Retrieval, ICTIR ’21, page 167–175, New York, NY, USA. Association for Computing Machinery.
  • Team et al. (2023) Gemini Team, Rohan Anil, Sebastian Borgeaud, Yonghui Wu, Jean-Baptiste Alayrac, Jiahui Yu, Radu Soricut, Johan Schalkwyk, Andrew M Dai, Anja Hauth, et al. 2023. Gemini: a family of highly capable multimodal models. arXiv preprint arXiv:2312.11805.
  • Upstage (2023) Upstage. 2023. upstage/llama-30b-instruct-2048. https://huggingface.co/upstage/llama-30b-instruct-2048.
  • Vtyurina et al. (2017) Alexandra Vtyurina, Denis Savenkov, Eugene Agichtein, and Charles L. A. Clarke. 2017. Exploring conversational search with humans, assistants, and wizards. In Proceedings of the 2017 CHI Conference Extended Abstracts on Human Factors in Computing Systems, CHI EA ’17, page 2187–2193, New York, NY, USA. Association for Computing Machinery.
  • Wang et al. (2023) Liang Wang, Nan Yang, and Furu Wei. 2023. Query2doc: Query expansion with large language models. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 9414–9423, Singapore. Association for Computational Linguistics.
  • Wang et al. (2015) Zhongyuan Wang, Haixun Wang, Ji-Rong Wen, and Yanghua Xiao. 2015. An inference approach to basic level of categorization. In Proceedings of the 24th ACM International on Conference on Information and Knowledge Management, CIKM ’15, page 653–662, New York, NY, USA. Association for Computing Machinery.
  • Wei et al. (2022) Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, Fei Xia, Ed Chi, Quoc V Le, Denny Zhou, et al. 2022. Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems, 35:24824–24837.
  • Zamani et al. (2020a) Hamed Zamani, Susan Dumais, Nick Craswell, Paul Bennett, and Gord Lueck. 2020a. Generating clarifying questions for information retrieval. In Proceedings of The Web Conference 2020, WWW ’20, page 418–428, New York, NY, USA. Association for Computing Machinery.
  • Zamani et al. (2020b) Hamed Zamani, Gord Lueck, Everest Chen, Rodolfo Quispe, Flint Luu, and Nick Craswell. 2020b. Mimics: A large-scale data collection for search clarification. In Proceedings of the 29th ACM International Conference on Information & Knowledge Management, CIKM ’20, page 3189–3196, New York, NY, USA. Association for Computing Machinery.
  • Zhang* et al. (2020) Tianyi Zhang*, Varsha Kishore*, Felix Wu*, Kilian Q. Weinberger, and Yoav Artzi. 2020. Bertscore: Evaluating text generation with bert. In International Conference on Learning Representations.
  • Zhao et al. (2023) Ziliang Zhao, Zhicheng Dou, Yu Guo, Zhao Cao, and Xiaohua Cheng. 2023. Improving search clarification with structured information extracted from search results. In Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, KDD ’23, page 3549–3558, New York, NY, USA. Association for Computing Machinery.

8.   Language Resource References

\c@NAT@ctr

 

  • Zamani et al. (2020) Hamed Zamani, Gord Lueck, Everest Chen, Rodolfo Quispe, Flint Luu, and Nick Craswell. 2020. Mimics: A large-scale data collection for search clarification. In Proceedings of the 29th ACM International Conference on Information & Knowledge Management, CIKM ’20, page 3189–3196, New York, NY, USA. Association for Computing Machinery.
Refer to caption
Figure 1: Overview of previous methods and our method.
query FD+M FD+M+E ground-truth
carrots carrots for sale, carrots care
carrots nutrition, carrots health benefits,
carrots recipes
grow carrots, cook carrots,
store carrots, freeze carrots
orange
orange tree, orange flower
orange fruit, orange juice,
orange tree, orange flower
orange the color, orange the fruit,
orange the company
firewall
firewall windows 10, windows 7,
windows 8, windows xp
firewall types, firewall software,
firewall hardware, firewall configuration
firewall hardware, firewall the movie
Table 7: Examples generated by FD+M and FD+M+E
Model
average number
of a set
average length of
generated facets
proportion of query
included in facet
proportion of duplicate
facets in a set
FD+M 2.39 17.9 61.87 0.038
FD+M+E 4.13 16.67 57.33 0.013
groud-truth 3.01 15 47.96 0
Table 8: Statistics of facets generated with FD+M, FD+M+E, and ground-truth
### User:
The facets for ‘{query}’ are ‘{facets}’. As in the format above,
generate facets related to the query within 5, separated by ‘,’.
### Assistant:
The facets for ‘{input query}’ are
Table 9: Prompt for E(zero) model. This prompt instructs LLM to generate the facet without any prior information, where {query} and {facets} are not examples, but strings themselves.
### User:
The facets for ‘{example query1}’ are ‘{correct facets1}’.
The facets for ‘{example query2}’ are ‘{correct facets2}’.
### Assistant:
The correct facets for ‘{input query}’ are
Table 10: Prompt for E(few) model. It shows only few-shot demonstrations, not facet modifications.
LLM Model Term Overlap (F1) Exact Match (F1) Set BLEU-mean Set BERTScore (F1)
HB 7B (HyperbeeAI, 2023) E(few) 0.1905 0.0241 0.206 0.8691
FD+M+E 0.2495 0.049 0.3657 0.878
OO 13B (Lee et al., 2023) E(few) 0.1852 0.0355 0.3148 0.8677
FD+M+E 0.2477 0.0477 0.3721 0.878
UP 30B (Upstage, 2023) E(few) 0.2101 0.0424 0.3511 0.8803
FD+M+E 0.2381 0.0518 0.3772 0.8812
Table 11: Performance based on other LLMs. Bold indicates the best performance.
LLM Average ARC HellaSwag MMLU TruthfulQA
UP 30B 67.02 64.93 84.94 61.9 56.3
OO 13B 63.19 61.52 82.27 58.85 50.11
HA 7B 59.89 56.31 79.01 52.55 51.68
Table 12: LLM benchmark performance reported on LLM leaderboard

Appendix A Details of Editing Prompt

Table 9 and Table 10 show prompts for E(zero) and E(few), respectively. Since E(zero) lacks information about facets, it provides structured information about the format and number of facets in the "### USER" section. We also experiment with other LLM sizes in Appendix C. In the prompt of OO 13B, special phrases are guided to use "Instruction" and "Response". Therefore, the prompt of OO 13B is composed of "User" and "Assistant" replaced with "Instruction" and "Response", respectively. We used the same generation configuration settings for all LLMs.

Chain-of-thought prompting (CoT) Wei et al. (2022) and contrastive chain-of-thought prompting (CCoT) Chia et al. (2023) are effective techniques in LLM prompting. CoT enhances performance by prompting LLM to generate rationales along with responses. Our goal is not to find a better prompt for facet identification but to improve performance by combining the fine-tuned small model and LLM. Therefore, we leave the exploration of better prompts for future research.

Appendix B Statistics of Generated Facet Set

Table 7 shows examples of our model. For carrots, LLM editing modifies facets through suffixes of nutrition, recipes, and health benefits. For orange, LLM editing adds two facets while maintaining some of the facets predicted by FD+M. For firewall, LLM editing adds a firewall prefix and creates a wider range of facets.

Table 8 shows facet statistics of the proposed model. LLM editing increases the number of existing facets from an average of 2.39 (FD+M) to 4.13 (FD+M+E) because LLM finds intentions that FD+M could not cover. The average length of each generated facet does not differ significantly. Furthermore, we also measure the extent to which the query string is important in constructing facets. As shown in the example in Table 7, the intention of the facet including the query string is intuitively clear. However, in terms of measurement results, the order of query inclusion rates is higher for FD+M > FD+M+E > ground-truth. In other words, we confirmed that query string is not an essential element in configuring facets, and LLM editing removes queries from facets as needed. For example, there are cases where the ground-truth sample has the query "internet explorer" and the facet is "windows 10". When the generated facet set contains duplicate facets, it can negatively impact the overall performance of the facet set. As a result of measurement, LLM editing further improves the facet set by removing duplicate facets. Through the distribution of the generated facet set, we gain insights that limiting the number of duplicate facets and the total number of generated facets can improve future performance.

Appendix C Effects of LLM Size

We conducted experiments on LLM(7B, 13B) in addition to LLM(30B). Table 12 shows LLM benchmark performance. As the size increases, LLM performance improves. Table 11 shows the facet generation performance of FD+M+E when using different LLMs(7B, 13B, 30B). We observe that even when utilizing LLM(7B, 13B), for editing instead of LLM(30B), there is still a notable performance improvement. Smaller LLMs tend to have higher performance in Term Overlap. Therefore, LLM editing is an effective prompt technique regardless of the size of LLM.

In the case of E(few), which does not use the small model, there is a significant difference in performance between LLM(30B) and LLM(13B,7B). That is, in few-shot inference, similar to the LLM benchmark, larger LLMs generally outperform smaller LLMs. However, when combined with a small model, the difference between FD+M+E is reduced, which shows that the role of the small model is crucial in generating facets of LLM. The small model serves as an intermediary bridge between the query and facets because it has learned the distribution of facets through the training dataset. With the assistance of this small model, LLM can generate the desired facets. We attempted various other LLMs, but it was challenging to find an LLM that excelled in all four metrics. This is considered a trade-off related to LLM size, and further research is needed to achieve better results in all metrics.