TY - JOUR
T1 - Improving conversational search with query reformulation using selective contextual history
AU - Al-Thani, Haya
AU - Elsayed, Tamer
AU - Jansen, Bernard J.
N1 - Publisher Copyright:
© 2022 The Authors
PY - 2023/6
Y1 - 2023/6
N2 - Automated responses to questions for conversational agents, known as conversation passage retrieval, is challenging due to omissions and implied context in user queries. To help address this challenge, queries can be re-written using pre-trained sequence-to-sequence models based on contextual clues from the conversation's history to resolve ambiguities. In this research, we use the TREC conversational assistant (CAsT) 2020 dataset, selecting relevant single sentences from conversation history for query reformulation to improve system effectiveness and efficiency by avoiding topic drift. We propose a practical query selection method that measures clarity score to determine whether to use response sentences during reformulation. We further explore query reformulation as a binary term classification problem and the effects of rank fusion using multiple retrieval models. T5 and BERT retrievals are inventively combined to better represent user information need. Using multi-model fusion, our best system outperforms the best CAsT 2020 run, with an NDCG@3 of 0.537. The implication is that a more selective system that varies the use of responses depending on the query produces a more effective conversational reformulation system. Combining different retrieval results also proved effective in improving system recall.
AB - Automated responses to questions for conversational agents, known as conversation passage retrieval, is challenging due to omissions and implied context in user queries. To help address this challenge, queries can be re-written using pre-trained sequence-to-sequence models based on contextual clues from the conversation's history to resolve ambiguities. In this research, we use the TREC conversational assistant (CAsT) 2020 dataset, selecting relevant single sentences from conversation history for query reformulation to improve system effectiveness and efficiency by avoiding topic drift. We propose a practical query selection method that measures clarity score to determine whether to use response sentences during reformulation. We further explore query reformulation as a binary term classification problem and the effects of rank fusion using multiple retrieval models. T5 and BERT retrievals are inventively combined to better represent user information need. Using multi-model fusion, our best system outperforms the best CAsT 2020 run, with an NDCG@3 of 0.537. The implication is that a more selective system that varies the use of responses depending on the query produces a more effective conversational reformulation system. Combining different retrieval results also proved effective in improving system recall.
KW - Conversational information seeking
KW - Conversational search systems
KW - Multi-stage retrieval systems
KW - Open-domain
UR - http://www.scopus.com/inward/record.url?scp=85144969291&partnerID=8YFLogxK
U2 - 10.1016/j.dim.2022.100025
DO - 10.1016/j.dim.2022.100025
M3 - Article
AN - SCOPUS:85144969291
SN - 2543-9251
VL - 7
JO - Data and Information Management
JF - Data and Information Management
IS - 2
M1 - 100025
ER -