--- license: cc-by-nc-4.0 task_categories: - text-generation language: - en tags: - math pretty_name: backtracing size_categories: - n>1k --- # Backtracing: Retrieving the Cause of the Query
Paper • Code • Video • Citation
**Authors:** Rose E. Wang, Pawan Wirawarn, Omar Khattab, Noah Goodman, Dorottya Demszky Findings of EACL, Long Paper, 2024. ## What is Backtracing?  Many online content portals allow users to ask questions to supplement their understanding (e.g., of lectures or news articles). While information retrieval (IR) systems may provide answers for such user queries, they do not directly assist content creators identify segments that _caused_ a user to ask those questions; this can be useful for several purposes like helping improve their content. We introduce the task of **backtracing**, in which systems retrieve the text segment that most likely provoked a user query. ## This Repository In this repository, you will find: 1. The first **benchmark for backtracing**, composed of three heterogeneous datasets and causal retrieval tasks: understanding the cause of (a) student confusion in the _Lecture_ domain, (b) reader curiosity in the _News Article_ domain, and (c) user emotion in the _Conversation_ domain. 2. **Evaluations of a suite of retrieval systems on backtracing**, including: BM25, bi-encoder methods, cross-encoder methods, re-ranker methods, gpt-3.5-turbo-16k, and several likelihood-based methods that use pre-trained language models to estimate the probability of the query conditioned on variations of the corpus. Our results reveal several limitations of these methods; for example, bi-encoder methods struggle when the query and target segment share limited similarity and likelihood-based methods struggle with modeling what may be unknown information to a user. Overall, these results suggest that Backtracing is a challenging task that requires new retrieval approaches. We hope our benchmark serves to evaluate and improve future retrieval systems for Backtracing, and ultimately, spawns systems that empower content creators to understand user queries, refine their content and provide users with better experiences. ## Citation If you find our work useful or interesting, please consider citing it! ``` @inproceedings{wang2024backtracing, title = {Backtracing: Retrieving the Cause of the Query}, booktitle = {Findings of the Association for Computational Linguistics: EACL 2024}, publisher = {Association for Computational Linguistics}, year = {2024}, author = {Wang, Rose E. and Wirawarn, Pawan and Khattab, Omar and Goodman, Noah and Demszky, Dorottya}, } ``` ## Reproducibility We ran our experiments with Python 3.11 and on A6000 machines. To reproduce the results in our work, please run the following commands: ``` >> conda create -n backtracing python=3.11 >> conda activate backtracing >> pip install -r requirements.txt # install all of our requirements >> source run_table_evaluations.sh ``` `run_table_evaluations.sh` outputs text files under `results/