Back to Main Conference 2024
LREC-COLING 2024main

Fusion-in-T5: Unifying Variant Signals for Simple and Effective Document Ranking with Attention Fusion

Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)

DOI:10.63317/3uakpgt28h8c

Abstract

Common document ranking pipelines in search systems are cascade systems that involve multiple ranking layers to integrate different information step-by-step. In this paper, we propose a novel re-ranker Fusion-in-T5 (FiT5), which integrates text matching information, ranking features, and global document information into one single unified model via templated-based input and global attention. Experiments on passage ranking benchmarks MS MARCO and TREC DL show that FiT5, as one single model, significantly improves ranking performance over complex cascade pipelines. Analysis finds that through attention fusion, FiT5 jointly utilizes various forms of ranking information via gradually attending to related documents and ranking features, and improves the detection of subtle nuances. Our code is open-sourced at https://github.com/OpenMatch/FiT5 . Keywords: document ranking, attention, fusion

Details

Paper ID
lrec2024-main-0667
Pages
pp. 7556-7561
BibKey
yu-etal-2024-fusion
Editor
N/A
Publisher
European Language Resources Association (ELRA) and ICCL
ISSN
2522-2686
ISBN
979-10-95546-34-4
Conference
Joint International Conference on Computational Linguistics, Language Resources and Evaluation
Location
Turin, Italy
Date
20 May 2024 25 May 2024

Authors

  • SY

    Shi Yu

  • CF

    Chenghao Fan

  • CX

    Chenyan Xiong

  • DJ

    David Jin

  • ZL

    Zhiyuan Liu

  • ZL

    Zhenghao Liu

Links