Back to Main Conference 2024
LREC-COLING 2024main

TP-Link: Fine-grained Pre-Training for Text-to-SQL Parsing with Linking Information

Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)

DOI:10.63317/3rx2tngto4r9

Abstract

In this paper, we introduce an innovative pre-training framework TP-Link, which aims to improve context-dependent Text-to-SQL Parsing by leveraging Linking information. This enhancement is achieved through better representation of both natural language utterances and the database schema, ultimately facilitating more effective text-to-SQL conversations. We present two novel pre-training objectives: (i) utterance linking prediction (ULP) task that models intricate syntactic relationships among natural language utterances in context-dependent text-to-SQL scenarios, and (ii) schema linking prediction (SLP) task that focuses on capturing fine-grained schema linking relationships between the utterances and the database schema. Extensive experiments demonstrate that our proposed TP-Link achieves state-of-the-art performance on two leading downstream benchmarks (i.e., SParC and CoSQL).

Details

Paper ID
lrec2024-main-1451
Pages
pp. 16686-16697
BibKey
liu-etal-2024-tp
Editor
N/A
Publisher
European Language Resources Association (ELRA) and ICCL
ISSN
2522-2686
ISBN
979-10-95546-34-4
Conference
Joint International Conference on Computational Linguistics, Language Resources and Evaluation
Location
Turin, Italy
Date
20 May 2024 25 May 2024

Authors

  • ZL

    Ziqiang Liu

  • SL

    Shujie Li

  • ZC

    Zefeng Cai

  • XL

    Xiangyu Li

  • YL

    Yunshui Li

  • CL

    Chengming Li

  • XH

    Xiping Hu

  • RX

    Ruifeng Xu

  • MY

    Min Yang

Links