Full text

Turn on search term navigation

© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.

Abstract

Recently, self-play fine-tuning (SPIN) has garnered widespread attention as it enables large language models (LLMs) to iteratively enhance their capabilities through simulated interactions with themselves, transforming a weak LLM into a strong one. However, applying SPIN to fine-tune text-to-SQL models presents substantial challenges. Notably, existing frameworks lack clear signal feedback during the training process and fail to adequately capture the implicit schema-linking characteristics between natural language questions and databases. To address these issues, we propose a novel self-play fine-tuning method for text-to-SQL models, termed ExSPIN, which incorporates explicit feedback. Specifically, during fine-tuning, the SQL query execution results predicted by the LLM are fed back into the model’s parameter update process. This feedback allows both the main player and the opponent to more accurately distinguish between negative and positive samples, thereby improving the fine-tuning outcomes. Additionally, we employ in-context learning techniques to provide explicit schema hints, enabling the LLM to better understand the schema-linking between the database and natural language queries during the self-play process. Evaluations on two real-world datasets show that our method significantly outperforms the state-of-the-art approaches.

Details

Title
ExSPIN: Explicit Feedback-Based Self-Play Fine-Tuning for Text-to-SQL Parsing
Author
Liang, Yan 1 ; Su, Jinhang 2 ; Liu, Chuanyi 3 ; Duan, Shaoming 4 ; Zhang, Yuhao 2 ; Li, Jianhang 2 ; Han, Peiyi 3 ; Liu, Ye 5 

 School of Computer Science and Technology, Harbin Institute of Technology, Shenzhen 518055, China; [email protected] (L.Y.); [email protected] (J.S.); [email protected] (Y.Z.); [email protected] (J.L.); Inspur Cloud Information Technology Co., Ltd., Jinan 250101, China 
 School of Computer Science and Technology, Harbin Institute of Technology, Shenzhen 518055, China; [email protected] (L.Y.); [email protected] (J.S.); [email protected] (Y.Z.); [email protected] (J.L.) 
 School of Computer Science and Technology, Harbin Institute of Technology, Shenzhen 518055, China; [email protected] (L.Y.); [email protected] (J.S.); [email protected] (Y.Z.); [email protected] (J.L.); Pengcheng Laboratory, Shenzhen 518000, China; Key Laboratory of Cyberspace and Data Security, Ministry of Emergency Management, Beijing 100010, China 
 Pengcheng Laboratory, Shenzhen 518000, China 
 Guangdong Power Grid Co., Ltd., Guangzhou 510000, China; [email protected] 
First page
235
Publication year
2025
Publication date
2025
Publisher
MDPI AG
e-ISSN
10994300
Source type
Scholarly Journal
Language of publication
English
ProQuest document ID
3181452201
Copyright
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.