-
Notifications
You must be signed in to change notification settings - Fork 14
Description
Thank you for providing such an interesting piece of work. In your paper, it is clearly stated:
- Zero-Shot LLMs for Text-to-SQL
As an alternative, zero-shot Text-to-SQL methods, such as DAIL-SQL (Gao et al., 2024a) and C3 (Dong et al., 2023), leverage the general knowledge encoded in LLMs to generate SQL queries without requiring task-specific fine-tuning, which eliminates the dependence on labeled datasets and computationally intensive training. While this approach offers a practical and cost-effective solution, it faces a fundamental challenge.
I believe that zero-shot learning, as demonstrated in your work, does not rely on any dataset fine-tuning or in-context learning (ICL) with partial data. This is where my question arises: DAIL-SQL employs a sophisticated method for selecting ICL examples, which has been widely acknowledged by other researchers (e.g., XiYan-SQL, OpenSearch-SQL). For this reason, I argue that DAIL-SQL should not be categorized as a zero-shot method, yet it is included in your paper and compared analytically. I am curious about how your paper defines zero-shot learning in this context.
I apologize if I have fallen into any mistaken assumptions—please feel free to correct and critique me directly! >_< Your response would be greatly appreciated.