Skip to content

Regarding the Discussion on Zero-Shot Learning #3

@qcjySONG

Description

@qcjySONG

Thank you for providing such an interesting piece of work. In your paper, it is clearly stated:


  • Zero-Shot LLMs for Text-to-SQL
    As an alternative, zero-shot Text-to-SQL methods, such as DAIL-SQL (Gao et al., 2024a) and C3 (Dong et al., 2023), leverage the general knowledge encoded in LLMs to generate SQL queries without requiring task-specific fine-tuning, which eliminates the dependence on labeled datasets and computationally intensive training. While this approach offers a practical and cost-effective solution, it faces a fundamental challenge.

I believe that zero-shot learning, as demonstrated in your work, does not rely on any dataset fine-tuning or in-context learning (ICL) with partial data. This is where my question arises: DAIL-SQL employs a sophisticated method for selecting ICL examples, which has been widely acknowledged by other researchers (e.g., XiYan-SQL, OpenSearch-SQL). For this reason, I argue that DAIL-SQL should not be categorized as a zero-shot method, yet it is included in your paper and compared analytically. I am curious about how your paper defines zero-shot learning in this context.

I apologize if I have fallen into any mistaken assumptions—please feel free to correct and critique me directly! >_< Your response would be greatly appreciated.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions