SOTAVerified

Open-SQL Framework: Enhancing Text-to-SQL on Open-source Large Language Models

2024-05-04Unverified0· sign in to hype

Xiaojun Chen, Tianle Wang, Tianhao Qiu, Jianbin Qin, Min Yang

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Despite the success of large language models (LLMs) in Text-to-SQL tasks, open-source LLMs encounter challenges in contextual understanding and response coherence. To tackle these issues, we present , a systematic methodology tailored for Text-to-SQL with open-source LLMs. Our contributions include a comprehensive evaluation of open-source LLMs in Text-to-SQL tasks, the strategy for effective question representation, and novel strategies for supervised fine-tuning. We explore the benefits of Chain-of-Thought in step-by-step inference and propose the method for enhanced few-shot learning. Additionally, we introduce token-efficient techniques, such as Variable-length Open DB Schema, Target Column Truncation, and Example Column Truncation, addressing challenges in large-scale databases. Our findings emphasize the need for further investigation into the impact of supervised fine-tuning on contextual learning capabilities. Remarkably, our method significantly improved Llama2-7B from 2.54\% to 41.04\% and Code Llama-7B from 14.54\% to 48.24\% on the BIRD-Dev dataset. Notably, the performance of Code Llama-7B surpassed GPT-4 (46.35\%) on the BIRD-Dev dataset.

Tasks

Reproductions