File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: Graphix-T5: Mixing Pre-trained Transformers with Graph-Aware Layers for Text-to-SQL Parsing

TitleGraphix-T5: Mixing Pre-trained Transformers with Graph-Aware Layers for Text-to-SQL Parsing
Authors
Issue Date26-Jun-2023
PublisherAssociation for the Advancement of Artificial Intelligence (AAAI)
Abstract

The task of text-to-SQL parsing, which aims at converting natural language questions into executable SQL queries, has garnered increasing attention in recent years. One of the major challenges in text-to-SQL parsing is domain generalization, i.e., how to generalize well to unseen databases. Recently, the pre-trained text-to-text transformer model, namely T5, though not specialized for text-to-SQL parsing, has achieved state-of-the-art performance on standard benchmarks targeting domain generalization. In this work, we explore ways to further augment the pre-trained T5 model with specialized components for text-to-SQL parsing. Such components are expected to introduce structural inductive bias into text-to-SQL parsers thus improving the model’s capacity on (potentially multi-hop) reasoning, which is critical for generating structure-rich SQLs. To this end, we propose a new architecture GRAPHIX-T5, a mixed model with the standard pre-trained transformer model augmented by specially-designed graph-aware layers. Extensive experiments and analysis demonstrate the effectiveness of GRAPHIX-T5 across four text-to-SQL benchmarks: SPIDER, SYN, REALISTIC and DK. GRAPHIX-T5 surpasses all other T5-based parsers with a significant margin, achieving new state-of-the-art performance. Notably, GRAPHIX-T5-large reaches performance superior to the original T5-large by 5.7% on exact match (EM) accuracy and 6.6% on execution accuracy (EX). This even outperforms the T5-3B by 1.2% on EM and 1.5% on EX


Persistent Identifierhttp://hdl.handle.net/10722/341727
ISSN

 

DC FieldValueLanguage
dc.contributor.authorLi, Jinyang-
dc.contributor.authorHui, Binyuan-
dc.contributor.authorCheng, Reynold-
dc.contributor.authorQin, Bowen-
dc.contributor.authorMa, Chenhao-
dc.contributor.authorHuo, Nan-
dc.contributor.authorHuang, Fei-
dc.contributor.authorDu, Wenyu-
dc.contributor.authorSi, Luo-
dc.contributor.authorLi, Yongbin-
dc.date.accessioned2024-03-20T06:58:36Z-
dc.date.available2024-03-20T06:58:36Z-
dc.date.issued2023-06-26-
dc.identifier.issn2159-5399-
dc.identifier.urihttp://hdl.handle.net/10722/341727-
dc.description.abstract<p>The task of text-to-SQL parsing, which aims at converting natural language questions into executable SQL queries, has garnered increasing attention in recent years. One of the major challenges in text-to-SQL parsing is domain generalization, i.e., how to generalize well to unseen databases. Recently, the pre-trained text-to-text transformer model, namely T5, though not specialized for text-to-SQL parsing, has achieved state-of-the-art performance on standard benchmarks targeting domain generalization. In this work, we explore ways to further augment the pre-trained T5 model with specialized components for text-to-SQL parsing. Such components are expected to introduce structural inductive bias into text-to-SQL parsers thus improving the model’s capacity on (potentially multi-hop) reasoning, which is critical for generating structure-rich SQLs. To this end, we propose a new architecture GRAPHIX-T5, a mixed model with the standard pre-trained transformer model augmented by specially-designed graph-aware layers. Extensive experiments and analysis demonstrate the effectiveness of GRAPHIX-T5 across four text-to-SQL benchmarks: SPIDER, SYN, REALISTIC and DK. GRAPHIX-T5 surpasses all other T5-based parsers with a significant margin, achieving new state-of-the-art performance. Notably, GRAPHIX-T5-large reaches performance superior to the original T5-large by 5.7% on exact match (EM) accuracy and 6.6% on execution accuracy (EX). This even outperforms the T5-3B by 1.2% on EM and 1.5% on EX</p>-
dc.languageeng-
dc.publisherAssociation for the Advancement of Artificial Intelligence (AAAI)-
dc.relation.ispartofProceedings of the AAAI Conference on Artificial Intelligence-
dc.titleGraphix-T5: Mixing Pre-trained Transformers with Graph-Aware Layers for Text-to-SQL Parsing-
dc.typeConference_Paper-
dc.identifier.doi10.1609/aaai.v37i11.26536-
dc.identifier.volume37-
dc.identifier.issnl2159-5399-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats