keyboard_arrow_up
TQWERE: Transformer-Based SQL Query Executor

Authors

Nir Regev, Asaf Shabtai and Lior Rokach, Ben-Gurion University of the Negev Be’er Sheva, Israel

Abstract

Recent developments in large language models (LLMs) trained on large-scale unstructured textual data have produced high-achieving models. However, it remains a challenge to train an LLM on vast structured (tabular) data for the task of understanding the information captured in the data and answering questions regarding the data. We propose a novel method - TQwerE, for approximating SQL aggregated queries' results over large data sets. Our main focus was to reduce query latency and incurred costs. Moreover, since we focus on large data sets, majority of models that scan raw data are not applicable. Instead, our method fine tunes Jurassic-2 to learn the relations between aggregated SQL queries and their results without referring directly to the underlying raw data. We demonstrate TQwerE's ability to approximate aggregated queries with state-of-the-art accuracy and speed. We evaluated TQwerE on twelve datasets, and our results demonstrated its superiority to both the state-of-the-art methods.

Keywords

LMM, SQL, AQ, LSTM

Full Text  Volume 15, Number 2