VLDB 2022, 48th International Conference on Very Large Databases, 5-9 September 2022, Sydney, Australia (Hybrid Conference) / Vol.15, N°12
In the last few years, the natural language processing community witnessed advances in neural representations of free texts with transformer-based language models (LMs). Given the importance of knowledge available in relational tables, recent research efforts extend LMs by developing neural representations for tabular data. In this tutorial, we present these proposals with two main goals. First, we introduce to a database audience the potentials and the limitations of current models. Second, we demonstrate the large variety of data applications that benefit from the transformer architecture. The tutorial aims at encouraging database researchers to engage and contribute to this new direction, and at empowering practitioners with a new set of tools for applications involving text and tabular data.
Type:
Tutorial
City:
Sydney
Date:
2022-09-05
Department:
Data Science
Eurecom Ref:
6923
Copyright:
© ACM, 2022. This is the author's version of the work. It is posted here by permission of ACM for your personal use. Not for redistribution. The definitive version was published in VLDB 2022, 48th International Conference on Very Large Databases, 5-9 September 2022, Sydney, Australia (Hybrid Conference) / Vol.15, N°12 https://doi.org/10.14778/3554821.3554890
See also: