You are my type! Type embeddings for pre-trained language models

Saeed, Mohammed; Papotti, Paolo
EMNLP 2022, Conference on Empirical Methods in Natural Language Processing, 7-11 December 2022, Abu Dhabi, UAE

One reason for the positive impact of Pretrained Language Models (PLMs) in NLP tasks is their ability to encode semantic types, such as ‘European City’ or ‘Woman’. While previous work has analyzed such information in the context of interpretability, it is not clear how to use types to steer the PLM output. For example, in a cloze statement, it is desirable to steer the model to generate a token that satisfies a user-specified type, e.g., predict a date rather than a location. In this work, we introduce Type Embeddings (TEs), an input embedding that promotes desired types in a PLM. Our proposal is to define a type by a small set of word examples. We empirically study the ability of TEs both in representing types and in steering masking predictions without changes to the
prompt text in BERT. Finally, using the LAMA datasets, we show how TEs highly improve the precision in extracting facts from PLMs.

Type:
Conférence
City:
Abu Dhabi
Date:
2022-12-07
Department:
Data Science
Eurecom Ref:
7095
Copyright:
Copyright ACL. Personal use of this material is permitted. The definitive version of this paper was published in EMNLP 2022, Conference on Empirical Methods in Natural Language Processing, 7-11 December 2022, Abu Dhabi, UAE and is available at :

PERMALINK : https://www.eurecom.fr/publication/7095