Abstract
WordNet is a database that represents relations between words and concepts as an abstraction of the contexts in which words are used. Contextualized language models represent words in contexts but leave the underlying concepts implicit. In this paper, we investigate how different layers of a pre-trained language model shape the abstract lexical relationship toward the actual contextual concept. Can we define the amount of contextualized concept forming needed given the abstracted representation of a word? Specifically, we consider samples of words with different polysemy profiles shared across three languages, assuming that words with a different polysemy profile require a different degree of concept shaping by context. We conduct probing experiments to investigate the impact of prior polysemy profiles on the representation in different layers. We analyze how contextualized models can approximate meaning through context and examine cross-lingual interference effects.
Original language | English |
---|---|
Title of host publication | Proceedings of the 12th Global Wordnet Conference |
Editors | German Rigau, Francis Bond, Alexandre Rademaker |
Publisher | Association for Computational Linguistics (ACL) |
Pages | 14-24 |
Number of pages | 11 |
ISBN (Electronic) | 9781713890881 |
Publication status | Published - 2023 |
Event | 12th Global Wordnet Conference, GWC 2023 - Donositia-San Sebastian, Spain Duration: 23 Jan 2023 → 27 Jan 2023 |
Conference
Conference | 12th Global Wordnet Conference, GWC 2023 |
---|---|
Country/Territory | Spain |
City | Donositia-San Sebastian |
Period | 23/01/23 → 27/01/23 |
Bibliographical note
Publisher Copyright:© 2023 12th Global Wordnet Conference, GWC 2023. All rights reserved.