@inproceedings{tufa-etal-2023-wordnet,
title = "A {W}ord{N}et View on Crosslingual Transformers",
author = "Tufa, Wondimagegnhue and
Beinborn, Lisa and
Vossen, Piek",
editor = "Rigau, German and
Bond, Francis and
Rademaker, Alexandre",
booktitle = "Proceedings of the 12th Global Wordnet Conference",
month = jan,
year = "2023",
address = "University of the Basque Country, Donostia - San Sebastian, Basque Country",
publisher = "Global Wordnet Association",
url = "https://aclanthology.org/2023.gwc-1.2/",
pages = "14--24",
abstract = "WordNet is a database that represents relations between words and concepts as an abstraction of the contexts in which words are used. Contextualized language models represent words in contexts but leave the underlying concepts implicit. In this paper, we investigate how different layers of a pre-trained language model shape the abstract lexical relationship toward the actual contextual concept. Can we define the amount of contextualized concept forming needed given the abstracted representation of a word? Specifically, we consider samples of words with different polysemy profiles shared across three languages, assuming that words with a different polysemy profile require a different degree of concept shaping by context. We conduct probing experiments to investigate the impact of prior polysemy profiles on the representation in different layers. We analyze how contextualized models can approximate meaning through context and examine crosslingual interference effects."
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="tufa-etal-2023-wordnet">
<titleInfo>
<title>A WordNet View on Crosslingual Transformers</title>
</titleInfo>
<name type="personal">
<namePart type="given">Wondimagegnhue</namePart>
<namePart type="family">Tufa</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Lisa</namePart>
<namePart type="family">Beinborn</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Piek</namePart>
<namePart type="family">Vossen</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2023-01</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<relatedItem type="host">
<titleInfo>
<title>Proceedings of the 12th Global Wordnet Conference</title>
</titleInfo>
<name type="personal">
<namePart type="given">German</namePart>
<namePart type="family">Rigau</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Francis</namePart>
<namePart type="family">Bond</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Alexandre</namePart>
<namePart type="family">Rademaker</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<originInfo>
<publisher>Global Wordnet Association</publisher>
<place>
<placeTerm type="text">University of the Basque Country, Donostia - San Sebastian, Basque Country</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">conference publication</genre>
</relatedItem>
<abstract>WordNet is a database that represents relations between words and concepts as an abstraction of the contexts in which words are used. Contextualized language models represent words in contexts but leave the underlying concepts implicit. In this paper, we investigate how different layers of a pre-trained language model shape the abstract lexical relationship toward the actual contextual concept. Can we define the amount of contextualized concept forming needed given the abstracted representation of a word? Specifically, we consider samples of words with different polysemy profiles shared across three languages, assuming that words with a different polysemy profile require a different degree of concept shaping by context. We conduct probing experiments to investigate the impact of prior polysemy profiles on the representation in different layers. We analyze how contextualized models can approximate meaning through context and examine crosslingual interference effects.</abstract>
<identifier type="citekey">tufa-etal-2023-wordnet</identifier>
<location>
<url>https://aclanthology.org/2023.gwc-1.2/</url>
</location>
<part>
<date>2023-01</date>
<extent unit="page">
<start>14</start>
<end>24</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Conference Proceedings
%T A WordNet View on Crosslingual Transformers
%A Tufa, Wondimagegnhue
%A Beinborn, Lisa
%A Vossen, Piek
%Y Rigau, German
%Y Bond, Francis
%Y Rademaker, Alexandre
%S Proceedings of the 12th Global Wordnet Conference
%D 2023
%8 January
%I Global Wordnet Association
%C University of the Basque Country, Donostia - San Sebastian, Basque Country
%F tufa-etal-2023-wordnet
%X WordNet is a database that represents relations between words and concepts as an abstraction of the contexts in which words are used. Contextualized language models represent words in contexts but leave the underlying concepts implicit. In this paper, we investigate how different layers of a pre-trained language model shape the abstract lexical relationship toward the actual contextual concept. Can we define the amount of contextualized concept forming needed given the abstracted representation of a word? Specifically, we consider samples of words with different polysemy profiles shared across three languages, assuming that words with a different polysemy profile require a different degree of concept shaping by context. We conduct probing experiments to investigate the impact of prior polysemy profiles on the representation in different layers. We analyze how contextualized models can approximate meaning through context and examine crosslingual interference effects.
%U https://aclanthology.org/2023.gwc-1.2/
%P 14-24
Markdown (Informal)
[A WordNet View on Crosslingual Transformers](https://aclanthology.org/2023.gwc-1.2/) (Tufa et al., GWC 2023)
ACL
- Wondimagegnhue Tufa, Lisa Beinborn, and Piek Vossen. 2023. A WordNet View on Crosslingual Transformers. In Proceedings of the 12th Global Wordnet Conference, pages 14–24, University of the Basque Country, Donostia - San Sebastian, Basque Country. Global Wordnet Association.