How do Transformer-Architecture Models Address Polysemy of Korean Adverbial Postpositions?

Seongmin Mun, Guillaume Desagulier

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

1 Scopus citations

Abstract

Postpositions, which are characterized as multiple form-function associations and thus polysemous, pose a challenge to automatic identification of their usage. Several studies have used contextualized word-embedding models to reveal the functions of Korean postpositions. Despite the superior classification performance of previous studies, the particular reason how these models resolve the polysemy of Korean postpositions is not enough clear. To add more interpretation, for this reason, we devised a classification model by employing two transformer-architecture models-BERT and GPT-2-and introduces a computational simulation that interactively demonstrates how these transformer-architecture models simulate human interpretation of word-level polysemy involving Korean adverbial postpositions -ey, -eyse, and -(u)lo. Results reveal that (i) the BERT model performs better than the GPT-2 model to classify the intended function of postpositions, (ii) there is an inverse relationship between the classification performance and the number of functions that each postposition manifests, (iii) model performance is affected by the corpus size of each function, (iv) the models' performance gradually improves as the epoch proceeds, and (vi) the models are affected by the scarcity of input and/or semantic closeness between the items.

Original languageEnglish
Title of host publicationDeeLIO 2022 - Deep Learning Inside Out
Subtitle of host publication3rd Workshop on Knowledge Extraction and Integration for Deep Learning Architectures, Proceedings of the Workshop
EditorsEneko Agirre, Marianna Apidianaki, Ivan Vulic
PublisherAssociation for Computational Linguistics (ACL)
Pages11-21
Number of pages11
ISBN (Electronic)9781955917322
StatePublished - 2022
EventDeep Learning Inside Out: 3rd Workshop on Knowledge Extraction and Integration for Deep Learning Architectures, DeeLIO 2022 - Virtual, Dublin, Ireland
Duration: 27 May 2022 → …

Publication series

NameDeeLIO 2022 - Deep Learning Inside Out: 3rd Workshop on Knowledge Extraction and Integration for Deep Learning Architectures, Proceedings of the Workshop

Conference

ConferenceDeep Learning Inside Out: 3rd Workshop on Knowledge Extraction and Integration for Deep Learning Architectures, DeeLIO 2022
Country/TerritoryIreland
CityVirtual, Dublin
Period27/05/22 → …

Fingerprint

Dive into the research topics of 'How do Transformer-Architecture Models Address Polysemy of Korean Adverbial Postpositions?'. Together they form a unique fingerprint.

Cite this