Content area
The disambiguation of word senses (Word Sense Disambiguation, WSD) plays a crucial role in various natural language processing (NLP) tasks, such as machine translation, sentiment analysis, and information retrieval. Due to the complex morphological structure and polysemy of the Korean language, the meaning of words can change depending on the context, making the WSD problem challenging. Since a single word can have multiple meanings, accurately distinguishing between them is essential for improving the performance of NLP models. Recently, large-scale pre-trained models like BERT and GPT, based on transfer learning, have shown promising results in addressing this issue. However, for languages with complex morphological structures, like Korean, the tokenization mismatch between pre-trained models and fine-tuning data prevents the rich contextual and lexical information learned by the pre-trained models from being fully utilized in downstream tasks. This paper proposes a novel method to address the tokenization mismatch issue during the fine-tuning of Korean WSD, leveraging BERT-based pre-trained models and the Sejong corpus, which has been annotated by language experts. Experimental results using various BERT-based pre-trained models and datasets from the Sejong corpus demonstrate that the proposed method improves performance by approximately 3–5% compared to existing approaches.
Details
Word sense disambiguation;
Datasets;
Information retrieval;
Words (language);
Context;
Corpus linguistics;
Machine translation;
Polysemy;
Machine learning;
Performance enhancement;
Sentiment analysis;
Neural networks;
Natural language processing;
Semantic analysis;
Multilingualism;
Algorithms;
Morphological complexity;
Annotations;
Korean language;
Morphology;
Semantics;
Meaning;
Data mining;
Retrieval;
Language shift;
Languages
