Content area

Abstract

Pre-trained Transformers are challenging human performances in many NLP tasks. The massive datasets used for pre-training seem to be the key to their success on existing tasks. In this paper, we explore how a range of pre-trained Natural Language Understanding models perform on definitely unseen sentences provided by classification tasks over a DarkNet corpus. Surprisingly, results show that syntactic and lexical neural networks perform on par with pre-trained Transformers even after fine-tuning. Only after what we call extreme domain adaptation, that is, retraining with the masked language model task on all the novel corpus, pre-trained Transformers reach their standard high results. This suggests that huge pre-training corpora may give Transformers unexpected help since they are exposed to many of the possible sentences.

Details

1009240
Title
The Dark Side of the Language: Pre-trained Transformers in the DarkNet
Publication title
arXiv.org; Ithaca
Publication year
2023
Publication date
Nov 17, 2023
Section
Computer Science
Publisher
Cornell University Library, arXiv.org
Source
arXiv.org
Place of publication
Ithaca
Country of publication
United States
University/institution
Cornell University Library arXiv.org
e-ISSN
2331-8422
Source type
Working Paper
Language of publication
English
Document type
Working Paper
Publication history
 
 
Online publication date
2024-05-07
Milestone dates
2022-01-14 (Submission v1); 2022-02-09 (Submission v2); 2023-11-17 (Submission v3)
Publication history
 
 
   First posting date
07 May 2024
ProQuest document ID
2621110576
Document URL
https://www.proquest.com/working-papers/dark-side-language-pre-trained-transformers/docview/2621110576/se-2?accountid=208611
Full text outside of ProQuest
Copyright
© 2023. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Last updated
2024-10-16
Database
ProQuest One Academic