4 |
HerBERT: Efficiently Pretrained Transformer-based Language Model for Polish ...
|
|
|
|
Abstract:
BERT-based models are currently used for solving nearly all Natural Language Processing (NLP) tasks and most often achieve state-of-the-art results. Therefore, the NLP community conducts extensive research on understanding these models, but above all on designing effective and efficient training procedures. Several ablation studies investigating how to train BERT-like models have been carried out, but the vast majority of them concerned only the English language. A training procedure designed for English does not have to be universal and applicable to other especially typologically different languages. Therefore, this paper presents the first ablation study focused on Polish, which, unlike the isolating English language, is a fusional language. We design and thoroughly evaluate a pretraining procedure of transferring knowledge from multilingual to monolingual BERT-based models. In addition to multilingual model initialization, other factors that possibly influence pretraining are also explored, i.e. training ... : Published in Proceedings of the 8th Workshop on Balto-Slavic Natural Language Processing ...
|
|
Keyword:
Computation and Language cs.CL; FOS Computer and information sciences; Machine Learning cs.LG
|
|
URL: https://arxiv.org/abs/2105.01735 https://dx.doi.org/10.48550/arxiv.2105.01735
|
|
BASE
|
|
Hide details
|
|
9 |
Semi-Supervised Neural System for Tagging, Parsing and Lematization ...
|
|
|
|
BASE
|
|
Show details
|
|
13 |
Overview of the SPMRL 2013 Shared Task: A Cross-Framework Evaluation of Parsing Morphologically Rich Languages
|
|
|
|
In: Proceedings of the Fourth Workshop on Statistical Parsing of Morphologically-Rich Languages ; https://hal.archives-ouvertes.fr/hal-00877096 ; Proceedings of the Fourth Workshop on Statistical Parsing of Morphologically-Rich Languages, Oct 2013, Seattle, Washington, United States. pp.146--182 (2013)
|
|
BASE
|
|
Show details
|
|
|
|