DE eng

Search in the Catalogues and Directories

Hits 1 – 10 of 10

1
DeepNet: Scaling Transformers to 1,000 Layers ...
Wang, Hongyu; Ma, Shuming; Dong, Li. - : arXiv, 2022
BASE
Show details
2
On the Representation Collapse of Sparse Mixture of Experts ...
Chi, Zewen; Dong, Li; Huang, Shaohan. - : arXiv, 2022
BASE
Show details
3
Allocating Large Vocabulary Capacity for Cross-lingual Language Model Pre-training ...
Zheng, Bo; Dong, Li; Huang, Shaohan. - : arXiv, 2021
BASE
Show details
4
Improving Pretrained Cross-Lingual Language Models via Self-Labeled Word Alignment ...
Chi, Zewen; Dong, Li; Zheng, Bo. - : arXiv, 2021
BASE
Show details
5
Multilingual Machine Translation Systems from Microsoft for WMT21 Shared Task ...
BASE
Show details
6
DeltaLM: Encoder-Decoder Pre-training for Language Generation and Translation by Augmenting Pretrained Multilingual Encoders ...
Ma, Shuming; Dong, Li; Huang, Shaohan. - : arXiv, 2021
BASE
Show details
7
XLM-E: Cross-lingual Language Model Pre-training via ELECTRA ...
Chi, Zewen; Huang, Shaohan; Dong, Li. - : arXiv, 2021
BASE
Show details
8
Adapt-and-Distill: Developing Small, Fast and Effective Pretrained Language Models for Domains ...
BASE
Show details
9
MiniLMv2: Multi-Head Self-Attention Relation Distillation for Compressing Pretrained Transformers ...
BASE
Show details
10
Text Morphing ...
Huang, Shaohan; Wu, Yu; Wei, Furu. - : arXiv, 2018
BASE
Show details

Catalogues
0
0
0
0
0
0
0
Bibliographies
0
0
0
0
0
0
0
0
0
Linked Open Data catalogues
0
Online resources
0
0
0
0
Open access documents
10
0
0
0
0
© 2013 - 2024 Lin|gu|is|tik | Imprint | Privacy Policy | Datenschutzeinstellungen ändern