DE eng

Search in the Catalogues and Directories

Page: 1 2
Hits 1 – 20 of 32

1
StableMoE: Stable Routing Strategy for Mixture of Experts ...
Dai, Damai; Dong, Li; Ma, Shuming. - : arXiv, 2022
BASE
Show details
2
DeepNet: Scaling Transformers to 1,000 Layers ...
Wang, Hongyu; Ma, Shuming; Dong, Li. - : arXiv, 2022
BASE
Show details
3
On the Representation Collapse of Sparse Mixture of Experts ...
Chi, Zewen; Dong, Li; Huang, Shaohan. - : arXiv, 2022
BASE
Show details
4
Controllable Natural Language Generation with Contrastive Prefixes ...
Qian, Jing; Dong, Li; Shen, Yelong. - : arXiv, 2022
BASE
Show details
5
Allocating Large Vocabulary Capacity for Cross-lingual Language Model Pre-training ...
Zheng, Bo; Dong, Li; Huang, Shaohan. - : arXiv, 2021
BASE
Show details
6
Improving Pretrained Cross-Lingual Language Models via Self-Labeled Word Alignment ...
Chi, Zewen; Dong, Li; Zheng, Bo. - : arXiv, 2021
BASE
Show details
7
s2s-ft: Fine-Tuning Pretrained Transformer Encoders for Sequence-to-Sequence Learning ...
Bao, Hangbo; Dong, Li; Wang, Wenhui. - : arXiv, 2021
BASE
Show details
8
Zero-shot Cross-lingual Transfer of Neural Machine Translation with Multilingual Pretrained Encoders ...
Chen, Guanhua; Ma, Shuming; Chen, Yun. - : arXiv, 2021
BASE
Show details
9
MT6: Multilingual Pretrained Text-to-Text Transformer with Translation Pairs ...
Chi, Zewen; Dong, Li; Ma, Shuming. - : arXiv, 2021
BASE
Show details
10
Multilingual Machine Translation Systems from Microsoft for WMT21 Shared Task ...
BASE
Show details
11
DeltaLM: Encoder-Decoder Pre-training for Language Generation and Translation by Augmenting Pretrained Multilingual Encoders ...
Ma, Shuming; Dong, Li; Huang, Shaohan. - : arXiv, 2021
BASE
Show details
12
XLM-E: Cross-lingual Language Model Pre-training via ELECTRA ...
Chi, Zewen; Huang, Shaohan; Dong, Li. - : arXiv, 2021
BASE
Show details
13
Adapt-and-Distill: Developing Small, Fast and Effective Pretrained Language Models for Domains ...
BASE
Show details
14
InfoXLM: An Information-Theoretic Framework for Cross-Lingual Language Model Pre-Training ...
Chi, Zewen; Dong, Li; Wei, Furu. - : arXiv, 2020
BASE
Show details
15
MiniLMv2: Multi-Head Self-Attention Relation Distillation for Compressing Pretrained Transformers ...
BASE
Show details
16
Harvesting and Refining Question-Answer Pairs for Unsupervised QA ...
Li, Zhongli; Wang, Wenhui; Dong, Li. - : arXiv, 2020
BASE
Show details
17
XLM-T: Scaling up Multilingual Machine Translation with Pretrained Cross-lingual Transformer Encoders ...
BASE
Show details
18
Can Monolingual Pretrained Models Help Cross-Lingual Classification? ...
Chi, Zewen; Dong, Li; Wei, Furu. - : arXiv, 2019
BASE
Show details
19
Cross-Lingual Natural Language Generation via Pre-Training ...
Chi, Zewen; Dong, Li; Wei, Furu. - : arXiv, 2019
BASE
Show details
20
Learning natural language interfaces with neural models
Dong, Li. - : The University of Edinburgh, 2019
BASE
Show details

Page: 1 2

Catalogues
0
0
1
0
0
0
0
Bibliographies
0
0
0
0
0
0
0
0
0
Linked Open Data catalogues
0
Online resources
0
0
0
0
Open access documents
31
0
0
0
0
© 2013 - 2024 Lin|gu|is|tik | Imprint | Privacy Policy | Datenschutzeinstellungen ändern