4 |
Structural Supervision Improves Few-Shot Learning and Syntactic Generalization in Neural Language Models
|
|
|
|
In: Association for Computational Linguistics (2021)
|
|
BASE
|
|
Show details
|
|
5 |
Structural Supervision Improves Learning of Non-Local Grammatical Dependencies
|
|
|
|
In: Association for Computational Linguistics (2021)
|
|
BASE
|
|
Show details
|
|
6 |
Event-Driven News Stream Clustering using Entity-Aware Contextual Embeddings ...
|
|
|
|
BASE
|
|
Show details
|
|
8 |
On the Evolution of Syntactic Information Encoded by BERT's Contextualized Representations ...
|
|
|
|
BASE
|
|
Show details
|
|
9 |
How much pretraining data do language models need to learn syntax? ...
|
|
|
|
BASE
|
|
Show details
|
|
14 |
Universal Dependencies 2.2
|
|
|
|
In: https://hal.archives-ouvertes.fr/hal-01930733 ; 2018 (2018)
|
|
BASE
|
|
Show details
|
|
18 |
Multilingual Neural Machine Translation with Task-Specific Attention ...
|
|
|
|
Abstract:
Multilingual machine translation addresses the task of translating between multiple source and target languages. We propose task-specific attention models, a simple but effective technique for improving the quality of sequence-to-sequence neural multilingual translation. Our approach seeks to retain as much of the parameter sharing generalization of NMT models as possible, while still allowing for language-specific specialization of the attention model to a particular language-pair or task. Our experiments on four languages of the Europarl corpus show that using a target-specific model of attention provides consistent gains in translation quality for all possible translation directions, compared to a model in which all parameters are shared. We observe improved translation quality even in the (extreme) low-resource zero-shot translation directions for which the model never saw explicitly paired parallel data. ... : COLING 2018 ...
|
|
Keyword:
Computation and Language cs.CL; FOS Computer and information sciences
|
|
URL: https://arxiv.org/abs/1806.03280 https://dx.doi.org/10.48550/arxiv.1806.03280
|
|
BASE
|
|
Hide details
|
|
19 |
Scheduled Multi-Task Learning: From Syntax to Translation ...
|
|
|
|
BASE
|
|
Show details
|
|
|
|