1 |
IGLUE: A Benchmark for Transfer Learning across Modalities, Tasks, and Languages ...
|
|
|
|
BASE
|
|
Show details
|
|
5 |
Modelling Latent Translations for Cross-Lingual Transfer ...
|
|
|
|
BASE
|
|
Show details
|
|
6 |
Minimax and Neyman–Pearson Meta-Learning for Outlier Languages ...
|
|
|
|
BASE
|
|
Show details
|
|
7 |
Mind the Context: The Impact of Contextualization in Neural Module Networks for Grounding Visual Referring Expressions ...
|
|
|
|
BASE
|
|
Show details
|
|
8 |
Back-Training excels Self-Training at Unsupervised Domain Adaptation of Question Generation and Passage Retrieval ...
|
|
|
|
BASE
|
|
Show details
|
|
9 |
Minimax and Neyman–Pearson Meta-Learning for Outlier Languages ...
|
|
|
|
BASE
|
|
Show details
|
|
10 |
Visually Grounded Reasoning across Languages and Cultures ...
|
|
|
|
BASE
|
|
Show details
|
|
11 |
Visually Grounded Reasoning across Languages and Cultures ...
|
|
|
|
BASE
|
|
Show details
|
|
12 |
Visually Grounded Reasoning across Languages and Cultures ...
|
|
|
|
BASE
|
|
Show details
|
|
15 |
Words aren't enough, their order matters: On the Robustness of Grounding Visual Referring Expressions ...
|
|
|
|
Abstract:
Visual referring expression recognition is a challenging task that requires natural language understanding in the context of an image. We critically examine RefCOCOg, a standard benchmark for this task, using a human study and show that 83.7% of test instances do not require reasoning on linguistic structure, i.e., words are enough to identify the target object, the word order doesn't matter. To measure the true progress of existing models, we split the test set into two sets, one which requires reasoning on linguistic structure and the other which doesn't. Additionally, we create an out-of-distribution dataset Ref-Adv by asking crowdworkers to perturb in-domain examples such that the target object changes. Using these datasets, we empirically show that existing methods fail to exploit linguistic structure and are 12% to 23% lower in performance than the established progress for this task. We also propose two methods, one based on contrastive learning and the other based on multi-task learning, to increase ... : ACL 2020 ...
|
|
Keyword:
Computation and Language cs.CL; Computer Vision and Pattern Recognition cs.CV; FOS Computer and information sciences
|
|
URL: https://arxiv.org/abs/2005.01655 https://dx.doi.org/10.48550/arxiv.2005.01655
|
|
BASE
|
|
Hide details
|
|
19 |
CoQA: A Conversational Question Answering Challenge
|
|
|
|
In: Transactions of the Association for Computational Linguistics, Vol 7, Pp 249-266 (2019) (2019)
|
|
BASE
|
|
Show details
|
|
20 |
Universal Dependencies 2.2
|
|
|
|
In: https://hal.archives-ouvertes.fr/hal-01930733 ; 2018 (2018)
|
|
BASE
|
|
Show details
|
|
|
|