DE eng

Search in the Catalogues and Directories

Hits 1 – 16 of 16

1
Winoground: Probing Vision and Language Models for Visio-Linguistic Compositionality ...
BASE
Show details
2
ANLIzing the Adversarial Natural Language Inference Dataset
In: Proceedings of the Society for Computation in Linguistics (2022)
BASE
Show details
3
Learning from the Worst: Dynamically Generated Datasets to Improve Online Hate Detection ...
BASE
Show details
4
I like fish, especially dolphins: Addressing Contradictions in Dialogue Modeling ...
BASE
Show details
5
Improving Question Answering Model Robustness with Synthetic Adversarial Data Generation ...
BASE
Show details
6
Reservoir Transformers ...
BASE
Show details
7
Gradient-based Adversarial Attacks against Text Transformers ...
BASE
Show details
8
DynaSent: A Dynamic Benchmark for Sentiment Analysis ...
BASE
Show details
9
On the Efficacy of Adversarial Data Collection for Question Answering: Results from a Large-Scale Randomized Study ...
BASE
Show details
10
Masked Language Modeling and the Distributional Hypothesis: Order Word Matters Pre-training for Little ...
BASE
Show details
11
Deep Artificial Neural Networks Reveal a Distributed Cortical Network Encoding Propositional Sentence-Level Meaning
In: J Neurosci (2021)
BASE
Show details
12
Inferring concept hierarchies from text corpora via hyperbolic embeddings ...
BASE
Show details
13
Inferring concept hierarchies from text corpora via hyperbolic embeddings
In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (ACL 2019) (2019)
BASE
Show details
14
Visually Grounded and Textual Semantic Models Differentially Decode Brain Activity Associated with Concrete and Abstract Nouns ...
Anderson, AJ; Kiela, Douwe; Clark, Stephen; Poesio, M. - : Apollo - University of Cambridge Repository, 2017
Abstract: Important advances have recently been made using computational semantic models to decode brain activity patterns associated with concepts; however, this work has almost exclusively focused on concrete nouns. How well these models extend to decoding abstract nouns is largely unknown. We address this question by applying state-of-the-art computational models to decode functional Magnetic Resonance Imaging (fMRI) activity patterns, elicited by participants reading and imagining a diverse set of both concrete and abstract nouns. One of the models we use is linguistic, exploiting the recent word2vec skipgram approach trained on Wikipedia. The second is visually grounded, using deep convolutional neural networks trained on Google Images. Dual coding theory considers concrete concepts to be encoded in the brain both linguistically and visually, and abstract concepts only linguistically. Splitting the fMRI data according to human concreteness ratings, we indeed observe that both models significantly decode the most ... : Stephen Clark is supported by ERC Starting Grant DisCoTex (306920). ...
URL: https://www.repository.cam.ac.uk/handle/1810/263633
https://dx.doi.org/10.17863/cam.8991
BASE
Hide details
15
Adaptive Communication: Languages with More Non-Native Speakers Tend to Have Fewer Word Forms.
Bentz, Christian; Verkerk, Annemarie; Kiela, Douwe. - : Public Library of Science (PLoS), 2015. : PLoS One, 2015
BASE
Show details
16
Adaptive Communication: Languages with More Non-Native Speakers Tend to Have Fewer Word Forms
Bentz, Christian; Verkerk, Annemarie; Kiela, Douwe. - : Public Library of Science, 2015
BASE
Show details

Catalogues
0
0
0
0
0
0
0
Bibliographies
0
0
0
0
0
0
0
0
0
Linked Open Data catalogues
0
Online resources
0
0
0
0
Open access documents
16
0
0
0
0
© 2013 - 2024 Lin|gu|is|tik | Imprint | Privacy Policy | Datenschutzeinstellungen ändern