DE eng

Search in the Catalogues and Directories

Page: 1 2
Hits 1 – 20 of 27

1
Probing Classifiers: Promises, Shortcomings, and Advances ...
Belinkov, Yonatan. - : arXiv, 2021
BASE
Show details
2
On the Pitfalls of Analyzing Individual Neurons in Language Models ...
Antverg, Omer; Belinkov, Yonatan. - : arXiv, 2021
BASE
Show details
3
Debiasing Methods in Natural Language Understanding Make Bias More Accessible ...
BASE
Show details
4
Causal Analysis of Syntactic Agreement Mechanisms in Neural Language Models ...
BASE
Show details
5
Similarity Analysis of Contextual Word Representation Models ...
BASE
Show details
6
Probing the Probing Paradigm: Does Probing Accuracy Entail Task Relevance? ...
BASE
Show details
7
The Sensitivity of Language Models and Humans to Winograd Schema Perturbations ...
BASE
Show details
8
Analyzing Individual Neurons in Pre-trained Language Models ...
Abstract: While a lot of analysis has been carried to demonstrate linguistic knowledge captured by the representations learned within deep NLP models, very little attention has been paid towards individual neurons.We carry outa neuron-level analysis using core linguistic tasks of predicting morphology, syntax and semantics, on pre-trained language models, with questions like: i) do individual neurons in pre-trained models capture linguistic information? ii) which parts of the network learn more about certain linguistic phenomena? iii) how distributed or focused is the information? and iv) how do various architectures differ in learning these properties? We found small subsets of neurons to predict linguistic tasks, with lower level tasks (such as morphology) localized in fewer neurons, compared to higher level task of predicting syntax. Our study also reveals interesting cross architectural comparisons. For example, we found neurons in XLNet to be more localized and disjoint when predicting properties compared to BERT ... : Accepted in EMNLP 2020 ...
Keyword: Computation and Language cs.CL; FOS Computer and information sciences
URL: https://dx.doi.org/10.48550/arxiv.2010.02695
https://arxiv.org/abs/2010.02695
BASE
Hide details
9
On the Linguistic Representational Power of Neural Machine Translation Models
In: Computational Linguistics, Vol 46, Iss 1, Pp 1-52 (2020) (2020)
BASE
Show details
10
Studying the history of the Arabic language: language technology and a large-scale historical corpus [<Journal>]
Shmidman, Avi [Verfasser]; Romanov, Maxim [Verfasser]; Barrón-Cedeño, Alberto [Verfasser].
DNB Subject Category Language
Show details
11
Exploring Compositional Architectures and Word Vector Representations for Prepositional Phrase Attachment
In: MIT Press (2019)
BASE
Show details
12
On the Linguistic Representational Power of Neural Machine Translation Models ...
BASE
Show details
13
On Adversarial Removal of Hypothesis-only Bias in Natural Language Inference ...
BASE
Show details
14
Improving Neural Language Models by Segmenting, Attending, and Predicting the Future ...
BASE
Show details
15
On Adversarial Removal of Hypothesis-only Bias in Natural Language Inference
BASE
Show details
16
On Evaluating the Generalization of LSTM Models in Formal Languages
Suzgun, Mirac; Belinkov, Yonatan; Shieber, Stuart. - : Society for Computation in Linguistics, 2019
BASE
Show details
17
Don't Take the Premise for Granted: Mitigating Artifacts in Natural Language Inference
Belinkov, Yonatan; Poliak, Adam; Shieber, Stuart. - : Association of Computational Linguistics, 2019
BASE
Show details
18
LSTM Networks Can Perform Dynamic Counting
BASE
Show details
19
On Evaluating the Generalization of LSTM Models in Formal Languages
In: Proceedings of the Society for Computation in Linguistics (2019)
BASE
Show details
20
Analysis Methods in Neural Language Processing: A Survey
In: Transactions of the Association for Computational Linguistics, Vol 7, Pp 49-72 (2019) (2019)
BASE
Show details

Page: 1 2

Catalogues
0
0
0
0
1
0
0
Bibliographies
0
0
0
0
0
0
0
0
0
Linked Open Data catalogues
0
Online resources
0
0
0
0
Open access documents
26
0
0
0
0
© 2013 - 2024 Lin|gu|is|tik | Imprint | Privacy Policy | Datenschutzeinstellungen ändern