1 |
Probing Classifiers: Promises, Shortcomings, and Advances ...
|
|
|
|
BASE
|
|
Show details
|
|
2 |
On the Pitfalls of Analyzing Individual Neurons in Language Models ...
|
|
|
|
BASE
|
|
Show details
|
|
3 |
Debiasing Methods in Natural Language Understanding Make Bias More Accessible ...
|
|
|
|
BASE
|
|
Show details
|
|
4 |
Causal Analysis of Syntactic Agreement Mechanisms in Neural Language Models ...
|
|
|
|
BASE
|
|
Show details
|
|
5 |
Similarity Analysis of Contextual Word Representation Models ...
|
|
|
|
BASE
|
|
Show details
|
|
6 |
Probing the Probing Paradigm: Does Probing Accuracy Entail Task Relevance? ...
|
|
|
|
BASE
|
|
Show details
|
|
7 |
The Sensitivity of Language Models and Humans to Winograd Schema Perturbations ...
|
|
|
|
BASE
|
|
Show details
|
|
8 |
Analyzing Individual Neurons in Pre-trained Language Models ...
|
|
|
|
BASE
|
|
Show details
|
|
9 |
On the Linguistic Representational Power of Neural Machine Translation Models
|
|
|
|
In: Computational Linguistics, Vol 46, Iss 1, Pp 1-52 (2020) (2020)
|
|
BASE
|
|
Show details
|
|
11 |
Exploring Compositional Architectures and Word Vector Representations for Prepositional Phrase Attachment
|
|
|
|
In: MIT Press (2019)
|
|
BASE
|
|
Show details
|
|
12 |
On the Linguistic Representational Power of Neural Machine Translation Models ...
|
|
|
|
Abstract:
Despite the recent success of deep neural networks in natural language processing (NLP), their interpretability remains a challenge. We analyze the representations learned by neural machine translation models at various levels of granularity and evaluate their quality through relevant extrinsic properties. In particular, we seek answers to the following questions: (i) How accurately is word-structure captured within the learned representations, an important aspect in translating morphologically-rich languages? (ii) Do the representations capture long-range dependencies, and effectively handle syntactically divergent languages? (iii) Do the representations capture lexical semantics? We conduct a thorough investigation along several parameters: (i) Which layers in the architecture capture each of these linguistic phenomena; (ii) How does the choice of translation unit (word, character, or subword unit) impact the linguistic properties captured by the underlying representations? (iii) Do the encoder and decoder ... : Accepted to appear in the Journal of Computational Linguistics ...
|
|
Keyword:
Computation and Language cs.CL; FOS Computer and information sciences
|
|
URL: https://dx.doi.org/10.48550/arxiv.1911.00317 https://arxiv.org/abs/1911.00317
|
|
BASE
|
|
Hide details
|
|
13 |
On Adversarial Removal of Hypothesis-only Bias in Natural Language Inference ...
|
|
|
|
BASE
|
|
Show details
|
|
14 |
Improving Neural Language Models by Segmenting, Attending, and Predicting the Future ...
|
|
|
|
BASE
|
|
Show details
|
|
15 |
On Adversarial Removal of Hypothesis-only Bias in Natural Language Inference
|
|
|
|
BASE
|
|
Show details
|
|
16 |
On Evaluating the Generalization of LSTM Models in Formal Languages
|
|
|
|
BASE
|
|
Show details
|
|
17 |
Don't Take the Premise for Granted: Mitigating Artifacts in Natural Language Inference
|
|
|
|
BASE
|
|
Show details
|
|
19 |
On Evaluating the Generalization of LSTM Models in Formal Languages
|
|
|
|
In: Proceedings of the Society for Computation in Linguistics (2019)
|
|
BASE
|
|
Show details
|
|
20 |
Analysis Methods in Neural Language Processing: A Survey
|
|
|
|
In: Transactions of the Association for Computational Linguistics, Vol 7, Pp 49-72 (2019) (2019)
|
|
BASE
|
|
Show details
|
|
|
|