DE eng

Search in the Catalogues and Directories

Page: 1 2
Hits 1 – 20 of 22

1
Estimating the Entropy of Linguistic Distributions ...
BASE
Show details
2
On Homophony and Rényi Entropy ...
BASE
Show details
3
On Homophony and Rényi Entropy ...
BASE
Show details
4
On Homophony and Rényi Entropy ...
BASE
Show details
5
Searching for Search Errors in Neural Morphological Inflection ...
BASE
Show details
6
Revisiting the Uniform Information Density Hypothesis ...
BASE
Show details
7
Revisiting the Uniform Information Density Hypothesis ...
BASE
Show details
8
Conditional Poisson Stochastic Beams ...
BASE
Show details
9
Language Model Evaluation Beyond Perplexity ...
BASE
Show details
10
A surprisal--duration trade-off across and within the world's languages ...
BASE
Show details
11
Determinantal Beam Search ...
BASE
Show details
12
Is Sparse Attention more Interpretable? ...
BASE
Show details
13
Revisiting the Uniform Information Density Hypothesis ...
BASE
Show details
14
A Plug-and-Play Method for Controlled Text Generation ...
BASE
Show details
15
Language Model Evaluation Beyond Perplexity ...
Meister, Clara Isabel; Cotterell, Ryan. - : ETH Zurich, 2021
BASE
Show details
16
Determinantal Beam Search ...
Abstract: Beam search is a go-to strategy for decoding neural sequence models. The algorithm can naturally be viewed as a subset optimization problem, albeit one where the corresponding set function does not reflect interactions between candidates. Empirically, this leads to sets often exhibiting high overlap, e.g., strings may differ by only a single word. Yet in use-cases that call for multiple solutions, a diverse or representative set is often desired. To address this issue, we propose a reformulation of beam search, which we call determinantal beam search. Determinantal beam search has a natural relationship to determinantal point processes (DPPs), models over sets that inherently encode intra-set interactions. By posing iterations in beam search as a series of subdeterminant maximization problems, we can turn the algorithm into a diverse subset selection process. In a case study, we use the string subsequence kernel to explicitly encourage n-gram coverage in text generated from a sequence model. We observe that ... : Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing ...
URL: https://dx.doi.org/10.3929/ethz-b-000519002
http://hdl.handle.net/20.500.11850/519002
BASE
Hide details
17
Is Sparse Attention more Interpretable? ...
BASE
Show details
18
A Cognitive Regularizer for Language Modeling ...
BASE
Show details
19
A Cognitive Regularizer for Language Modeling ...
BASE
Show details
20
A Cognitive Regularizer for Language Modeling ...
BASE
Show details

Page: 1 2

Catalogues
0
0
0
0
0
0
0
Bibliographies
0
0
0
0
0
0
0
0
0
Linked Open Data catalogues
0
Online resources
0
0
0
0
Open access documents
22
0
0
0
0
© 2013 - 2024 Lin|gu|is|tik | Imprint | Privacy Policy | Datenschutzeinstellungen ändern