DE eng

Search in the Catalogues and Directories

Page: 1 2 3 4 5 6 7...52
Hits 41 – 60 of 1.029

41
KOAS: Korean Text Offensiveness Analysis System ...
BASE
Show details
42
Contrastive Code Representation Learning ...
BASE
Show details
43
Does Putting a Linguist in the Loop Improve NLU Data Collection ...
BASE
Show details
44
What are we learning from language? ...
BASE
Show details
45
Machine Translation Decoding beyond Beam Search ...
BASE
Show details
46
Say `YES' to Positivity: Detecting Toxic Language in Workplace Communications ...
BASE
Show details
47
Unsupervised Multi-View Post-OCR Error Correction With Language Models ...
BASE
Show details
48
AttentionRank: Unsupervised Keyphrase Extraction using Self and Cross Attentions ...
BASE
Show details
49
ProtoInfoMax: Prototypical Networks with Mutual Information Maximization for Out-of-Domain Detection ...
BASE
Show details
50
Multi-granularity Textual Adversarial Attack with Behavior Cloning ...
BASE
Show details
51
Automatic Fact-Checking with Document-level Annotations using BERT and Multiple Instance Learning ...
BASE
Show details
52
Towards the Early Detection of Child Predators in Chat Rooms: A BERT-based Approach ...
BASE
Show details
53
TSDAE: Using Transformer-based Sequential Denoising Auto-Encoder for Unsupervised Sentence Embedding Learning ...
BASE
Show details
54
WebSRC: A Dataset for Web-Based Structural Reading Comprehension ...
BASE
Show details
55
Improving Math Word Problems with Pre-trained Knowledge and Hierarchical Reasoning ...
BASE
Show details
56
Semantic Categorization of Social Knowledge for Commonsense Question Answering ...
BASE
Show details
57
Adversarial Examples for Evaluating Math Word Problem Solvers ...
BASE
Show details
58
Pre-train or Annotate? Domain Adaptation with a Constrained Budget ...
Abstract: Anthology paper link: https://aclanthology.org/2021.emnlp-main.409/ Abstract: Recent work has demonstrated that pre-training in-domain language models can boost performance when adapting to a new domain. However, the costs associated with pre-training raise an important question: given a fixed budget, what steps should an NLP practitioner take to maximize performance? In this paper, we study domain adaptation under budget constraints, and approach it as a customer choice problem between data annotation and pre-training. Specifically, we measure the annotation cost of three procedural text datasets and the pre-training cost of three in-domain language models. Then we evaluate the utility of different combinations of pre-training and data annotation under varying budget constraints to assess which combination strategy works best. We find that, for small budgets, spending all funds on annotation leads to the best performance; once the budget becomes large enough, a combination of data annotation and in-domain ...
Keyword: Computational Linguistics; Language Models; Machine Learning; Machine Learning and Data Mining; Natural Language Processing
URL: https://underline.io/lecture/37963-pre-train-or-annotatequestion-domain-adaptation-with-a-constrained-budget
https://dx.doi.org/10.48448/z1gf-n855
BASE
Hide details
59
Corpus-based Open-Domain Event Type Induction ...
BASE
Show details
60
Learning with Different Amounts of Annotation: From Zero to Many Labels ...
BASE
Show details

Page: 1 2 3 4 5 6 7...52

Catalogues
0
0
0
0
0
0
0
Bibliographies
0
0
0
0
0
0
0
0
0
Linked Open Data catalogues
0
Online resources
0
0
0
0
Open access documents
1.029
0
0
0
0
© 2013 - 2024 Lin|gu|is|tik | Imprint | Privacy Policy | Datenschutzeinstellungen ändern