1 |
Universal Dependencies and Semantics for English and Hebrew Child-directed Speech
|
|
|
|
In: Proceedings of the Society for Computation in Linguistics (2022)
|
|
Abstract:
While corpora of child speech and child-directed speech (CDS) have enabled major contributions to the study of child language acquisition, semantic annotation for such corpora is still scarce and lacks a uniform standard. We compile two CDS corpora—in English and Hebrew—with syntactic and semantic annotations. We employ a methodology that enforces a cross-linguistically consistent representation, building on recent advances in dependency representation and semantic parsing. Our semi-automatic syntactic annotation follows the Universal Dependencies standard (UD; de Marneffe et al., 2021), adapted to suit the CDS genre. To induce semantic forms, we develop an automatic method for transducing UD structures into sentential logical forms (LFs). The two representations have complementary strengths: UD structures are language-neutral and support direct annotation, whereas LFs are neutral as to the syntax-semantics interface, and transparently encode semantic distinctions.
|
|
Keyword:
child-directed speech; Computational Linguistics; corpus annotation; language acquisition; syntax-semantics interface; Universal Dependencies
|
|
URL: https://scholarworks.umass.edu/scil/vol5/iss1/25 https://scholarworks.umass.edu/cgi/viewcontent.cgi?article=1254&context=scil
|
|
BASE
|
|
Hide details
|
|
2 |
Do Infants Really Learn Phonetic Categories?
|
|
|
|
In: EISSN: 2470-2986 ; Open Mind ; https://hal.archives-ouvertes.fr/hal-03550830 ; Open Mind, MIT Press, 2021, 5, pp.113-131. ⟨10.1162/opmi_a_00046⟩ (2021)
|
|
BASE
|
|
Show details
|
|
3 |
Early phonetic learning without phonetic categories -- Insights from large-scale simulations on realistic input
|
|
|
|
In: ISSN: 0027-8424 ; EISSN: 1091-6490 ; Proceedings of the National Academy of Sciences of the United States of America ; https://hal.archives-ouvertes.fr/hal-03070566 ; Proceedings of the National Academy of Sciences of the United States of America , National Academy of Sciences, 2021, 118 (7), pp.e2001844118. ⟨10.1073/pnas.2001844118⟩ (2021)
|
|
BASE
|
|
Show details
|
|
4 |
Black or White but never neutral: How readers perceive identity from yellow or skin-toned emoji ...
|
|
|
|
BASE
|
|
Show details
|
|
6 |
Cross-linguistically Consistent Semantic and Syntactic Annotation of Child-directed Speech ...
|
|
|
|
BASE
|
|
Show details
|
|
7 |
Do Infants Really Learn Phonetic Categories?
|
|
|
|
In: Open Mind (Camb) (2021)
|
|
BASE
|
|
Show details
|
|
8 |
Early phonetic learning without phonetic categories: Insights from large-scale simulations on realistic input
|
|
|
|
In: Proc Natl Acad Sci U S A (2021)
|
|
BASE
|
|
Show details
|
|
9 |
Multilingual acoustic word embedding models for processing zero-resource languages ...
|
|
|
|
BASE
|
|
Show details
|
|
10 |
Improved acoustic word embeddings for zero-resource languages using multilingual transfer ...
|
|
|
|
BASE
|
|
Show details
|
|
12 |
Inflecting when there's no majority: Limitations of encoder-decoder neural networks as cognitive models for German plurals ...
|
|
|
|
BASE
|
|
Show details
|
|
13 |
Evaluating computational models of infant phonetic learning across languages ...
|
|
|
|
BASE
|
|
Show details
|
|
14 |
Multilingual and Unsupervised Subword Modeling for Zero-Resource Languages
|
|
|
|
In: http://infoscience.epfl.ch/record/277105 (2020)
|
|
BASE
|
|
Show details
|
|
15 |
On understanding character-level models for representing morphology
|
|
|
|
BASE
|
|
Show details
|
|
16 |
Methods for morphology learning in low(er)-resource scenarios
|
|
|
|
BASE
|
|
Show details
|
|
17 |
Discovering and analysing lexical variation in social media text
|
|
|
|
BASE
|
|
Show details
|
|
18 |
Are we there yet? Encoder-decoder neural networks as cognitive models of English past tense inflection ...
|
|
|
|
BASE
|
|
Show details
|
|
19 |
Analyzing ASR pretraining for low-resource speech-to-text translation ...
|
|
|
|
BASE
|
|
Show details
|
|
|
|