1 |
Universal Dependencies and Semantics for English and Hebrew Child-directed Speech
|
|
|
|
In: Proceedings of the Society for Computation in Linguistics (2022)
|
|
BASE
|
|
Show details
|
|
2 |
Do Infants Really Learn Phonetic Categories?
|
|
|
|
In: EISSN: 2470-2986 ; Open Mind ; https://hal.archives-ouvertes.fr/hal-03550830 ; Open Mind, MIT Press, 2021, 5, pp.113-131. ⟨10.1162/opmi_a_00046⟩ (2021)
|
|
BASE
|
|
Show details
|
|
3 |
Early phonetic learning without phonetic categories -- Insights from large-scale simulations on realistic input
|
|
|
|
In: ISSN: 0027-8424 ; EISSN: 1091-6490 ; Proceedings of the National Academy of Sciences of the United States of America ; https://hal.archives-ouvertes.fr/hal-03070566 ; Proceedings of the National Academy of Sciences of the United States of America , National Academy of Sciences, 2021, 118 (7), pp.e2001844118. ⟨10.1073/pnas.2001844118⟩ (2021)
|
|
BASE
|
|
Show details
|
|
4 |
Black or White but never neutral: How readers perceive identity from yellow or skin-toned emoji ...
|
|
|
|
BASE
|
|
Show details
|
|
6 |
Cross-linguistically Consistent Semantic and Syntactic Annotation of Child-directed Speech ...
|
|
|
|
BASE
|
|
Show details
|
|
7 |
Do Infants Really Learn Phonetic Categories?
|
|
|
|
In: Open Mind (Camb) (2021)
|
|
BASE
|
|
Show details
|
|
8 |
Early phonetic learning without phonetic categories: Insights from large-scale simulations on realistic input
|
|
|
|
In: Proc Natl Acad Sci U S A (2021)
|
|
Abstract:
Before they even speak, infants become attuned to the sounds of the language(s) they hear, processing native phonetic contrasts more easily than nonnative ones. For example, between 6 to 8 mo and 10 to 12 mo, infants learning American English get better at distinguishing English and [l], as in “rock” vs. “lock,” relative to infants learning Japanese. Influential accounts of this early phonetic learning phenomenon initially proposed that infants group sounds into native vowel- and consonant-like phonetic categories—like and [l] in English—through a statistical clustering mechanism dubbed “distributional learning.” The feasibility of this mechanism for learning phonetic categories has been challenged, however. Here, we demonstrate that a distributional learning algorithm operating on naturalistic speech can predict early phonetic learning, as observed in Japanese and American English infants, suggesting that infants might learn through distributional learning after all. We further show, however, that, contrary to the original distributional learning proposal, our model learns units too brief and too fine-grained acoustically to correspond to phonetic categories. This challenges the influential idea that what infants learn are phonetic categories. More broadly, our work introduces a mechanism-driven approach to the study of early phonetic learning, together with a quantitative modeling framework that can handle realistic input. This allows accounts of early phonetic learning to be linked to concrete, systematic predictions regarding infants’ attunement.
|
|
Keyword:
Social Sciences
|
|
URL: https://doi.org/10.1073/pnas.2001844118 http://www.ncbi.nlm.nih.gov/pmc/articles/PMC7924220/ http://www.ncbi.nlm.nih.gov/pubmed/33510040
|
|
BASE
|
|
Hide details
|
|
9 |
Multilingual acoustic word embedding models for processing zero-resource languages ...
|
|
|
|
BASE
|
|
Show details
|
|
10 |
Improved acoustic word embeddings for zero-resource languages using multilingual transfer ...
|
|
|
|
BASE
|
|
Show details
|
|
12 |
Inflecting when there's no majority: Limitations of encoder-decoder neural networks as cognitive models for German plurals ...
|
|
|
|
BASE
|
|
Show details
|
|
13 |
Evaluating computational models of infant phonetic learning across languages ...
|
|
|
|
BASE
|
|
Show details
|
|
14 |
Multilingual and Unsupervised Subword Modeling for Zero-Resource Languages
|
|
|
|
In: http://infoscience.epfl.ch/record/277105 (2020)
|
|
BASE
|
|
Show details
|
|
15 |
On understanding character-level models for representing morphology
|
|
|
|
BASE
|
|
Show details
|
|
16 |
Methods for morphology learning in low(er)-resource scenarios
|
|
|
|
BASE
|
|
Show details
|
|
17 |
Discovering and analysing lexical variation in social media text
|
|
|
|
BASE
|
|
Show details
|
|
18 |
Are we there yet? Encoder-decoder neural networks as cognitive models of English past tense inflection ...
|
|
|
|
BASE
|
|
Show details
|
|
19 |
Analyzing ASR pretraining for low-resource speech-to-text translation ...
|
|
|
|
BASE
|
|
Show details
|
|
|
|