1 |
One model for the learning of language.
|
|
|
|
In: Proceedings of the National Academy of Sciences of the United States of America, vol 119, iss 5 (2022)
|
|
BASE
|
|
Show details
|
|
2 |
One model for the learning of language
|
|
|
|
In: Proc Natl Acad Sci U S A (2022)
|
|
Abstract:
A major goal of linguistics and cognitive science is to understand what class of learning systems can acquire natural language. Until recently, the computational requirements of language have been used to argue that learning is impossible without a highly constrained hypothesis space. Here, we describe a learning system that is maximally unconstrained, operating over the space of all computations, and is able to acquire many of the key structures present in natural language from positive evidence alone. We demonstrate this by providing the same learning model with data from 74 distinct formal languages which have been argued to capture key features of language, have been studied in experimental work, or come from an interesting complexity class. The model is able to successfully induce the latent system generating the observed strings from small amounts of evidence in almost all cases, including for regular (e.g., a(n), [Formula: see text] , and [Formula: see text]), context-free (e.g., [Formula: see text] , and [Formula: see text]), and context-sensitive (e.g., [Formula: see text] , and xx) languages, as well as for many languages studied in learning experiments. These results show that relatively small amounts of positive evidence can support learning of rich classes of generative computations over structures. The model provides an idealized learning setup upon which additional cognitive constraints and biases can be formalized.
|
|
Keyword:
Social Sciences
|
|
URL: https://doi.org/10.1073/pnas.2021865119 http://www.ncbi.nlm.nih.gov/pubmed/35074868 http://www.ncbi.nlm.nih.gov/pmc/articles/PMC8812683/
|
|
BASE
|
|
Hide details
|
|
5 |
The Natural Stories corpus: a reading-time corpus of English texts containing rare syntactic constructions
|
|
|
|
In: Springer Netherlands (2020)
|
|
BASE
|
|
Show details
|
|
6 |
Recursive sequence generation in monkeys, children, U.S. adults, and native Amazonians
|
|
|
|
In: Sci Adv (2020)
|
|
BASE
|
|
Show details
|
|
7 |
Post Hoc Analysis Decisions Drive the Reported Reading Time Effects in Hackl, Koster-Hale & Varvoutis (2012)
|
|
|
|
In: Other repository (2019)
|
|
BASE
|
|
Show details
|
|
8 |
Table of assumptions used in our estimates from Humans store about 1.5 megabytes of information during language acquisition ...
|
|
|
|
BASE
|
|
Show details
|
|
9 |
Table of assumptions used in our estimates from Humans store about 1.5 megabytes of information during language acquisition ...
|
|
|
|
BASE
|
|
Show details
|
|
10 |
Supplementary material from "Humans store about 1.5 megabytes of information during language acquisition" ...
|
|
|
|
BASE
|
|
Show details
|
|
11 |
Supplementary material from "Humans store about 1.5 megabytes of information during language acquisition" ...
|
|
|
|
BASE
|
|
Show details
|
|
13 |
Word Forms Are Structured for Efficient Use
|
|
|
|
In: Prof. Gibson via Courtney Crummett (2018)
|
|
BASE
|
|
Show details
|
|
14 |
Color naming across languages reflects color use
|
|
|
|
In: National Academy of Sciences (2018)
|
|
BASE
|
|
Show details
|
|
15 |
Words cluster phonetically beyond phonotactic regularities
|
|
|
|
In: Prof. Gibson via Courtney Crummett (2017)
|
|
BASE
|
|
Show details
|
|
18 |
Wordform Similarity Increases With Semantic Similarity: An Analysis of 100 Languages
|
|
|
|
In: Prof. Gibson via Courtney Crummett (2016)
|
|
BASE
|
|
Show details
|
|
20 |
A Corpus Investigation of Syntactic Embedding in Piraha
|
|
|
|
In: PLoS (2015)
|
|
BASE
|
|
Show details
|
|
|
|