Learning in any domain depends on how the data for learning are represented. In the domain of language acquisition, children's representations of the speech they hear determine what generalizations they can draw about their target grammar. But these input representations change over development as a function of children's developing linguistic knowledge, and may be incomplete or inaccurate when children lack the knowledge to parse their input veridically. How does learning succeed in the face of potentially misleading data? We address this issue using the case study of "non-basic" clauses in verb learning. A young infant hearing What did Amy fix? might not recognize that what stands in for the direct object of fix, and might think that fix is occurring without a direct object. We follow a previous proposal that children might filter nonbasic clauses out of the data for learning verb argument structure, but offer a new approach. Instead of assuming that children identify the data to filter in advance, we demonstrate computationally that it is possible for learners to infer a filter on their input without knowing which clauses are nonbasic. We instantiate a learner that considers the possibility that it misparses some of the sentences it hears, and learns to filter out those parsing errors in order to correctly infer transitivity for the majority of 50 frequent verbs in child-directed speech. Our learner offers a novel solution to the problem of learning from immature input representations: Learners may be able to avoid drawing faulty inferences from misleading data by identifying a filter on their input, without knowing in advance what needs to be filtered.
Download full-text PDF |
Source |
---|---|
http://dx.doi.org/10.1111/cogs.13080 | DOI Listing |
Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!