Format

Send to

Choose Destination
Cogn Sci. 2017 Apr;41 Suppl 4:855-884. doi: 10.1111/cogs.12360. Epub 2016 Mar 14.

Semantic Coherence Facilitates Distributional Learning.

Author information

1
Department of Psychology, Stanford University.
2
Department of Cognitive Science, University of California San Diego.

Abstract

Computational models have shown that purely statistical knowledge about words' linguistic contexts is sufficient to learn many properties of words, including syntactic and semantic category. For example, models can infer that "postman" and "mailman" are semantically similar because they have quantitatively similar patterns of association with other words (e.g., they both tend to occur with words like "deliver," "truck," "package"). In contrast to these computational results, artificial language learning experiments suggest that distributional statistics alone do not facilitate learning of linguistic categories. However, experiments in this paradigm expose participants to entirely novel words, whereas real language learners encounter input that contains some known words that are semantically organized. In three experiments, we show that (a) the presence of familiar semantic reference points facilitates distributional learning and (b) this effect crucially depends both on the presence of known words and the adherence of these known words to some semantic organization.

KEYWORDS:

Distributional learning; Semantic coherence; Word learning

PMID:
26988338
DOI:
10.1111/cogs.12360
[Indexed for MEDLINE]

Supplemental Content

Full text links

Icon for Wiley
Loading ...
Support Center