Browse by author
Lookup NU author(s): Professor Christopher Petkov
Full text for this publication is not currently held within this repository. Alternative links are provided below where available.
© 2020 Cognitive Science Society, IncHuman languages all have a grammar, that is, rules that determine how symbols in a language can be combined to create complex meaningful expressions. Despite decades of research, the evolutionary, developmental, cognitive, and computational bases of grammatical abilities are still not fully understood. “Artificial Grammar Learning” (AGL) studies provide important insights into how rules and structured sequences are learned, the relevance of these processes to language in humans, and whether the cognitive systems involved are shared with other animals. AGL tasks can be used to study how human adults, infants, animals, or machines learn artificial grammars of various sorts, consisting of rules defined typically over syllables, sounds, or visual items. In this introduction, we distill some lessons from the nine other papers in this special issue, which review the advances made from this growing body of literature. We provide a critical synthesis, identify the questions that remain open, and recognize the challenges that lie ahead. A key observation across the disciplines is that the limits of human, animal, and machine capabilities have yet to be found. Thus, this interdisciplinary area of research firmly rooted in the cognitive sciences has unearthed exciting new questions and venues for research, along the way fostering impactful collaborations between traditionally disconnected disciplines that are breaking scientific ground.
Author(s): ten Cate C, Gervain J, Levelt CC, Petkov CI, Zuidema W
Publication type: Review
Publication status: Published
Journal: Topics in Cognitive Science
Pages: epub ahead of print
Online publication date: 05/03/2020
Acceptance date: 08/01/2020
ISSN (print): 1756-8757
ISSN (electronic): 1756-8765