Bayesian learning of visual chunks by human observers

Gergö Orbán, József Fiser, Richard N. Aslin, Máté Lengyel

Research output: Contribution to journalArticle

121 Citations (Scopus)

Abstract

Efficient and versatile processing of any hierarchically structured information requires a learning mechanism that combines lower-level features into higher-level chunks. We investigated this chunking mechanism in humans with a visual pattern-learning paradigm. We developed an ideal learner based on Bayesian model comparison that extracts and stores only those chunks of information that are minimally sufficient to encode a set of visual scenes. Our ideal Bayesian chunk learner not only reproduced the results of a large set of previous empirical findings in the domain of human pattern learning but also made a key prediction that we confirmed experimentally. In accordance with Bayesian learning but contrary to associative learning, human performance was well above chance when pair-wise statistics in the exemplars contained no relevant information. Thus, humans extract chunks from complex visual patterns by generating accurate yet economical representations and not by encoding the full correlational structure of the input.

Original languageEnglish
Pages (from-to)2745-2750
Number of pages6
JournalProceedings of the National Academy of Sciences of the United States of America
Volume105
Issue number7
DOIs
Publication statusPublished - Feb 19 2008

    Fingerprint

Keywords

  • Bayesian inference
  • Probabilistic modeling
  • Vision

ASJC Scopus subject areas

  • General

Cite this