The use of written symbols is a major achievement of human cultural evolution. However, how abstract letter representations might be learned from vision is still an unsolved problem. Here, we present a large-scale computational model of letter recognition based on deep neural networks, which develops a hierarchy of increasingly more complex internal representations in a completely unsupervised way by fitting a probabilistic, generative model to the visual input. In line with the hypothesis that learning written symbols partially recycles pre-existing neuronal circuits for object recognition, earlier processing levels in the model exploit domain-general visual features learned from natural images, while domain-specific features emerge in upstream neurons following exposure to printed letters. We show that these high-level representations can be easily mapped to letter identities even for noise-degraded images, producing accurate simulations of a broad range of empirical findings on letter perception in human observers. Our model shows that by reusing natural visual primitives, learning written symbols only requires limited, domain-specific tuning, supporting the hypothesis that their shape has been culturally selected to match the statistical structure of natural environments.
Letter perception emerges from unsupervised deep learning and recycling of natural image features
Tipo Pubblicazione:
Articolo
Publisher:
Springer Nature
Source:
Nature human behaviour Online 1 (2017): 657–664. doi:10.1038/s41562-017-0186-2
info:cnr-pdr/source/autori:Testolin, Alberto; Stoianov, Ivilin; Zorzi, Marco/titolo:Letter perception emerges from unsupervised deep learning and recycling of natural image features/doi:10.1038/s41562-017-0186-2/rivista:Nature human behaviour Online/anno:
Date:
2017
Resource Identifier:
http://www.cnr.it/prodotto/i/432197
https://dx.doi.org/10.1038/s41562-017-0186-2
info:doi:10.1038/s41562-017-0186-2
https://www.nature.com/articles/s41562-017-0186-2
Language:
Eng