ITMO
ru/ ru

ISSN: 1023-5086

ru/

ISSN: 1023-5086

Scientific and technical

Opticheskii Zhurnal

A full-text English translation of the journal is published by Optica Publishing Group under the title “Journal of Optical Technology”

Article submission Подать статью
Больше информации Back

DOI: 10.17586/1023-5086-2024-91-08-99-109

УДК: 612.821

Vadim Glezer’s modules model is a possible basis of vision perception description

For Russian citation (Opticheskii Zhurnal):

Бондарко В.М. Модель модулей В.Д. Глезера — возможная основа для описания зрительного восприятия // Оптический журнал. 2024. Т. 91. № 8. С. 99–109. http://doi.org/10.17586/1023-5086-2024-91-08-99-109

 

Bondarko V.M. Vadim Glezer’s modules model is a possible basis of vision perception description [in Russian] // Opticheskii Zhurnal. 2024. V. 91. № 8. P. 99–109. http://doi.org/10.17586/1023-5086-2024-91-08-99-109

For citation (Journal of Optical Technology):
-
Abstract:

Subject of study. The mechanisms of visual perception were studied. Aim of study is to determine a possibility of using the modules model to describe visual perception. This model was proposed by V.D. Glezer on the basis of obtained electrophysiological data. Method. Psychophysical methods and modeling were used. Experiments were carried out on image recognition and size estimation of spatial intervals and circles in the Delboeuf and Ebbinghaus illusions. The experimental data were fitted by models using full image spectra and a modules model in which the images were decomposed into finite Fourier series. For each image, a module of the optimal size was selected to save the most of the energy in the image. Two combinations of such modules were considered. Main results. It was shown that the optimal module size depended on image. Recognition images errors and their proximity estimations correlate with the distances between images calculated in the feature space as the norm of the difference their spectra, and partially with the distances obtained in one of the variants of the modules model. The modules model adequately approximated the segmentation and size estimation data, which was confirmed by the analysis of the paintings. Thus, the module model can describe the mechanisms of size estimation and segmentation. At the same time, for recognition it is necessary to improve this model: to introduce interaction between modules of different sizes. Therefore, the module model can be considered as a first approximation to the description of visual perception. Practical significance. The modules model can be used to analyze images and create artificial neural networks that provide segmentation and object recognition.

Keywords:

recognition, size estimation, segmentation, modeling, spatial frequency analysis, modules model, optical illusions

Acknowledgements:

the study was supported by the State funding allocated to the Pavlov Institute of Physiology Russian Academy of Sciences (№ 1021062411653-4-3.1.8).

OCIS codes: 330.7326, 330.4060, 330.5510, 330.5370

References:

1. Yavna D.V., Babenko V.V., Gorbenkova O.A. et al. Classification of objects and scenes by a neural network with pretrained input modules to decode spatial texture inhomogeneities // Journal of Optical Technology. 2023. V. 90(1). P. 20–25. https://doi.org/10.1364/ JOT.90.000020
2. Ermachenkova M.K., Malashin R.O., Boyko A.A. Training of neural networks for classification of thermal imaging based on images of the visible spectrum [in Russian] // Journal of Optical Technology. 2023. V. 90(10). P. 590–600. https://doi.org/10.1364/JOT.90.000590
3. Tsytsulin A.K., Bobrovskiy A.I., Morozov A.V. et al. Using convolutional neural networks to automatically select small artificial space objects on optical images of a starry sky // Journal of Optical Technology. 2019. V. 86(10). P. 627–633. https://doi.org/10.1364/ JOT.86.000627
4. Zhukova O.V., Malakhova E.Yu., Shelepin Yu.E. La Gioconda and the indeterminacy of smile recognition by a person and by an artificial neural network // Journal of Optical Technology. 2019. V. 86(11). P. 706–715. https://doi.org/10.1364/JOT.86.000706
5. Lutsiv V.R. Convolutional artificial neural networks of deep learning [in Russian] // Journal of Optical Technology. 2015. T. 82. № 8. P. 499–508. https://doi. org/10.1364/JOT.82.000499
6. Malakhova E.Yu. Information representation space in artificial and biological neural networks // Journal of Optical Technology. 2020. V. 87(10). P. 598–603. https://doi.org/10.1364/JOT.87.000598
7. Malakhova E.Yu. Representation of categories through prototypes formed based on coordinated activity of units in convolutional neural networks // Journal of Optical Technology. 2021. V. 88. № 12. P. 706–709. https://doi.org/10.1364/JOT.88.000706
8. Glezer V.D. Vision and thinking [in Russian]. L.: Nauka, 1985. 300 p.
9. Glezer V.D. Vision and mind: Modeling mental functions. NJ.: Lawrens Erlbaum Ass., 1995. 274 p.
10. Glezer V.D., Tcherbach T.A., Gauzelman V.E., Bondarko V.M. Linear and non-linear properties of simple and complex receptive fields in area 17 of the cat visual cortex: A model of the fields // Biol. Cybern. 1980. V. 37. P. 195–208.
11. Vol I.A. Spatial-frequency model of hyperacuity of the visual system [in Russian] // Sensory systems. 1988. V. 2. № 2. P. 133–138.
12. Kaliteevsky N.A., Semenov V.E., Glezer V.D., Gauselman V.E. Algorithm of invariant image description by the use of a modified Gabor transform // Applied optics. 1994. V. 33. № 23. Р. 5256–5261.
13. Glezer V.D., Yakovlev V.V., Gauselman V.E. Harmonic basis function for spatial coding in the cat striate cortex // Visual Neurosci. 1989. V. 3. P. 351–383.
14. Burt P.J. Fast filter transforms for image processing // Comput. Graph. and Image Proc. 1981. V. 16. P. 20–51.
15. Bondarko V.M., Danilova M.V., Chikhman V.N. Segmentation of visual images: experimental data and modelling // J. Opt. Technol. 2021. V. 88(12). P. 692–699. https://doi.org/10.1364/JOT.88.000692
16. Bondarko V.M., Danilova M.V. Relation of local window size in a model of modules with estimation of the size of visual images and their segmentation // J. Opt. Technol. 2022. V. 89(8). P. 461–468. https://doi.org/10.1364/JOT.89.000461

17. Blakemore G., Campbell F.W. On the existence in human visual system of neurones selectively sensitive to the orientation and size of retinal image // J. Physiology. 1969. V. 203. № 1. P. 237–260.
18. Campbell F.W., Robson J.G. Application of Fourier analyses to the visibility of gratings // J. Physiol. 1968. V. 197. P. 551–566.
19. Vol I.A., Pavlovskaya M.B. Correlation between proximity of Fourier spectra of images and errors in their identification // Human Physiology. 1986. V. 12. P. 400–406.
20. Gervais M. J., Harvey L.O., Roberts Y.O. Identification confusions among letters of the alphabet // J. Exp. Psychol. Hum. Percept. 1984. V. 10. № 5. P. 655–666.
21. Vol I.A., Pavlovskaya M.B., Bondarko V.M. Similarity between Fourier transform of images predicts their experimental confusion // Perception and Psychophysics. 1990. V. 47. № 1. P. 12–21.
22. Patil A., Rane M. Convolutional neural networks: an overview and its applications in pattern recognition // Information and Communication Technology for Intelligent Systems: Proceedings of ICTIS 2020. 2021. V. 1. P. 21–30. https://doi.org/10.1007/978-981-15-7078-0_3
23. Rawat W., Wang Z. Deep convolutional neural networks for image classification: A comprehensive review // Neural computation. 2017. V. 29. № 9. P. 2352–2449. https://doi.org/: 10.1162/neco_a_00990
24. Sharma N., Jain V., Mishra A. An analysis of convolutional neural networks for image classification // Procedia computer science. 2018. V. 132. P. 377–384. https://doi.org/10.1016/j.procs.2018.05.198
25. Titarenko M.A., Malashin R.O. Study of the ability of neural networks to extract and use semantic information when they are trained to reconstruct noisy images // Journal of Optical Technology. 2022. V. 89(2). P. 81–88. https://doi.org/10.1364/JOT.89.000081
26. Alies M.Yu., Antonov E.A., Kalugin A.I., Zaripov M.R. Application of artificial neural networks for the analysis of multispectral images // Journal of Optical Technology. 2021. V. 88(8). P. 441–444. https://doi.org/10.1364/JOT.88.000441
27. Mihaylova M., Stomonyakov V., Vassilev A. Peripheral and central delay in processing high spatial frequencies: reaction time and VEP latency studies // Vision research. 1999. V. 39. № 4. P. 699–705. https://doi.org/10.1016/s0042-6989(98)00165-5
28. Bondarko V.M., Semenov L.A. Size estimation in the Ebbinghaus illusion in adults and children of different ages // Human Physiology. 2004. V. 30. P. 24–30.
29. Weintraub D.J., Schneck M.K. Fragments of Delboeuf and Ebbinghaus illusions: Contour context explorations of misjudged circle size // Perception and Psychophysics. 1986. V. 40 (3). P. 147–158.