BIAS-3D Brain inspired attentional search model fashioned after what and where/how pathways for target search in 3D environment

Show simple item record

dc.contributor.author Kumari, Sweta
dc.contributor.author Shobha Amala, V.Y.
dc.contributor.author Nivethithan, M.
dc.contributor.author Chakravarthy, V. Srinivasa
dc.date.accessioned 2023-04-18T05:20:10Z
dc.date.available 2023-04-18T05:20:10Z
dc.date.issued 2022-10
dc.identifier.issn 16625188
dc.identifier.uri http://localhost:8080/xmlui/handle/123456789/2059
dc.description This paper is submitted by the author of IIT (BHU), Varanasi, India en_US
dc.description.abstract We propose a brain inspired attentional search model for target search in a 3D environment, which has two separate channels—one for the object classification, analogous to the “what” pathway in the human visual system, and the other for prediction of the next location of the camera, analogous to the “where” pathway. To evaluate the proposed model, we generated 3D Cluttered Cube datasets that consist of an image on one vertical face, and clutter or background images on the other faces. The camera goes around each cube on a circular orbit and determines the identity of the image pasted on the face. The images pasted on the cube faces were drawn from: MNIST handwriting digit, QuickDraw, and RGB MNIST handwriting digit datasets. The attentional input of three concentric cropped windows resembling the high-resolution central fovea and low-resolution periphery of the retina, flows through a Classifier Network and a Camera Motion Network. The Classifier Network classifies the current view into one of the target classes or the clutter. The Camera Motion Network predicts the camera's next position on the orbit (varying the azimuthal angle or “θ”). Here the camera performs one of three actions: move right, move left, or do not move. The Camera-Position Network adds the camera's current position (θ) into the higher features level of the Classifier Network and the Camera Motion Network. The Camera Motion Network is trained using Q-learning where the reward is 1 if the classifier network gives the correct classification, otherwise 0. Total loss is computed by adding the mean square loss of temporal difference and cross entropy loss. Then the model is trained end-to-end by backpropagating the total loss using Adam optimizer. Results on two grayscale image datasets and one RGB image dataset show that the proposed model is successfully able to discover the desired search pattern to find the target face on the cube, and also classify the target face accurately. en_US
dc.description.sponsorship Pavan Holla and Vigneswaran en_US
dc.language.iso en_US en_US
dc.publisher Frontiers Media S.A. en_US
dc.relation.ispartofseries Frontiers in Computational Neuroscience;Volume 16
dc.subject Attention en_US
dc.subject Convolutional neural network en_US
dc.subject flip-flop neurons en_US
dc.subject Human visual system en_US
dc.subject Memory en_US
dc.subject Search in 3D en_US
dc.subject What and where pathway en_US
dc.title BIAS-3D Brain inspired attentional search model fashioned after what and where/how pathways for target search in 3D environment en_US
dc.type Article en_US


Files in this item

This item appears in the following Collection(s)

Show simple item record

Search in IDR


Advanced Search

Browse

My Account