An asymmetric coupling between the peripheral visual and olfactory sensory systems allows for enhanced steering response to discrete objects in mosquitos.
The human brain can desensitize background motion and focus on smaller moving objects in the foreground as a result of activity in the middle temporal visual area. However, our ability to pick out smaller objects changes over time. Younger people are better at picking out foreground objects moving, while those over 65 have heightened awareness of objects moving in the background.
People are able to form the correct mental model of puzzles from either visual or haptic experiences alone and are able to predict haptic properties from visual ones. Findings suggest humans segment scenes into objects without explicit boundary cues by using purely statistical information.
Optical illusions are helping researchers better understand attention and visual perception. Findings suggest attention operates periodically on the perceptual binding of visual information.
Results from a new artificial intelligence study indicate number sense is spontaneously created by the visual system, without prior experience of counting.
Current deep learning models are able to create images strongly enough to activate specific neurons in the visual cortex. However, researchers say more accurate artificial neural network models should be developed to help produce more accurate control.
CT1 cells connect around 1400 areas in the fly brain. Each cell area works like a separate neuron, allowing CT1 to access information from the fly's eye and support local motion detection.
Researchers have developed a new technique that helps identify the functional properties of individual synapses that link the two hemispheres of the brain.