A small, preliminary study helps researchers identify a region of the brain that specializes in the processing of auditory words.
Artificial IntelligenceDeep LearningFeaturedMachine LearningNeuroscienceNeurotechRoboticsVisual Neuroscience··5 min read
MEG neuroimaging implicates the occipital place area (OPA) in our ability to rapidly sense our surroundings. The findings may advance improving machine learning and robotics technology aimed at mimicking visual processes in the human brain.
Using MEG neuroimaging, researchers identify abnormalities in functional activity in brain regions which look structurally normal on conventional MRI scans. The findings could help with early detection of primary progressive aphasia.
Eating processed food which contains Propionic Acid (PPA) during pregnancy may affect neural stem cell development in the fetal brain. Excessive PPA reduces the number of neurons and over-produces glial cells, causing inflammation. Additionally, PPA shortens neural pathways. The combination of damaged pathways and reduced neurons may be associated with behavioral deficits associated with ASD.
Researchers have identified the location of dysfunctional brain networks that lead to impaired sentence production and word-finding in primary progressive aphasia (PPA). PPA can occur in those with neurodegenerative diseases, such as frontotemporal dementia and Alzheimer's disease. Mapping the networks allows clinicians to apply non-invasive brain stimulation to potentially improve speech in those with PPA.
Combining neuroimaging and a constructed virtual reality town, researchers found the brain uses three different systems to perceive environment. One system allows us to recognize a place, another helps navigate through that location, and the third helps navigate from one place to another. The parahippocampal place area (PPA) helps encode location recognition, while the retrosplenial complex allows for mentally mapping the locations of specific places.