Neuroscience

Articles and news from the latest research reports.

325 notes

Don’t Underestimate Your Mind’s Eye
Take a look around, and what do you see? Much more than you think you do, thanks to your finely tuned mind’s eye, which processes images without your even knowing.
A University of Arizona study has found that objects in our visual field of which we are not consciously aware still may influence our decisions. The findings refute traditional ideas about visual perception and cognition, and they could shed light on why we sometimes make decisions — stepping into a street, choosing not to merge into a traffic lane — without really knowing why.
Laura Cacciamani, who recently earned her doctorate in psychology with a minor in neuroscience, has found supporting evidence. Cacciamani’s is the lead author on a co-authored study, published online in the journal Attention, Perception and Psychophysics, shows that the brain’s subconscious processing has an impact on behavior and decision-making.
This seems to make evolutionary sense, Cacciamani said. Early humans would have required keen awareness of their surroundings on a subliminal level in order to survive.
"Your brain is always monitoring for meaning in the world, to be aware of your general surroundings and potential predators," Cacciamani said. "You can be focused on a task, but your brain is assessing the meaning of everything around you – even objects that you’re not consciously perceiving."
The study builds on the findings of earlier research by Jay Sanguinetti, who also was a doctoral candidate in the UA Department of Psychology. Both studies go against conventional wisdom among vision scientists.
"According to the traditional view, the brain accesses the meaning – or the memory – of an object after you perceive it," Cacciamani said. "Against this view, we have now shown that the meaning of an object can be accessed before conscious perception.
"We’re showing that there’s more interplay between memory and perception than previously has been assumed," she said.
Cacciamani asked participants in her study to classify nouns that appeared on a computer screen as naming a natural object or artificial object by pressing one of two buttons labeled “natural” or “artificial.” For example, the word “leaf” indicates an object found in nature, while “anchor” indicates a man-made or artificial object.
But before each word appeared on the screen, the computer flashed a black silhouette that – unknown to participants – had portions of natural or artificial objects suggested along the white outside regions (called the “ground” regions) of the image. Participants were not told to look for anything in the silhouettes, and they were flashed so quickly – 50 milliseconds – that it would have been difficult to notice the objects in the ground regions even if someone knew what to look for. Participants never were aware that the silhouette’s grounds suggested recognizable objects.
Cacciamani measured how well study participants performed at categorizing the words as natural or artificial by recording speed and accuracy.
"We found that participants performed better on the natural/artificial word task when that word followed a silhouette whose ground contained an object of the same rather than a different category," Cacciamani said.
This indicates that the brain accessed the meaning of the objects in the silhouette’s grounds even though study participants didn’t know the objects were there, she said.
"Every day our visual systems are bombarded with more information than we can consciously be aware of," Cacciamani said. "We’re showing that your brain might still be accessing information without your conscious awareness, and that could influence your behavior."

Don’t Underestimate Your Mind’s Eye

Take a look around, and what do you see? Much more than you think you do, thanks to your finely tuned mind’s eye, which processes images without your even knowing.

A University of Arizona study has found that objects in our visual field of which we are not consciously aware still may influence our decisions. The findings refute traditional ideas about visual perception and cognition, and they could shed light on why we sometimes make decisions — stepping into a street, choosing not to merge into a traffic lane — without really knowing why.

Laura Cacciamani, who recently earned her doctorate in psychology with a minor in neuroscience, has found supporting evidence. Cacciamani’s is the lead author on a co-authored study, published online in the journal Attention, Perception and Psychophysics, shows that the brain’s subconscious processing has an impact on behavior and decision-making.

This seems to make evolutionary sense, Cacciamani said. Early humans would have required keen awareness of their surroundings on a subliminal level in order to survive.

"Your brain is always monitoring for meaning in the world, to be aware of your general surroundings and potential predators," Cacciamani said. "You can be focused on a task, but your brain is assessing the meaning of everything around you – even objects that you’re not consciously perceiving."

The study builds on the findings of earlier research by Jay Sanguinetti, who also was a doctoral candidate in the UA Department of Psychology. Both studies go against conventional wisdom among vision scientists.

"According to the traditional view, the brain accesses the meaning – or the memory – of an object after you perceive it," Cacciamani said. "Against this view, we have now shown that the meaning of an object can be accessed before conscious perception.

"We’re showing that there’s more interplay between memory and perception than previously has been assumed," she said.

Cacciamani asked participants in her study to classify nouns that appeared on a computer screen as naming a natural object or artificial object by pressing one of two buttons labeled “natural” or “artificial.” For example, the word “leaf” indicates an object found in nature, while “anchor” indicates a man-made or artificial object.

But before each word appeared on the screen, the computer flashed a black silhouette that – unknown to participants – had portions of natural or artificial objects suggested along the white outside regions (called the “ground” regions) of the image. Participants were not told to look for anything in the silhouettes, and they were flashed so quickly – 50 milliseconds – that it would have been difficult to notice the objects in the ground regions even if someone knew what to look for. Participants never were aware that the silhouette’s grounds suggested recognizable objects.

Cacciamani measured how well study participants performed at categorizing the words as natural or artificial by recording speed and accuracy.

"We found that participants performed better on the natural/artificial word task when that word followed a silhouette whose ground contained an object of the same rather than a different category," Cacciamani said.

This indicates that the brain accessed the meaning of the objects in the silhouette’s grounds even though study participants didn’t know the objects were there, she said.

"Every day our visual systems are bombarded with more information than we can consciously be aware of," Cacciamani said. "We’re showing that your brain might still be accessing information without your conscious awareness, and that could influence your behavior."

Filed under visual perception decision making visual awareness object perception psychology neuroscience science

  1. passinpassion reblogged this from neurosciencestuff
  2. richardsbattella reblogged this from neurosciencestuff
  3. nordicmansworld reblogged this from neurosciencestuff
  4. thisisallpractice reblogged this from abstractnerd and added:
    This is why I study what I do. This is so god damn cool. Brains are bad as fuck.
  5. abstractnerd reblogged this from neurosciencestuff
  6. icewindandthisisinteresting reblogged this from neurosciencestuff
  7. hiphopandinsubordination reblogged this from adistinguishedvillain
  8. limisfolie reblogged this from neurosciencestuff
  9. generatepineal reblogged this from babygirltripps
  10. baebulsaur reblogged this from neurosciencestuff
  11. microbiomusings reblogged this from neurosciencestuff
  12. virtualistik reblogged this from lackadaisicallilliputian
  13. sart-rosa reblogged this from neurosciencestuff
  14. maverickanoidite reblogged this from speculative-evolution
  15. roaminc reblogged this from speculative-evolution
  16. lackadaisicallilliputian reblogged this from speculative-evolution
  17. ahuitilozotlec reblogged this from speculative-evolution
  18. speculative-evolution reblogged this from neurosciencestuff
  19. kentuckyfriedbitches reblogged this from neurosciencestuff
  20. rom-sp4ceknight reblogged this from starsaremymuse
  21. isenpaikun reblogged this from neurosciencestuff
  22. cutie3pnt14159 reblogged this from starsaremymuse
free counters