Neuroscience

Articles and news from the latest research reports.

272 notes

Computer can read letters directly from the brain
By analysing MRI images of the brain with an elegant mathematical model, it is possible to reconstruct thoughts more accurately than ever before. In this way, researchers from Radboud University Nijmegen have succeeded in determining which letter a test subject was looking at. The journal Neuroimage has accepted the article, which will be published soon. A preliminary version of the article can be read online.
Functional MRI scanners have been used in cognition research primarily to determine which brain areas are active while test subjects perform a specific task. The question is simple: is a particular brain region on or off? A research group at the Donders Institute for Brain, Cognition and Behaviour at Radboud University has gone a step further: they have used data from the scanner to determine what a test subject is looking at. The researchers ‘taught’ a model how small volumes of 2x2x2 mm from the brain scans - known as voxels - respond to individual pixels. By combining all the information about the pixels from the voxels, it became possible to reconstruct the image viewed by the subject. The result was not a clear image, but a somewhat fuzzy speckle pattern. In this study, the researchers used hand-written letters.
Prior knowledge improves model performance‘After this we did something new’, says lead researcher Marcel van Gerven. ‘We gave the model prior knowledge: we taught it what letters look like. This improved the recognition of the letters enormously. The model compares the letters to determine which one corresponds most exactly with the speckle image, and then pushes the results of the image towards that letter. The result was the actual letter, a true reconstruction.’
‘Our approach is similar to how we believe the brain itself combines prior knowledge with sensory information. For example, you can recognise the lines and curves in this article as letters only after you have learned to read. And this is exactly what we are looking for: models that show what is happening in the brain in a realistic fashion. We hope to improve the models to such an extent that we can also apply them to the working memory or to subjective experiences such as dreams or visualisations. Reconstructions indicate whether the model you have created approaches reality.’
Improved resolution; more possibilities‘In our further research we will be working with a more powerful MRI scanner,’ explains Sanne Schoenmakers, who is working on a thesis about decoding thoughts. ‘Due to the higher resolution of the scanner, we hope to be able to link the model to more detailed images. We are currently linking images of letters to 1200 voxels in the brain; with the more powerful scanner we will link images of faces to 15,000 voxels.’

Computer can read letters directly from the brain

By analysing MRI images of the brain with an elegant mathematical model, it is possible to reconstruct thoughts more accurately than ever before. In this way, researchers from Radboud University Nijmegen have succeeded in determining which letter a test subject was looking at. The journal Neuroimage has accepted the article, which will be published soon. A preliminary version of the article can be read online.

Functional MRI scanners have been used in cognition research primarily to determine which brain areas are active while test subjects perform a specific task. The question is simple: is a particular brain region on or off? A research group at the Donders Institute for Brain, Cognition and Behaviour at Radboud University has gone a step further: they have used data from the scanner to determine what a test subject is looking at. The researchers ‘taught’ a model how small volumes of 2x2x2 mm from the brain scans - known as voxels - respond to individual pixels. By combining all the information about the pixels from the voxels, it became possible to reconstruct the image viewed by the subject. The result was not a clear image, but a somewhat fuzzy speckle pattern. In this study, the researchers used hand-written letters.

Prior knowledge improves model performance
‘After this we did something new’, says lead researcher Marcel van Gerven. ‘We gave the model prior knowledge: we taught it what letters look like. This improved the recognition of the letters enormously. The model compares the letters to determine which one corresponds most exactly with the speckle image, and then pushes the results of the image towards that letter. The result was the actual letter, a true reconstruction.’

‘Our approach is similar to how we believe the brain itself combines prior knowledge with sensory information. For example, you can recognise the lines and curves in this article as letters only after you have learned to read. And this is exactly what we are looking for: models that show what is happening in the brain in a realistic fashion. We hope to improve the models to such an extent that we can also apply them to the working memory or to subjective experiences such as dreams or visualisations. Reconstructions indicate whether the model you have created approaches reality.’

Improved resolution; more possibilities
‘In our further research we will be working with a more powerful MRI scanner,’ explains Sanne Schoenmakers, who is working on a thesis about decoding thoughts. ‘Due to the higher resolution of the scanner, we hope to be able to link the model to more detailed images. We are currently linking images of letters to 1200 voxels in the brain; with the more powerful scanner we will link images of faces to 15,000 voxels.’

Filed under neuroimaging brain activity brain scans neuroscience science

  1. viirulentscience reblogged this from neurosciencestuff
  2. icewindandthisisinteresting reblogged this from neurosciencestuff
  3. themargatron reblogged this from joshbyard
  4. tinkles reblogged this from peculiaraura
  5. peculiaraura reblogged this from neurosciencestuff
  6. satanicpizzaparty reblogged this from neurosciencestuff
  7. feeols reblogged this from neurosciencestuff
  8. moons-and-mirrors reblogged this from joshbyard
  9. modifiedcat reblogged this from neurosciencestuff
  10. nursinghealthcommunication reblogged this from neurosciencestuff and added:
    I am tracking down this forthcoming article that is previewed below.
  11. hotdogcephalopod reblogged this from medicalwatson
  12. celsampaio reblogged this from joshbyard and added:
    Radboud University Nijmegen rocks!!! I just hope I can be back soon…
  13. empathy-vs-apathy-nyc reblogged this from consciousnesscrisis
  14. consciousnesscrisis reblogged this from joshbyard
  15. sabannatheginger reblogged this from joshbyard
  16. caguilar649 reblogged this from neurosciencestuff
  17. averyunqiueurl reblogged this from joshbyard
  18. panda-owner reblogged this from joshbyard
  19. psychedelisex reblogged this from joshbyard
  20. juxtaposedinthought reblogged this from joshbyard
  21. aloh0mora reblogged this from joshbyard
  22. cmsmithsk reblogged this from joshbyard
  23. kaleidos-dope reblogged this from rahrainah
  24. leaderoftherevolution reblogged this from joshbyard
free counters