Neuroscience

Articles and news from the latest research reports.

216 notes

Inside the Letterbox: How Literacy Transforms the Human Brain

Although I find the diversity of the world’s writing systems bewildering, there is also a striking regularity that remains hidden. Whenever we read—whether our language is Japanese, Hebrew, English, or Italian—each of us relies on very similar brain networks. In particular, a small region of the visual cortex becomes active with remarkable reproducibility in the brains of all readers. A brief localizer scan, during which images of brain activity are collected as a person responds to written words, faces, objects, and other visual stimuli, serves to identify this region. Written words never fail to activate a small region at the base of the left hemisphere, always at the same place, give or take a few millimeters.
Experts call this region the visual word form area, but in a recent book for the general public, I dubbed it the brain’s letterbox, because it concentrates much of our visual knowledge of letters and their configurations. Indeed, this site is amazingly specialized. The letterbox responds to written words more than it does to most other categories of visual stimuli, including pictures of faces, objects, houses, and even Arabic numerals.Its efficiency is so great that it even responds to words that we fail to recognize consciously—words made subliminal by flashing them for a fraction of a second. Yet it performs highly sophisticated operations that are indispensable to fluent reading. For instance, the letterbox is the first visual area that recognizes that “READ” and “read” depict the same word by representing strings of letters invariantly for changes in case, which is no small feat if you consider that uppercase and lowercase letters such as “A” and “a” bear very little similarity. Furthermore, if it is impaired or disconnected via brain surgery or a cerebral infarct (type of stroke), the patient may develop a syndrome called pure alexia. He or she will be unable to recognize even a single word, as well as faces, objects, digits, and Arabic numerals. Yet many of these patients can still speak and understand spoken language fluently, and they may even still write; only their visual capacity to process letter strings seems dramatically affected.
The brain of any educated adult contains a circuit specialized for reading. But how is this possible, given that reading is an extremely recent and highly variable cultural activity? The alphabet is only about 4,000 years old, and until recently, only a very small fraction of humanity could read. Thus, there was no time for Darwinian evolution to shape our genome and adapt our brain networks to the particularities of reading. How is it, then, that we all possess a specialized letterbox area?

Read more

Inside the Letterbox: How Literacy Transforms the Human Brain

Although I find the diversity of the world’s writing systems bewildering, there is also a striking regularity that remains hidden. Whenever we read—whether our language is Japanese, Hebrew, English, or Italian—each of us relies on very similar brain networks. In particular, a small region of the visual cortex becomes active with remarkable reproducibility in the brains of all readers. A brief localizer scan, during which images of brain activity are collected as a person responds to written words, faces, objects, and other visual stimuli, serves to identify this region. Written words never fail to activate a small region at the base of the left hemisphere, always at the same place, give or take a few millimeters.

Experts call this region the visual word form area, but in a recent book for the general public, I dubbed it the brain’s letterbox, because it concentrates much of our visual knowledge of letters and their configurations. Indeed, this site is amazingly specialized. The letterbox responds to written words more than it does to most other categories of visual stimuli, including pictures of faces, objects, houses, and even Arabic numerals.Its efficiency is so great that it even responds to words that we fail to recognize consciously—words made subliminal by flashing them for a fraction of a second. Yet it performs highly sophisticated operations that are indispensable to fluent reading. For instance, the letterbox is the first visual area that recognizes that “READ” and “read” depict the same word by representing strings of letters invariantly for changes in case, which is no small feat if you consider that uppercase and lowercase letters such as “A” and “a” bear very little similarity. Furthermore, if it is impaired or disconnected via brain surgery or a cerebral infarct (type of stroke), the patient may develop a syndrome called pure alexia. He or she will be unable to recognize even a single word, as well as faces, objects, digits, and Arabic numerals. Yet many of these patients can still speak and understand spoken language fluently, and they may even still write; only their visual capacity to process letter strings seems dramatically affected.

The brain of any educated adult contains a circuit specialized for reading. But how is this possible, given that reading is an extremely recent and highly variable cultural activity? The alphabet is only about 4,000 years old, and until recently, only a very small fraction of humanity could read. Thus, there was no time for Darwinian evolution to shape our genome and adapt our brain networks to the particularities of reading. How is it, then, that we all possess a specialized letterbox area?

Read more

Filed under letterbox visual stimuli brain activity brain circuitry psychology neuroscience science

264 notes

MACH system from MIT can coach those with social anxiety
Plenty of people out there have a serious phobia of public speaking and there are tons of other disorders, such as Asperger’s, that severely limit a person’s ability to handle even simple social interactions. M. Ehsan Hoque, a student at the MIT Media Lab, has made these subjects the focus of her latest project: MACH (My Automated Conversation coacH). At the heart of MACH is a complex system of facial and speech recognition algorithms that can detect subtle nuances in intonation while tracking smiles, head nods and eye movement. The latter is especially important since the front end of MACH is a computer generated avatar that can tell when you break eye contact and shift your attention elsewhere.
The software then provides feedback about your performance, helping to prep you for that big presentation or just guide you out of your shell. Experimental data suggests that coaching from MACH could even help you perform better in a job interview. What’s particularly exciting is that the program requires no special hardware; it’s designed to be used with a standard webcam and microphone on a laptop. So it might not be too long before we start seeing apps designed to help users through social awkwardness.

MACH system from MIT can coach those with social anxiety

Plenty of people out there have a serious phobia of public speaking and there are tons of other disorders, such as Asperger’s, that severely limit a person’s ability to handle even simple social interactions. M. Ehsan Hoque, a student at the MIT Media Lab, has made these subjects the focus of her latest project: MACH (My Automated Conversation coacH). At the heart of MACH is a complex system of facial and speech recognition algorithms that can detect subtle nuances in intonation while tracking smiles, head nods and eye movement. The latter is especially important since the front end of MACH is a computer generated avatar that can tell when you break eye contact and shift your attention elsewhere.

The software then provides feedback about your performance, helping to prep you for that big presentation or just guide you out of your shell. Experimental data suggests that coaching from MACH could even help you perform better in a job interview. What’s particularly exciting is that the program requires no special hardware; it’s designed to be used with a standard webcam and microphone on a laptop. So it might not be too long before we start seeing apps designed to help users through social awkwardness.

Filed under MACH social interaction social anxiety public speaking technology science

207 notes

People can sense a smile before it appears on the face

But a forced or polite smile does not transmit the same signals, meaning we only detect it when it is visible, reports journal Psychological Science.


Researchers say the study reflects the unique social value of a heartfelt smile, which involves specific movements of muscles around the eyes.


A team from Bangor University had noted that pairs of strangers getting to know one another not only exchanged smiles, they almost always matched the particular smile type, whether genuine or polite.


But they responded much more quickly to their partners’ genuine smiles than their polite smiles, suggesting that they were anticipating the genuine smiles.


In the lab, the results were repeated and data from electrical sensors on participants’ faces revealed that they engaged smile-related muscles when they expected a genuine smile to appear but showed no such activity when expecting polite smiles.
The different responses suggest that genuine smiles are more valuable social rewards, said Dr Erin Heerey.
She said: “These findings give us the first clear suggestion that the basic processes that guide responses to reward also play a role in guiding social behaviour on a moment-to-moment basis during interactions.
"No two interactions are alike, yet people still manage to smoothly coordinate their speech and nonverbal behaviors with those of another person."
She said that polite smiles typically occur when sociocultural norms dictate that smiling is appropriate.
Genuine smiles, on the other hand, signify pleasure, occur spontaneously, and are indicated by engagement of specific muscles around the eye.
She said the study could help those who find social interactions tricky.
She explained: “As we progress in our understanding of how social interactions unfold, these findings may help to guide the development of interventions for people who find social interactions difficult, such as those with social anxiety, autism, or schizophrenia.”

People can sense a smile before it appears on the face

But a forced or polite smile does not transmit the same signals, meaning we only detect it when it is visible, reports journal Psychological Science.

Researchers say the study reflects the unique social value of a heartfelt smile, which involves specific movements of muscles around the eyes.

A team from Bangor University had noted that pairs of strangers getting to know one another not only exchanged smiles, they almost always matched the particular smile type, whether genuine or polite.

But they responded much more quickly to their partners’ genuine smiles than their polite smiles, suggesting that they were anticipating the genuine smiles.

In the lab, the results were repeated and data from electrical sensors on participants’ faces revealed that they engaged smile-related muscles when they expected a genuine smile to appear but showed no such activity when expecting polite smiles.

The different responses suggest that genuine smiles are more valuable social rewards, said Dr Erin Heerey.

She said: “These findings give us the first clear suggestion that the basic processes that guide responses to reward also play a role in guiding social behaviour on a moment-to-moment basis during interactions.

"No two interactions are alike, yet people still manage to smoothly coordinate their speech and nonverbal behaviors with those of another person."

She said that polite smiles typically occur when sociocultural norms dictate that smiling is appropriate.

Genuine smiles, on the other hand, signify pleasure, occur spontaneously, and are indicated by engagement of specific muscles around the eye.

She said the study could help those who find social interactions tricky.

She explained: “As we progress in our understanding of how social interactions unfold, these findings may help to guide the development of interventions for people who find social interactions difficult, such as those with social anxiety, autism, or schizophrenia.”

Filed under smiles social interaction social anxiety psychology neuroscience science

241 notes

Scientists discover new layer of the human cornea
Scientists at The University of Nottingham have discovered a previously undetected layer in the cornea, the clear window at the front of the human eye.
The breakthrough, announced in a study published in the academic journal Ophthalmology, could help surgeons to dramatically improve outcomes for patients undergoing corneal grafts and transplants.
The new layer has been dubbed the Dua’s Layer after the academic Professor Harminder Dua who discovered it.
Professor Dua, Professor of Ophthalmology and Visual Sciences, said: “This is a major discovery that will mean that ophthalmology textbooks will literally need to be re-written. Having identified this new and distinct layer deep in the tissue of the cornea, we can now exploit its presence to make operations much safer and simpler for patients.
“From a clinical perspective, there are many diseases that affect the back of the cornea which clinicians across the world are already beginning to relate to the presence, absence or tear in this layer.”
Tough and strong
The human cornea is the clear protective lens on the front of the eye through which light enters the eye. Scientists previously believed the cornea to be comprised of five layers, from front to back, the corneal epithelium, Bowman’s layer, the corneal stroma, Descemet’s membrane and the corneal endothelium.
The new layer that has been discovered is located at the back of the cornea between the corneal stroma and Descemet’s membrane. Although it is just 15 microns thick — the entire cornea is around 550 microns thick or 0.5mm — it is incredibly tough and is strong enough to be able to withstand one and a half to two bars of pressure.
The scientists proved the existence of the layer by simulating human corneal transplants and grafts on eyes donated for research purposes to eye banks located in Bristol and Manchester.
During this surgery, tiny bubbles of air were injected into the cornea to gently separate the different layers. The scientists then subjected the separated layers to electron microscopy, allowing them to study them at many thousand times their actual size.
Better outcomes
Understanding the properties and location of the new Dua’s layer could help surgeons to better identify where in the cornea these bubbles are occurring and take appropriate measures during the operation. If they are able to inject a bubble next to the Dua’s layer, its strength means that it is less prone to tearing, meaning a better outcome for the patient.
The discovery will have an impact on advancing understanding of a number of diseases of the cornea, including acute hydrops, Descematocele and pre-Descemet’s dystrophies.
The scientists now believe that corneal hydrops, a bulging of the cornea caused by fluid build up that occurs in patients with keratoconus (conical deformity of the cornea), is caused by a tear in the Dua layer, through which water from inside the eye rushes in and causes waterlogging.

Scientists discover new layer of the human cornea

Scientists at The University of Nottingham have discovered a previously undetected layer in the cornea, the clear window at the front of the human eye.

The breakthrough, announced in a study published in the academic journal Ophthalmology, could help surgeons to dramatically improve outcomes for patients undergoing corneal grafts and transplants.

The new layer has been dubbed the Dua’s Layer after the academic Professor Harminder Dua who discovered it.

Professor Dua, Professor of Ophthalmology and Visual Sciences, said: “This is a major discovery that will mean that ophthalmology textbooks will literally need to be re-written. Having identified this new and distinct layer deep in the tissue of the cornea, we can now exploit its presence to make operations much safer and simpler for patients.

“From a clinical perspective, there are many diseases that affect the back of the cornea which clinicians across the world are already beginning to relate to the presence, absence or tear in this layer.”

Tough and strong

The human cornea is the clear protective lens on the front of the eye through which light enters the eye. Scientists previously believed the cornea to be comprised of five layers, from front to back, the corneal epithelium, Bowman’s layer, the corneal stroma, Descemet’s membrane and the corneal endothelium.

The new layer that has been discovered is located at the back of the cornea between the corneal stroma and Descemet’s membrane. Although it is just 15 microns thick — the entire cornea is around 550 microns thick or 0.5mm — it is incredibly tough and is strong enough to be able to withstand one and a half to two bars of pressure.

The scientists proved the existence of the layer by simulating human corneal transplants and grafts on eyes donated for research purposes to eye banks located in Bristol and Manchester.

During this surgery, tiny bubbles of air were injected into the cornea to gently separate the different layers. The scientists then subjected the separated layers to electron microscopy, allowing them to study them at many thousand times their actual size.

Better outcomes

Understanding the properties and location of the new Dua’s layer could help surgeons to better identify where in the cornea these bubbles are occurring and take appropriate measures during the operation. If they are able to inject a bubble next to the Dua’s layer, its strength means that it is less prone to tearing, meaning a better outcome for the patient.

The discovery will have an impact on advancing understanding of a number of diseases of the cornea, including acute hydrops, Descematocele and pre-Descemet’s dystrophies.

The scientists now believe that corneal hydrops, a bulging of the cornea caused by fluid build up that occurs in patients with keratoconus (conical deformity of the cornea), is caused by a tear in the Dua layer, through which water from inside the eye rushes in and causes waterlogging.

Filed under vision human eye cornea Dua’s layer science

86 notes

Can you feel me now?

New array measures vibrations across the skin, may help engineers design optimal, wearable tactile displays.

image

In the near future, a buzz in your belt or a pulse from your jacket may give you instructions on how to navigate your surroundings.

Think of it as tactile Morse code: vibrations from a wearable, GPS-linked device that tell you to turn right or left, or stop, depending on the pattern of pulses you feel. Such a device could free drivers from having to look at maps, and could also serve as a tactile guide for the visually and hearing impaired.

Lynette Jones, a senior research scientist in MIT’s Department of Mechanical Engineering, designs wearable tactile displays. Through her work, she’s observed that the skin is a sensitive — though largely untapped — medium for communication.

“If you compare the skin to the retina, you have about the same number of sensory receptors, you just have them over almost two square meters of space, unlike the eye where it’s all concentrated in an extremely small area,” Jones says. “The skin is generally as useful as a very acute area. It’s just that you need to disperse the information that you’re presenting.”

Knowing just how to disperse tactile information across the skin is tricky. For instance, people may be much more sensitive to stimuli on areas like the hand, as opposed to the forearm, and may respond best to certain patterns of vibrations. Such information on skin responsiveness could help designers determine the best configuration of motors in a display, given where on the skin a device would be worn.

Now Jones has built an array that precisely tracks a motor’s vibrations through skin in three dimensions. The array consists of eight miniature accelerometers and a single pancake motor — a type of vibrating motor used in cellphones. She used the array to measure motor vibrations in three locations: the palm of the hand, the forearm and the thigh. From her studies with eight healthy participants, Jones found that a motor’s mechanical vibrations through skin drop off quickly in all three locations, within 8 millimeters from where the vibrations originated.

Jones also gauged participants’ perception of vibrations, fitting them with a 3-by-3 array of pancake motors in these three locations on the body. While skin generally stopped vibrating 8 millimeters from the source, most people continued to perceive the vibrations as far away as 24 millimeters.

When participants were asked to identify specific locations of motors within the array, they were much more sensitive on the palm than on the forearm or thigh. But in all three locations, people were better at picking out vibrations in the four corners of the array, versus the inner motors, leading Jones to posit that perhaps people use the edges of their limbs to localize vibrations and other stimuli.

“For a lot of sensory modalities, you have to work out what it is people can process, as one of the dictates for how you design,” says Jones, whose results will appear in the journal IEEE Transactions on Haptics. “There’s no point in making things much more compact, which may be a desirable feature from an engineering point of view, but from a human-use point of view, doesn’t make a difference.”

Mapping good vibrations

In addition to measuring skin’s sensitivity to vibrations, Jones and co-author Katherine Sofia ’12 found that skin has a strong effect on motor vibrations. The researchers compared a pancake motor’s frequency of vibrations when mounted on a rigid structure or on more compliant skin. They found that in general, skin reduced a motor’s vibrations by 28 percent, with the forearm and thigh having a slightly stronger dampening effect than the palm of the hand.

The skin’s damping of motor vibrations is significant, Jones says, if engineers plan to build tactile displays that incorporate different frequencies of vibrations. For instance, the difference between two motors — one slightly faster than the other — may be indistinguishable in certain parts of the skin. Likewise, two motors spaced a certain distance apart may be differentiable in one area but not another.

“Should I have eight motors, or is four enough that 90 percent of the time, I’ll know that when this one’s on, it’s this one and not that one?” Jones says. “We’re answering those sorts of questions in the context of what information you want to present using a device.”

Roberta Klatzky, a professor of psychology at Carnegie Mellon University, says that measurements taken by Jones’ arrays can be used to set up displays in which the location of a stimulus — for example, a pattern to convey a letter — is important.

“A major challenge is to enable people to tell the difference between patterns applied to the skin as, for example, blind people do when reading Braille,” says Klatzky, who specializes in the study of spatial cognition. “Lynette’s work sets up a methodology and potential guidelines for effective pattern displays.”

Creating a buzz

Jones sees promising applications for wearable tactile displays. In addition to helping drivers navigate, she says tactile stimuli may direct firefighters through burning buildings, or emergency workers through disaster sites. In more mundane scenarios, she says tactile displays may help joggers traverse an unfamiliar city, taking directions from a buzzing wristband, instead of having to look at a smartphone. 

Using data from their mechanical and perceptual experiments, Jones’ group is designing arrays that can be worn across the back and around the wrist, and is investigating various ways to present vibrations. For example, a row of vibrations activated sequentially from left to right may tell a driver to turn right; a single motor that buzzes with increasing frequency may be a warning to slow down.

“There’s a lot of things you can do with these displays that are fairly intuitive in terms of how people respond,” Jones says, “which is important because no one’s going to spend hours and hours in any application, learning what a signal means.”

(Source: web.mit.edu)

Filed under sensory receptors haptics sensory modalities somatosensory modalities neuroscience science

197 notes

From the mouths of babes – The truth about toddler talk

The sound of small children chattering has always been considered cute – but not particularly sophisticated. However, research by a Newcastle University expert has shown their speech is far more advanced than previously understood.

image

Dr Cristina Dye, a lecturer in child language development, found that two to three- year-olds are using grammar far sooner than expected.

She studied fifty French speaking youngsters aged between 23 and 37 months, capturing tens of thousands of their utterances.

Dr Dye, who carried out the research while at Cornell University in the United States, found that the children were using ‘little words’ which form the skeleton of sentences such as a, an, can, is, an, far sooner than previously thought.

Dr Dye and her team used advanced recording technology including highly sensitive microphones placed close to the children, to capture the precise sounds the children voiced. They spent years painstakingly analysing every minute sound made by the toddlers and the context in which it was produced.

They found a clear, yet previously undetected, pattern of sounds and puffs of air, which consistently replaced grammatical words in many of the children’s utterances.

Dr Dye said: “Many of the toddlers we studied made a small sound, a soft breath, or a pause, at exactly the place that a grammatical word would normally be uttered.” 

“The fact that this sound was always produced in the correct place in the sentence leads us to believe that young children are knowledgeable of grammatical words. They are far more sophisticated in their grammatical competence than we ever understood.

“Despite the fact the toddlers we studied were acquiring French, our findings are expected to extend to other languages. I believe we should give toddlers more credit – they’re much more amazing than we realised.”

For decades the prevailing view among developmental specialists has been that children’s early word combinations are devoid of grammatical words. On this view, children then undergo a ‘tadpole to frog’ transformation where due to an unknown mechanism, they start to develop grammar in their speech. Dye’s results now challenge the old view.

Dr Dye said: “The research sheds light on a really important part of a child’s development. Language is one of the things that makes us human and understanding how we acquire it shows just how amazing children are.

“There are also implications for understanding language delay in children. When children don’t learn to speak normally it  can lead to serious issues later in life. For example, those who have it are more likely to suffer from mental illness or be unemployed later in life. If we can understand what is ‘normal’ as early as possible then we can intervene sooner to help those children.”

The research was originally published in the Journal of Linguistics.

(Source: ncl.ac.uk)

Filed under language development speech toddlers grammar auxiliaries semantics neuroscience psychology science

35 notes

Functional MRI provides support in operations on the brain

Researchers at the MedUni Vienna have proved in a so far unique multicenter study that clinical functional magnetic resonance tomography (fMRI), in the area in which the MedUni Vienna has a leading role internationally, is a safe method in brain surgery. With the aid of fMRI imaging can pinpoint to the millimetre where critical nerve fibres (e.g. vital for speech or hand function) lie and which have to be avoided – in operations on brain tumours for example.

image

"With the assistance of functional magnetic resonance tomography we are, if you like, drawing a red line for the surgeon so he knows where not to make an incision so as to avoid damage," says Roland Beisteiner from the University Department of Neurology at the MedUni Vienna. The neurologist and president of the Austrian Society for fMRI was playing a part in the development of fMRI as early as 1992, initiating its development in Austria. Since then this method has been developed and implemented at the University Department of Neurology and the High Field MRI Center of Excellence.

Now Beisteiner’s team have been able for the first time to demonstrate in a current paper in the top journal “Radiology" that functional magnetic resonance tomography provides diagnostic certainty in operations on the brain – no matter what the equipment is (whether a 7Tesla magnetic resonance tomograph as in Vienna or even only a 1.5Tesla), no matter in which location and also irrespective of who is operating it. The Medical Universities in Innsbruck and Salzburg, the Heinrich Heine University of Düsseldorf and the Stiftungsklinikum Koblenz (Koblenz Hospital Foundation) also took part in the study.

The “Imaging and Cognition Biology” Research Cluster of the MedUni and Vienna University

Likewise, with the help of functional magnetic resonance tomography, the teams of Beisteiner and Tecumseh Fitch (Faculty of Life Sciences of the University of Vienna) are investigating in a joint research cluster belonging to the MedUni Vienna and the University of Vienna whether the structural and syntactic processing of music takes place in similar areas of the brain as does the processing of speech. Says Beisteiner: “It is never exactly the same area of the brain; however, brain activities can overlap when talking or playing an instrument.”

The main focus of the research cluster is to determine precisely the common areas of the brain involved and to develop new treatments by activating them. These could perhaps then be used on people suffering from aphasia, which is a loss of language as the result of brain damage mostly to the left half of the brain.

According to Beisteiner there have been some astonishing results: “People, who could no longer speak because of their aphasia, have been able to sing the words they have learned to the matching tune.” From this one can conclude that it would seem to make sense to also practise music skills during speech therapy.

The “Imaging and Cognition Biology” research cluster is one of six joint clusters at the MedUni Vienna with the University of Vienna, which were set up in 2011. Further information: http://forschungscluster.meduniwien.ac.at/.

(Source: meduniwien.ac.at)

Filed under brain brain activity fMRI neuroimaging neuroscience science

146 notes

Stress Test and Brain Scans Pinpoint Two Distinct Forms of Gulf War Illness
Researchers at Georgetown University Medical Center say their new work suggests that Gulf War illness may have two distinct forms depending on which brain regions have atrophied. Their study of Gulf War veterans, published online today in PLOS ONE, may help explain why clinicians have consistently encountered veterans with different symptoms and complaints.
Using brain imaging that was acquired before and after exercise tests, the researchers studied the effects of physical stress on the veterans and controls. Following exercise, subgroups were evident. In 18 veterans, they found that pain levels increased after completion of the exercise stress tests exercised; fMRI scans in these participants showed loss of brain matter in adjacent regions associated with pain regulation.  
During cognitive tasks, this group showed an increased use of the basal ganglia — a potential compensatory strategy the brain uses that is also seen in neurodegenerative disorders such as Alzheimer’s disease. Following exercise, this group lost the ability to employ their basal ganglia, suggesting an adverse response to a physiological stressor.
In addition, “a separate group of 10 veterans had a very different clinical alteration,” says lead author Rakib Rayhan, a researcher in the lab of the study’s senior investigator, James Baraniuk, MD, a professor of medicine at GUMC.
In these 10 veterans, the researchers found substantial increases in heart rate. They also discovered that this subgroup had atrophy in the brain stem, which regulates heart rate. .
In addition, brain scans during a cognitive task performed prior to exercise showed increased compensatory use of the cerebellum, again a trait seen in neurodegenerative disorders. Like the other group, this cohort lost the ability to use this compensatory area after exercise.
Alterations in cognition, brain structure and exercise-induced symptoms found in the veterans were absent in the 10-participant matched control group, the researchers say.
“The use of other brain areas to compensate for a damaged area is seen in other disorders, such as Alzheimer’s disease, which is why we believe our data show that these veterans are suffering from central nervous system dysfunction,” Rayhan explains. He adds, however, that because such changes are similar to other neurodegenerative states, it doesn’t mean that veterans will progress to Alzheimer’s or other diseases.
These findings — a surprise to researchers — follow a study in Gulf War veterans published in March in PLOS ONE that reported abnormalities in the bundle of nerve fibers connecting the brain areas involved in the processing and perception of pain and fatigue.
Gulf War Illness is the mysterious malady believed to have affected more than 200,000 military personnel who served in the 1990-1991 Operation Desert Shield and Desert Storm.
Although veterans were exposed to nerve agents, pesticides and herbicides (among other toxic chemicals), no one has definitively linked any single exposure or underlying mechanism to Gulf War illness.
The symptoms of Gulf War illness — which have not been widely accepted by the public or medical professionals — range from mild to debilitating and can include widespread pain, fatigue and headache, as well as cognitive and gastrointestinal dysfunctions.
“Our findings help explain and validate what these veterans have long said about their illness,” Rayhan says.

Stress Test and Brain Scans Pinpoint Two Distinct Forms of Gulf War Illness

Researchers at Georgetown University Medical Center say their new work suggests that Gulf War illness may have two distinct forms depending on which brain regions have atrophied. Their study of Gulf War veterans, published online today in PLOS ONE, may help explain why clinicians have consistently encountered veterans with different symptoms and complaints.

Using brain imaging that was acquired before and after exercise tests, the researchers studied the effects of physical stress on the veterans and controls. Following exercise, subgroups were evident. In 18 veterans, they found that pain levels increased after completion of the exercise stress tests exercised; fMRI scans in these participants showed loss of brain matter in adjacent regions associated with pain regulation.  

During cognitive tasks, this group showed an increased use of the basal ganglia — a potential compensatory strategy the brain uses that is also seen in neurodegenerative disorders such as Alzheimer’s disease. Following exercise, this group lost the ability to employ their basal ganglia, suggesting an adverse response to a physiological stressor.

In addition, “a separate group of 10 veterans had a very different clinical alteration,” says lead author Rakib Rayhan, a researcher in the lab of the study’s senior investigator, James Baraniuk, MD, a professor of medicine at GUMC.

In these 10 veterans, the researchers found substantial increases in heart rate. They also discovered that this subgroup had atrophy in the brain stem, which regulates heart rate. .

In addition, brain scans during a cognitive task performed prior to exercise showed increased compensatory use of the cerebellum, again a trait seen in neurodegenerative disorders. Like the other group, this cohort lost the ability to use this compensatory area after exercise.

Alterations in cognition, brain structure and exercise-induced symptoms found in the veterans were absent in the 10-participant matched control group, the researchers say.

“The use of other brain areas to compensate for a damaged area is seen in other disorders, such as Alzheimer’s disease, which is why we believe our data show that these veterans are suffering from central nervous system dysfunction,” Rayhan explains. He adds, however, that because such changes are similar to other neurodegenerative states, it doesn’t mean that veterans will progress to Alzheimer’s or other diseases.

These findings — a surprise to researchers — follow a study in Gulf War veterans published in March in PLOS ONE that reported abnormalities in the bundle of nerve fibers connecting the brain areas involved in the processing and perception of pain and fatigue.

Gulf War Illness is the mysterious malady believed to have affected more than 200,000 military personnel who served in the 1990-1991 Operation Desert Shield and Desert Storm.

Although veterans were exposed to nerve agents, pesticides and herbicides (among other toxic chemicals), no one has definitively linked any single exposure or underlying mechanism to Gulf War illness.

The symptoms of Gulf War illness — which have not been widely accepted by the public or medical professionals — range from mild to debilitating and can include widespread pain, fatigue and headache, as well as cognitive and gastrointestinal dysfunctions.

“Our findings help explain and validate what these veterans have long said about their illness,” Rayhan says.

Filed under Gulf War illness brain imaging stress cognitive tasks psychology neuroscience science

49 notes

Scientists identify neurons that control feeding behavior in Drosophila

Scientists at the University of Massachusetts Medical School have developed a novel transgenic system which allows them to remotely activate individual brain cells in the model organism Drosophila using ambient temperature. This powerful new tool for identifying and characterizing neural circuitry has lead to the identification of a pair of neurons – now called Fdg neurons – in the fruit fly that decide when to eat and initiate the subsequent feeding action. Discovery of these neurons may help neurobiologists better understand how the brain uses memory and stimuli to produce classically conditioned responses, such as those often associated with phobias or drug tolerance. The study appears in the journal Nature.

image

"For any organism, the decision to eat is a complex integration of internal and external stimuli leading to the activation of an organized sequence of motor patterns," said Motojiro Yoshihara, PhD, assistant professor of neurobiology at the University of Massachusetts Medical School and lead author of the Nature study. “By developing genetic tools to remotely activate individual brain cells in Drosophila, we’ve been able to isolate a pair of neurons that are critical to the act of eating in fruit flies. More importantly, we now have a powerful new tool with which we can answer important questions about the function and composition of neural circuitry.”

To isolate the neurons responsible for sensing food and initiating the complex feeding program in Drosophila, UMMS scientists had to develop a method of studying the behavior of freely moving flies while targeting and manipulating individual neurons. To accomplish this, Dr. Yoshihara expressed temperature activated genes in random neurons in more than 800 Drosophila lines. Placing these genetically modified flies in a small temperature-controlled chamber, he was able to active these genes by increasing and decreasing the ambient temperature. This, in turn, activated the corresponding neurons.

Under wild conditions, when a hungry fly comes in contact with food it ceases motion and executives eight basic motor functions resulting in the consumption of the food. When the temperature in the chamber was increased, Yoshihara and colleagues were able to isolate a single Drosophila line which exhibited these eight motor functions, even in the absence of food or other stimuli. Subsequent experiments revealed that the feeding mechanism initiated by activating the transgenes was being controlled by a single pair of neurons in the fly’s brain. Furthermore, these feeding (Fdg) neurons were responsible for synthesizing cues about available food and hunger, and using them to start the feeding mechanism.

"Our results showed that these neurons become active in the presence of a food source for the fly, but the response was contingent on whether the animal was hungry," said Yoshihara. "This means that these neurons are integrating both internal and external stimuli in order to initiate a complex feeding behavior with multiple motor programs."

Yoshihara believes this discovery will provide researchers with a powerful new tool for isolating, analyzing and characterizing aspects of the brain’s neural circuitry and studying how information is integrated in the brain. In the future, Yoshihara plans to use the Fdg-neurons to study the biological basis of classical or Pavlovian conditioning. Doing so, he hopes to uncover how memory integrates stimuli to illicit a conditioned behavior.

(Source: eurekalert.org)

Filed under neurons fdg neurons fruit flies neural circuitry motor response feeding conditioned behavior neuroscience science

74 notes

Memory-Boosting Chemical Is Identified in Mice
Memory improved in mice injected with a small, drug-like molecule discovered by UCSF San Francisco researchers studying how cells respond to biological stress.
The same biochemical pathway the molecule acts on might one day be targeted in humans to improve memory, according to the senior author of the study, Peter Walter, PhD, UCSF professor of biochemistry and biophysics and a Howard Hughes Investigator.
The discovery of the molecule and the results of the subsequent memory tests in mice were published in eLife, an online scientific open-access journal, on May 28, 2013.
In one memory test included in the study, normal mice were able to relocate a submerged platform about three times faster after receiving injections of the potent chemical than mice that received sham injections.
The mice that received the chemical also better remembered cues associated with unpleasant stimuli – the sort of fear conditioning that could help a mouse avoid being preyed upon.
Notably, the findings suggest that despite what would seem to be the importance of having the best biochemical mechanisms to maximize the power of memory, evolution does not seem to have provided them, Walter said.
“It appears that the process of evolution has not optimized memory consolidation; otherwise I don’t think we could have improved upon it the way we did in our study with normal, healthy mice,” Walter said.
The memory-boosting chemical was singled out from among 100,000 chemicals screened at the Small Molecule Discovery Center at UCSF for their potential to perturb a protective biochemical pathway within cells that is activated when cells are unable to keep up with the need to fold proteins into their working forms.
However, UCSF postdoctoral fellow Carmela Sidrauski, PhD, discovered that the chemical acts within the cell beyond the biochemical pathway that activates this unfolded protein response, to more broadly impact what’s known as the integrated stress response. In this response, several biochemical pathways converge on a single molecular lynchpin, a protein called eIF2 alpha.
Scientists have known that in organisms ranging in complexity from yeast to humans different kinds of cellular stress — a backlog of unfolded proteins, DNA-damaging UV light, a shortage of the amino acid building blocks needed to make protein, viral infection, iron deficiency — trigger different enzymes to act downstream to switch off eIF2 alpha.
“Among other things, the inactivation of eIF2 alpha is a brake on memory consolidation,” Walter said, perhaps an evolutionary consequence of a cell or organism becoming better able to adapt in other ways.
Turning off eIF2 alpha dials down production of most proteins, some of which may be needed for memory formation, Walter said. But eIF2 alpha inactivation also ramps up production of a few key proteins that help cells cope with stress.
Study co-author Nahum Sonenberg, PhD, of McGill University previously linked memory and eIF2 alpha in genetic studies of mice, and his lab group also conducted the memory tests for the current study.
The chemical identified by the UCSF researchers is called ISRIB, which stands for integrated stress response inhibitor. ISRIB counters the effects of eIF2 alpha inactivation inside cells, the researchers found.
“ISRIB shows good pharmacokinetic properties [how a drug is absorbed, distributed and eliminated], readily crosses the blood-brain barrier, and exhibits no overt toxicity in mice, which makes it very useful for studies in mice,” Walter said. These properties also indicate that ISRIB might serve as a good starting point for human drug development, according to Walter.
Walter said he is looking for scientists to collaborate with in new studies of cognition and memory in mouse models of neurodegenerative diseases and aging, using ISRIB or related molecules.
In addition, chemicals such as ISRIB could play a role in fighting cancers, which take advantage of stress responses to fuel their own growth, Walter said. Walter already is exploring ways to manipulate the unfolded protein response to inhibit tumor growth, based on his earlier discoveries.
At a more basic level, Walter said, he and other scientists can now use ISRIB to learn more about the role of the unfolded protein response and the integrated stress response in disease and normal physiology.

Memory-Boosting Chemical Is Identified in Mice

Memory improved in mice injected with a small, drug-like molecule discovered by UCSF San Francisco researchers studying how cells respond to biological stress.

The same biochemical pathway the molecule acts on might one day be targeted in humans to improve memory, according to the senior author of the study, Peter Walter, PhD, UCSF professor of biochemistry and biophysics and a Howard Hughes Investigator.

The discovery of the molecule and the results of the subsequent memory tests in mice were published in eLife, an online scientific open-access journal, on May 28, 2013.

In one memory test included in the study, normal mice were able to relocate a submerged platform about three times faster after receiving injections of the potent chemical than mice that received sham injections.

The mice that received the chemical also better remembered cues associated with unpleasant stimuli – the sort of fear conditioning that could help a mouse avoid being preyed upon.

Notably, the findings suggest that despite what would seem to be the importance of having the best biochemical mechanisms to maximize the power of memory, evolution does not seem to have provided them, Walter said.

“It appears that the process of evolution has not optimized memory consolidation; otherwise I don’t think we could have improved upon it the way we did in our study with normal, healthy mice,” Walter said.

The memory-boosting chemical was singled out from among 100,000 chemicals screened at the Small Molecule Discovery Center at UCSF for their potential to perturb a protective biochemical pathway within cells that is activated when cells are unable to keep up with the need to fold proteins into their working forms.

However, UCSF postdoctoral fellow Carmela Sidrauski, PhD, discovered that the chemical acts within the cell beyond the biochemical pathway that activates this unfolded protein response, to more broadly impact what’s known as the integrated stress response. In this response, several biochemical pathways converge on a single molecular lynchpin, a protein called eIF2 alpha.

Scientists have known that in organisms ranging in complexity from yeast to humans different kinds of cellular stress — a backlog of unfolded proteins, DNA-damaging UV light, a shortage of the amino acid building blocks needed to make protein, viral infection, iron deficiency — trigger different enzymes to act downstream to switch off eIF2 alpha.

“Among other things, the inactivation of eIF2 alpha is a brake on memory consolidation,” Walter said, perhaps an evolutionary consequence of a cell or organism becoming better able to adapt in other ways.

Turning off eIF2 alpha dials down production of most proteins, some of which may be needed for memory formation, Walter said. But eIF2 alpha inactivation also ramps up production of a few key proteins that help cells cope with stress.

Study co-author Nahum Sonenberg, PhD, of McGill University previously linked memory and eIF2 alpha in genetic studies of mice, and his lab group also conducted the memory tests for the current study.

The chemical identified by the UCSF researchers is called ISRIB, which stands for integrated stress response inhibitor. ISRIB counters the effects of eIF2 alpha inactivation inside cells, the researchers found.

“ISRIB shows good pharmacokinetic properties [how a drug is absorbed, distributed and eliminated], readily crosses the blood-brain barrier, and exhibits no overt toxicity in mice, which makes it very useful for studies in mice,” Walter said. These properties also indicate that ISRIB might serve as a good starting point for human drug development, according to Walter.

Walter said he is looking for scientists to collaborate with in new studies of cognition and memory in mouse models of neurodegenerative diseases and aging, using ISRIB or related molecules.

In addition, chemicals such as ISRIB could play a role in fighting cancers, which take advantage of stress responses to fuel their own growth, Walter said. Walter already is exploring ways to manipulate the unfolded protein response to inhibit tumor growth, based on his earlier discoveries.

At a more basic level, Walter said, he and other scientists can now use ISRIB to learn more about the role of the unfolded protein response and the integrated stress response in disease and normal physiology.

Filed under memory microRNA stress response proteins memory consolidation neuroscience science

free counters