Neuroscience

Articles and news from the latest research reports.

86 notes

Is anxiety related to premature aging? A new study by researchers at Brigham and Women’s Hospital (BWH) shows that a common form of anxiety, known as phobic anxiety, was associated with shorter telomeres in middle-aged and older women. The study suggests that phobic anxiety is a possible risk factor for accelerated aging.
Read more: Anxiety Linked to Shortened Telomeres, Accelerated Aging

// ]]]]>]]>]]>

Is anxiety related to premature aging? A new study by researchers at Brigham and Women’s Hospital (BWH) shows that a common form of anxiety, known as phobic anxiety, was associated with shorter telomeres in middle-aged and older women. The study suggests that phobic anxiety is a possible risk factor for accelerated aging.

Read more: Anxiety Linked to Shortened Telomeres, Accelerated Aging

Filed under science neuroscience brain psychology anxiety

33 notes

Oligodendroglia cells protect neurons against neurodegeneration

July 11, 2012

(Medical Xpress) — Johns Hopkins researchers say they have discovered that the central nervous system’s oligodendroglia cells, long believed to simply insulate nerves as they “fire” signals, are unexpectedly also vital to the survival of neurons. Damage to these insulators appears to contribute to brain injury in neurodegenerative diseases such as amyotrophic lateral sclerosis (ALS), also known as Lou Gehrig’s disease for the Yankee baseball great who died from the disease.

The discovery, described online in the journal Nature, suggests that a previously unknown — and unexpected — function of these cells is to supply nutrition to the principal brain cells, neurons. This new pathway may prove to be an important and novel therapeutic target for ALS, the researchers say, and potentially other diseases that attack the body’s nerve fibers, such as multiple sclerosis.

"More than 100 years after their discovery, we have now found a fundamentally new property in the way oligodendroglia work in the brain, laying the foundation for a new approach to try to treat debilitating neurodegenerative diseases,” says Jeffrey D. Rothstein, M.D., Ph.D., a professor of neurology and neuroscience at the Johns Hopkins University School of Medicine, and the study’s leader. “We’ve added a whole new category to what they do in the brain.”

The cells responsible for the transfer of information and electrical impulses around the body, neurons work by transferring electrical charges from neuron to neuron. Axons, the wire-like extensions of the neurons, help move the messages, in some cases over many feet, from cell to cell. Oligodendroglia insulate axons, like rubber coating around an electrical wire, to speed up the conduction of information. Axonal death is a hallmark of ALS and most other neurodegenerative disorders, Rothstein says.

Rothstein and his colleagues say the other principal brain cells, the astroglia, were believed to be primarily responsible for providing energy to neurons in the form of glucose, but their experiments show that oligodendroglia are surprisingly crucial in feeding neurons — in the form of less energy-rich lactate, without which neurons and their axons die. Lactate has long been seen as a minor player in this process, but the Johns Hopkins team says it appears to be far more important to nerve cell survival. Moreover, they found that the protein MCT1, the dominant transporter of lactate in the brain, is only found in oligodendroglia.

Rothstein says their discovery was rooted in experiments during which scientists, using mice, knocked out the gene that makes the MCT1 protein and saw axons begin to die, even though they were still getting plenty of glucose.

As part of these experiments, the researchers engineered mice whose cells would light up if they were expressing MCT1. The scientists then determined that only oligodendroglia cells lit up, showing that MCTI is located on this type of cell alone. They also knocked out the MCT1 in cell cultures and found that neurons would begin to die, but would recover when fed lactate, proving the importance of MCT1 in providing this nutritional compound. They conducted the same experiments in mice and got similar results.

Finally, the researchers turned their attention to ALS, a disease where they had recently uncovered abnormalities related to oligodendroglia. In ALS mice, they found that MCT1 was missing in brain cells well before the disease developed, and they found similar results in ALS patients. Rothstein says the findings suggest that oligodendroglia injury — specifically injury to the mechanism that produces MCT1 — may be an important event in the onset and progression of ALS.

Rothstein, who is director of the Johns Hopkins University School of Medicine’s Brain Science Institute, says he hopes further research can establish that the activation of MCT1 in people will protect axons in those with ALS and other degenerative diseases.

Provided by Johns Hopkins University School of Medicine

Source: medicalxpress.com

Filed under science neuroscience brain psychology neuron neurodegenerative disorders

15 notes

Two proteins offer a ‘clearer’ way to treat Huntington’s disease

July 11, 2012

In a paper published in the July 11 online issue of Science Translational Medicine, researchers at the University of California, San Diego School of Medicine have identified two key regulatory proteins critical to clearing away misfolded proteins that accumulate and cause the progressive, deadly neurodegeneration of Huntington’s disease (HD).

This is a human neuron. UC San Diego scientists have identified a pair of proteins that help clear away other misfolded proteins responsible for the progressive degeneration of brain cells in Huntington’s disease. Credit: UC San Diego School of Medicine

The findings explain a fundamental aspect of how HD wreaks havoc within cells and provides “clear, therapeutic opportunities,” said principal investigator Albert R. La Spada, MD, PhD, professor of cellular and molecular medicine, chief of the Division of Genetics in the Department of Pediatrics and associate director of the Institute for Genomic Medicine at UC San Diego.

"We think the implications are significant," said La Spada. "It’s a lead we can vigorously pursue, not just for Huntington’s disease, but also for similar neurodegenerative conditions like Parkinson’s disease and maybe even Alzheimer’s disease.”

In HD, an inherited mutation in the huntingtin (htt) gene results in misfolded htt proteins accumulating in certain central nervous system cells, leading to progressive deterioration of involuntary movement control, cognitive decline and psychological problems. More than 30,000 Americans have HD. There are no effective treatments currently to either cure the disease or slow its progression.

La Spada and colleagues focused on a protein called PGC-1alpha, which helps regulate the creation and operation of mitochondria, the tiny organelles that generate the fuel required for every cell to function.

"It’s all about energy," La Spada said. "Neurons have a constant, high demand for it. They’re always on the edge for maintaining adequate levels of energy production. PGC-1alpha regulates the function of transcription factors that promote the creation of mitochondria and allow them to run at full capacity.”

Previous studies by La Spada and others discovered that the mutant form of the htt gene interfered with normal levels and functioning of PGC-1alpha. “This study confirms that,” La Spada said. More surprising was the discovery that elevated levels of PGC-1alpha in a mouse model of HD virtually eliminated the problematic misfolded proteins.

Specifically, PGC-1alpha influenced expression of another protein vital to autophagy – the process in which healthy cells degrade and recycle old, unneeded or dangerous parts and products, including oxidative, damaging molecules generated by metabolism. For neurons, which must last a lifetime, the self-renewal is essential to survival.

"Mitochondria get beat up and need to be recycled," La Spada said. "PGC-1alpha drives this pathway through another protein called transcription factor EB or TFEB. We were unaware of this connection before, because TFEB is a relatively new player, though clearly emerging as a leading actor. We discovered that even without PGC-1alpha induction, TFEB can prevent htt aggregation and neurotoxicity."

In their experiments, HD mice crossbred with mice that produced greater levels of PGC-1alpha showed dramatic improvement. Production of misfolded proteins was essentially eliminated and the mice behaved normally. “Degeneration of brain cells is prevented. Neurons don’t die,” said La Spada.

PGC-1alpha and TFEB provide two new therapeutic targets for Huntington’s disease, according to La Spada. “If you can induce the bioenergetics and protein quality control pathways of nervous system cells to function properly, by activating the PGC-1alpha pathway and promoting greater TFEB function, you stand a good chance of maintaining neural function for an extended period of time. If we could achieve the level of increased function necessary to eliminate misfolded proteins, we might nip the disease process in the bud. That would go a long way toward treating this devastating condition.”

Provided by University of California - San Diego

Source: medicalxpress.com

Filed under science neuroscience brain psychology huntington protein

18 notes

Alzheimer’s plaques in PET brain scans identify future cognitive decline

July 11, 2012

Among patients with mild or no cognitive impairment, brain scans using a new radioactive dye can detect early evidence of Alzheimer’s disease that may predict future decline, according to a multi-center study led by researchers at Duke University Medical Center.

PET images using florbetapir dye to highlight beta-amyloid plaques show (A), a cognitively normal subject; (B) an amyloid-positive patient with Alzheimer’s disease; (C) a patient with mild cognitive impairment; and (D) a patient with mild cognitive impairment who progressed to dementia during the study. Credit: Slide courtesy of the journal Neurology.

The finding is published online July 11, 2012, in the journal Neurology, the medical journal of the American Academy of Neurology. It expands on smaller studies demonstrating that early detection of tell-tale plaques could be a predictive tool to help guide care and treatment decisions for patients with Alzheimer’s disease.

"Even at a short follow-up of 18 months we can see how the presence of amyloid plaques affects cognitive function," said P. Murali Doraiswamy, M.D., professor of psychiatry at Duke who co-led the study with R. Edward Coleman, M.D., professor of radiology at Duke . "Most people who come to the doctor with mild impairment really want to know the short-term prognosis and potential long-term effect."

Doraiswamy said such knowledge also has some pitfalls. There is no cure for Alzheimer’s disease, which afflicts 5.4 million people in the United States and is the sixth-leading cause of death among U.S. adults. But he said numerous drugs are being investigated, and identifying earlier disease would improve research into their potential benefits and speed new discoveries, while also enhancing care and treatment of current patients.

In the Neurology study, 151 people who had enrolled in a multi-center test of a new radioactive dye called florbetapir (Amyvid) were recruited to participate in a 36-month analysis. Of those participants, 69 had normal cognitive function at the start of the study, 51 had been diagnosed with mild impairment, and 31 had Alzheimer’s dementia.

All completed cognitive tests and underwent a brain scan using Positron Emission Tomography, or PET imaging. The technology uses radioactive tracers designed to highlight specific tissue to create a three-dimensional picture of an organ or a biological function.

The dye used in the study, florbetapir, was recently approved by the U.S. Food and Drug Administration for PET imaging of the brain to estimate beta-amyloid plaque density in patients who are being evaluated for cognitive impairment. It binds to the amyloid plaques that characterize Alzheimer’s disease, providing a window into the brain to see if the plaques have formed, and how extensively.

Patients in the study were reassessed with additional cognitive exams at 18 months and 36 months. At the 18-month point, patients with mild cognitive impairment who had PET evidence of plaque at the trial’s start worsened to a great degree on cognitive tests than patients who had no evidence of plaque at the trial’s start. Twenty-nine percent of the plaque-positive patients in this group developed Alzheimer’s dementia, compared to 10 percent who started with no plaque.

Cognitively normal patients with a plaque-positive PET scan at the start of the study also showed more mental decline at 18 months compared to those who were negative for plaque.

The study additionally found that people with negative scans reversed from minimally impaired to normal more often than people with positive PET scan, suggesting test anxiety or concentration problems could have affected their initial performance.

"For the most part we have been blind about who would progress and who wouldn’t, so this approach is a step toward having a biomarker that predicts risk of decline in people who are experiencing cognitive impairment," Doraiswamy said.

He said the study’s results provide initial data that needs to be verified by additional research. Final, 36-month data from the study has been completed and will be presented at the Alzheimer’s Association International Conference this week in Vancouver, Canada. Doraiswamy also cautioned that florbetapir is currently not approved to predict the development of dementia or other neurologic conditions and stressed that it should not be used as a screening tool in otherwise normal or minimally impaired people. Likewise, a positive scan is not necessarily diagnostic for Alzheimer’s by itself.

Provided by Duke University Medical Center

Source: medicalxpress.com

Filed under science neuroscience brain psychology alzheimer neuroimaging

42 notes

Individual differences in altruism explained by brain region involved in empathy

July 11, 2012

What can explain extreme differences in altruism among individuals, from Ebenezer Scrooge to Mother Teresa? It may all come down to variation in the size and activity of a brain region involved in appreciating others’ perspectives, according to a study published in the July 12th issue of the journal Neuron. The findings also provide a neural explanation for why altruistic tendencies remain stable over time.

The junction (yellow) between the parietal and the temporal lobes, in which the relative proportion of gray matter is significantly positively correlated with the propensity for altruistic behavior. Credit: University of Zurich

"This is the first study to link both brain anatomy and brain activation to human altruism,” says senior study author Ernst Fehr of the University of Zurich. “The findings suggest that the development of altruism through appropriate training or social practices might occur through changes in the brain structure and the neural activations that we identified in our study.”

Individuals who excel at understanding others’ intents and beliefs are more altruistic than those who struggle at this task. The ability to understand others’ perspectives has previously been associated with activity in a brain region known as the temporoparietal junction (TPJ). Based on these past findings, Fehr and his team reasoned that the size and activation of the TPJ would relate to individual differences in altruism.

In the new study, subjects underwent a brain imaging scan and played a game in which they had to decide how to split money between themselves and anonymous partners. Subjects who made more generous decisions had a larger TPJ in the right hemisphere of the brain compared with subjects who made stingy decisions.

Moreover, activity in the TPJ reflected each subject’s specific cutoff value for the maximal cost the subject was willing to endure to increase the partner’s payoff. Activity in the TPJ was higher during hard decisions—when the personal cost of an altruistic act was just below the cutoff value—than during easy decisions associated with a very low or very high cost.

"The structure of the TPJ strongly predicts an individual’s setpoint for altruistic behavior, while activity in this brain region predicts an individual’s acceptable cost for altruistic actions," says study author Yosuke Morishima of the University of Zurich. "We have elucidated the relationship between the hardware and software of human altruistic behavior."

Provided by Cell Press

Source: medicalxpress.com

Filed under science neuroscience brain psychology empathy emotion

13 notes

H1N1 Vaccine Associated With Small but Significant Risk of Guillain-Barre Syndrome

July 10th, 2012

Guillain-Barre syndrome (GBS) is usually characterized by rapidly developing motor weakness and areflexia (the absence of reflexes). “The disease is thought to be autoimmune and triggered by a stimulus of external origin.

In 1976-1977, an unusually high rate of GBS was identified in the United States following the administration of inactivated ‘swine’ influenza A(H1N1) vaccines. In 2003, the Institute of Medicine (IOM) concluded that the evidence favored acceptance of a causal relationship between the 1976 swine influenza vaccines and GBS in adults.

Studies of seasonal influenza vaccines administered in subsequent years have found small or no increased risk,” according to background information in the article. “In a more recent assessment of epidemiologic studies on seasonal influenza vaccines, experimental studies in animals, and case reports in humans, the IOM Committee to Review Adverse Effects of Vaccines concluded that the evidence was inadequate to accept or reject a causal relationship.”

Analysis of recent H1N1 vaccination data indicated a small but significant risk of GBS following influenza A(H1N1) vaccinations.

Philippe De Wals, M.D., Ph.D., of Laval University, Quebec City, Canada and colleagues conducted a study to assess the risk of GBS following pandemic influenza vaccine administration. In fall 2009 in Quebec an immunization campaign was launched against the 2009 influenza A(H1N1) pandemic strain. By the end of the year, 4.4 million residents had been vaccinated. The study included follow-up over the 6-month period of October 2009 through March 2010 for suspected and confirmed GBS cases reported by physicians, mostly neurologists, during active surveillance or identified in the provincial hospital summary discharge database. Immunization status was verified.

Over the 6-month period, 83 confirmed GBS cases were identified. Twenty-five confirmed cases had been vaccinated against 2009 influenza A(H1N1) 8 or fewer weeks before disease onset, with most (19/25) vaccinated 4 or fewer weeks before onset. Analysis of data indicated a small but significant risk of GBS following influenza A(H1N1) vaccination. The number of cases attributable to vaccination was approximately 2 per 1 million doses. The excess risk was observed only in persons 50 years of age or older.

“In Quebec, the individual risk of hospitalization following a documented influenza A(H1N1) infection was 1 per 2,500 and the risk of death was 1/73,000. The H1N1 vaccine was very effective in preventing infections and complications. It is likely that the benefits of immunization outweigh the risks,” the authors write.

Source: Neuroscience News

Filed under science neuroscience disease guillain-barre H1N1

21 notes

Potential Cause of HIV-Associated Dementia Revealed

ScienceDaily (July 10, 2012) — Researchers at Georgetown University Medical Center appear to have solved the mystery of why some patients infected with HIV, who are using antiretroviral therapy and show no signs of AIDS, develop serious depression as well as profound problems with memory, learning, and motor function. The finding might also provide a way to test people with HIV to determine their risk for developing dementia.

They say the answer, published in the July 11 issue of the Journal of Neuroscience, may ultimately lead to a therapeutic solution that helps these patients as well as others suffering from brain ailments that appear to develop through the same pathway, including those that occur in the aged.

"We believe we have discovered a general mechanism of neuronal decline that even explains what happens in some elderly folks," says the study’s lead investigator, Italo Mocchetti, Ph.D., professor and vice chair of the department of neuroscience at Georgetown University Medical Center. "The HIV-infected patients who develop this syndrome are usually quite young, but their brains act old."

The research team found that even though HIV does not infect neurons, it tries to stop the brain from producing a protein growth factor — mature brain derived neurotrophic factor (mature BDNF) — that Mocchetti says acts like “food” for brain neurons. Reduced mature BDNF results in the shortening of the axons and their branches that neurons use to connect to each other, and when they lose this communication, the neurons die.

"The loss of neurons and their connections is profound in these patients," Mocchetti says. HIV-associated dementia occurs in two to three percent of HIV-infected patients using retroviral therapies, all of who appear to be otherwise healthy, and in 30 percent of HIV-positive patients who are not on medication.

Mocchetti believes that HIV stops production of mature BDNF because that protein interferes with the ability of the virus to attack other brain cells. It does this through the potent gp120 envelope protein that sticks out from the viral shell — the same protein that hooks on to brain macrophages and microglial cells to infect them. “In earlier experiments, when we dumped gp120 into neuronal tissue culture, there was a 30-40 percent loss of neurons overnight. That makes gp120 a remarkable neurotoxin.”

This study is the product of years of work that has resulted in a string of publications. It began when Mocchetti and his colleagues were given a grant from the National Institutes on Drug Abuse to determine whether there was a connection between the use of cocaine and morphine, and dementia. (A substantial number of HIV-positive patients have been or currently are intravenous drugs users.)

They found that it was the virus that was responsible for the dementia, not the drugs, and so they set out to discover how the virus was altering neuronal function.

Their scientific break came when the researchers were able to study the blood of 130 women who were enrolled in the 17 year-old, nationwide WIHS (Women’s Interagency HIV Study, directed at Georgetown by Mary Young, M.D.), which has focused on the effects of HIV in infected females. In one seminal discovery, Mocchetti and colleagues found that when there was less BDNF in the blood, patients were at risk of developing brain abnormalities. He published this finding in 2011 in the May 15 issue of AIDS.

In this study, Mocchetti, Alessia Bachis, Ph.D., and their colleagues studied the brains of HIV-positive patients who had died, and who had developed HIV-associated dementia. They also found that neurons had shrunk, and that mature BDNF had substantially decreased.

He and his colleagues then worked out the mechanism responsible for this destruction of neurons.

Normally, neurons release a long form of BDNF known as proBDNF, and then certain enzymes, including one called furin, cleave proBDNF to produce mature BDNF, which then nurtures brain neurons. When uncut, proBDNF is toxic, leading to “synaptic simplification,” or the shortening of axons. It does this by binding to a receptor, p75NTR, that contains a death domain.

"HIV interferes with that normal process of cleaving proBDNF, resulting in neurons primarily secreting a toxic form of BDNF," Mocchetti says. The same imbalance between mature BDNF and proBDNF occurs as we age, he says, although no one knows how that happens. "The link between depression and lack of mature BDNF is also known, as is the link to issues of learning and memory. That’s why I say HIV-associated dementia resembles the aging brain."

Loss of mature BDNF has also been suggested to be a risk factor in chronic diseases such as Parkinson’s and Huntington’s diseases, Mocchetti says.

The findings suggest a possible therapeutic intervention, he adds. “One way would be to use a small molecule to block the p75NTR receptor that proBDNF uses to kill neurons. A small molecule like that could get through the blood-brain barrier.

"If this works in HIV-dementia, it may also work in other brain issues caused by proBDNF, such as aging," Mocchetti adds.

The finding also suggests that measuring proBDNF in HIV-positive patients may provide a biomarker of risk for development of dementia, he adds.

"This finding is extremely important for both basic scientists and physicians, because it suggests a new avenue to understand, and treat, a fairly widespread cause of dementia," Mocchetti says.

Source: Science Daily

Filed under science neuroscience brain psychology HIV dementia

16 notes

Blood-brain barrier less permeable in newborns than adults after acute stroke

July 10, 2012

The ability for substances to pass through the blood-brain barrier is increased after adult stroke, but not after neonatal stroke, according to a new study the UCSF that will be published July 11 in the Journal of Neuroscience.

The blood-brain barrier is selectively permeable and blocks unwanted molecules from entering into the brain. The selectivity is achieved through fine coordination in function of many transporting systems in endothelial cells, which line the interior of blood vessels, and communication between endothelial cells and several types of cells in the brain. When blood flow in an artery to the brain is blocked by a blood clot, as occurs in arterial stroke, brain energy metabolism is compromised, and ion and other transporting systems malfunction, leading to blood-brain disruption.

The new finding suggests, the researchers said, that drugs used to treat stroke need to be tailored to the specific makeup of the neonate blood-brain barrier.

"How the blood-brain barrier responds to stroke in adults and neonates currently is poorly understood,” said senior author Zinaida Vexler, PhD, director of research at the Neonatal Brain Disorders Center at the Department of Neurology at UCSF.

"The assumption has been that at birth the blood-brain barrier is immature and thus permeable and that a neonatal brain responds in the same way to injury as an adult brain. This would mean that, after a stroke, the blood-brain barrier is an open gate and different molecules could go in and out, like a floodgate,” she said. “But in neonatal stroke the situation is very different, and this study shows that the neonatal brain has the ability to protect itself by limiting blood-brain barrier permeability.”

In the study, the scientists examined the structural and functional aspects of the blood-brain barrier in live rats that had acute stroke, and found that the blood-brain barrier was markedly more intact in neonatal rats than in adult rats.

The study compared vascular responses to injury in an adult arterial stroke model and an age-appropriate model of neonatal arterial stroke using several blood-brain barrier permeability procedures. Injected molecules that remained in blood vessels under normal conditions leaked into the injured tissue of the adult rats, but the same molecules remained in vessels of neonatal injured rats within 24 hours after injury.

Importantly, the vessels remained intact for molecules of various sizes. The study also showed a different composition of several barrier structural proteins in neonates versus adults, as well as a differential response to stroke at both ages, findings that likely are to contribute to the higher resistance of the neonatal blood-brain barrier after stroke. The study also showed age-related differences in communication between circulating white blood cells and the blood-brain barrier. Neutrophils — a subtype of leukocytes — stuck to injured vasculature and entered the adult brain shortly after stroke, releasing toxic molecules and reactive oxidants and producing damage. In contrast, only a few neutrophils were able to enter the injured neonatal brain. However, pharmacological change – in communication of neutrophils with injured vessels in the neonate made injury worse.

"This study is a very critical step towards developing therapeutics, but these findings are a tip of the iceberg and a lot is still to be learned," said Vexler. "We’re moving to characterize the potential for neonatal repair. Some brain damage can’t be diagnosed early, but might show up later. Now we are experimenting with postponing certain treatments or tweaking some signaling mechanisms to see if we can enhance the capacity of the immature brain to repair itself."

Provided by University of California, San Francisco

Source: medicalxpress.com

Filed under science neuroscience brain psychology stroke

35 notes

Deaf Brain Processes Touch Differently: Lacking Sound Input, the Primary Auditory Cortex ‘Feels’ Touch

ScienceDaily (July 10, 2012) — People who are born deaf process the sense of touch differently than people who are born with normal hearing, according to research funded by the National Institutes of Health. The finding reveals how the early loss of a sense — in this case hearing — affects brain development. It adds to a growing list of discoveries that confirm the impact of experiences and outside influences in molding the developing brain.

People who are born deaf process the sense of touch differently than people who are born with normal hearing, according to research funded by the National Institutes of Health. The finding reveals how the early loss of a sense — in this case hearing — affects brain development. (Credit: © James Steidl / Fotolia)

The study is published in the July 11 online issue of The Journal of Neuroscience.

The researchers, Christina M. Karns, Ph.D., a postdoctoral research associate in the Brain Development Lab at the University of Oregon, Eugene, and her colleagues, show that deaf people use the auditory cortex to process touch stimuli and visual stimuli to a much greater degree than occurs in hearing people. The finding suggests that since the developing auditory cortex of profoundly deaf people is not exposed to sound stimuli, it adapts and takes on additional sensory processing tasks.

"This research shows how the brain is capable of rewiring in dramatic ways," said James F. Battey, Jr., M.D., Ph.D., director of the NIDCD. "This will be of great interest to other researchers who are studying multisensory processing in the brain."

Previous research, including studies performed by the lab director, Helen Neville Ph.D., has shown that people who are born deaf are better at processing peripheral vision and motion. Deaf people may process vision using many different brain regions, especially auditory areas, including the primary auditory cortex. However, no one has tackled whether vision and touch together are processed differently in deaf people, primarily because in experimental settings, it is more difficult to produce the kind of precise tactile stimuli needed to answer this question.

Dr. Karns and her colleagues developed a unique apparatus that could be worn like headphones while subjects were in a magnetic resonance imaging (MRI) scanner. Flexible tubing, connected to a compressor in another room, delivered soundless puffs of air above the right eyebrow and to the cheek below the right eye. Visual stimuli — brief pulses of light — were delivered through fiber optic cables mounted directly below the air-puff nozzle. Functional MRI was used to measure reactions to the stimuli in Heschl’s gyrus, the site of the primary auditory cortex in the human brain’s temporal lobe as well as other brain areas.

The researchers took advantage of an already known perceptual illusion in hearing people known as the auditory induced double flash, in which a single flash of light paired with two or more brief auditory events is perceived as multiple flashes of light. In their experiment, the researchers used a double puff of air as a tactile stimulus to replace the auditory stimulus, but kept the single flash of light. Subjects were also exposed to tactile stimuli and light stimuli separately and time-periods without stimuli to establish a baseline for brain activity.

Hearing people exposed to two puffs of air and one flash of light claimed only to see a single flash. However, when exposed to the same mix of stimuli, the subjects who were deaf saw two flashes. Looking at the brain scans of those who saw the double flash, the scientists observed much greater activity in Heschl’s gyrus, although not all deaf brains responded to the same degree. The deaf individuals with the highest levels of activity in the primary auditory cortex in response to touch also had the strongest response to the illusion.

"We designed this study because we thought that touch and vision might have stronger interactions in the auditory cortices of deaf people," said Dr. Karns." As it turns out, the primary auditory cortex in people who are profoundly deaf focuses on touch, even more than vision, in our experiment."

There are several ways the finding may help deaf people. For example, if touch and vision interact more in the deaf, touch could be used to help deaf students learn math or reading. The finding also has the potential to help clinicians improve the quality of hearing after cochlear implants, especially among congenitally deaf children who are implanted after the ages of 3 or 4. These children, who have lacked auditory input since birth, may struggle with comprehension and speech because their auditory cortex has taken on the processing of other senses, such as touch and vision. These changes may make it more challenging for the auditory cortex to recover auditory processing function after cochlear implantation. Being able to measure how much the auditory cortex has been taken over by other sensory processing could offer doctors insights into the kinds of intervention programs that would help the brain retrain and devote more capacity to auditory processing.

Source: Science Daily

Filed under science neuroscience brain psychology auditory cortex

6 notes

Preclinical development shows promise to treat hearing loss with Usher syndrome III

July 10, 2012

A new study published in the July 11 issue of the Journal of Neuroscience details the development of the first mouse model engineered to carry the most common mutation in Usher syndrome III causative gene (Clarin-1) in North America. Further, the research team from Case Western Reserve University School of Medicine used this new model to understand why mutation in Clarin-1 leads to hearing loss.

Usher Syndrome is an incurable genetic disease and it is the most common cause of the dual sensory deficits of deafness and blindness. It affects an estimated 50,000 Americans and many more worldwide. Clinically it is subdivided into types I-III based on the degree of deafness and the presence of balance disorder and each type is associated with distinct genes. While the progression of the disease is different with each type, all patients ultimately arrive at the same consequence. The focus of this study is Usher type III. More than a dozen genetic mutations are associated with Usher III, with ‘N48K’ mutation in Clarin-1 being the most prevalent mutation in Usher III patients in North America. Since N48K mutation originated in Europe, results of this study will be of significance to a subset of Usher III patients in Europe as well.

"With the prospective of designing and exploring therapies for Usher III patients with N48K mutation, this is a significant preclinical finding," says Kumar Alagramam, PhD, associate professor of otolaryngology head & neck surgery, genetics, and neurosciences and senior author of the manuscript. "This key understanding of how deafness occurs in Usher III is based on three years of collaborative work."

This new study reports on the first mouse model that mimicked the N48K mutation in Usher III patients. The genetically engineered mouse developed hearing loss similar to clinical presentations observed in Usher III patients with N48K mutation. This model allowed researchers to understand the pathophysiology in fine detail, as there is no non-invasive way to evaluate soft tissue pathology in the human inner ear.

The new study explains why the mutation in the N48K mutation in Clarin-1 leads to hearing loss – mislocalization of mutant protein in mechanosensory hair cells of the inner ear. Using this new Usher III model, researchers can now explore prospective therapeutics to rescue mutant protein localization and hearing. If successful, this approach could serve as a model to treat Usher I and II associated with missense mutation.

In 2009, Alagramam et al reported on the first mouse model of Usher III. The first mouse model was gene knockout mutation and most recent mouse model is a missense mutation, the first model of its kind for Usher III.

Provided by Case Western Reserve University

Source: medicalxpress.com

Filed under science neuroscience brain psychology hearing

free counters