Posts tagged science

Posts tagged science
July 11, 2012
Among patients with mild or no cognitive impairment, brain scans using a new radioactive dye can detect early evidence of Alzheimer’s disease that may predict future decline, according to a multi-center study led by researchers at Duke University Medical Center.

PET images using florbetapir dye to highlight beta-amyloid plaques show (A), a cognitively normal subject; (B) an amyloid-positive patient with Alzheimer’s disease; (C) a patient with mild cognitive impairment; and (D) a patient with mild cognitive impairment who progressed to dementia during the study. Credit: Slide courtesy of the journal Neurology.
The finding is published online July 11, 2012, in the journal Neurology, the medical journal of the American Academy of Neurology. It expands on smaller studies demonstrating that early detection of tell-tale plaques could be a predictive tool to help guide care and treatment decisions for patients with Alzheimer’s disease.
"Even at a short follow-up of 18 months we can see how the presence of amyloid plaques affects cognitive function," said P. Murali Doraiswamy, M.D., professor of psychiatry at Duke who co-led the study with R. Edward Coleman, M.D., professor of radiology at Duke . "Most people who come to the doctor with mild impairment really want to know the short-term prognosis and potential long-term effect."
Doraiswamy said such knowledge also has some pitfalls. There is no cure for Alzheimer’s disease, which afflicts 5.4 million people in the United States and is the sixth-leading cause of death among U.S. adults. But he said numerous drugs are being investigated, and identifying earlier disease would improve research into their potential benefits and speed new discoveries, while also enhancing care and treatment of current patients.
In the Neurology study, 151 people who had enrolled in a multi-center test of a new radioactive dye called florbetapir (Amyvid) were recruited to participate in a 36-month analysis. Of those participants, 69 had normal cognitive function at the start of the study, 51 had been diagnosed with mild impairment, and 31 had Alzheimer’s dementia.
All completed cognitive tests and underwent a brain scan using Positron Emission Tomography, or PET imaging. The technology uses radioactive tracers designed to highlight specific tissue to create a three-dimensional picture of an organ or a biological function.
The dye used in the study, florbetapir, was recently approved by the U.S. Food and Drug Administration for PET imaging of the brain to estimate beta-amyloid plaque density in patients who are being evaluated for cognitive impairment. It binds to the amyloid plaques that characterize Alzheimer’s disease, providing a window into the brain to see if the plaques have formed, and how extensively.
Patients in the study were reassessed with additional cognitive exams at 18 months and 36 months. At the 18-month point, patients with mild cognitive impairment who had PET evidence of plaque at the trial’s start worsened to a great degree on cognitive tests than patients who had no evidence of plaque at the trial’s start. Twenty-nine percent of the plaque-positive patients in this group developed Alzheimer’s dementia, compared to 10 percent who started with no plaque.
Cognitively normal patients with a plaque-positive PET scan at the start of the study also showed more mental decline at 18 months compared to those who were negative for plaque.
The study additionally found that people with negative scans reversed from minimally impaired to normal more often than people with positive PET scan, suggesting test anxiety or concentration problems could have affected their initial performance.
"For the most part we have been blind about who would progress and who wouldn’t, so this approach is a step toward having a biomarker that predicts risk of decline in people who are experiencing cognitive impairment," Doraiswamy said.
He said the study’s results provide initial data that needs to be verified by additional research. Final, 36-month data from the study has been completed and will be presented at the Alzheimer’s Association International Conference this week in Vancouver, Canada. Doraiswamy also cautioned that florbetapir is currently not approved to predict the development of dementia or other neurologic conditions and stressed that it should not be used as a screening tool in otherwise normal or minimally impaired people. Likewise, a positive scan is not necessarily diagnostic for Alzheimer’s by itself.
Provided by Duke University Medical Center
Source: medicalxpress.com
July 11, 2012
What can explain extreme differences in altruism among individuals, from Ebenezer Scrooge to Mother Teresa? It may all come down to variation in the size and activity of a brain region involved in appreciating others’ perspectives, according to a study published in the July 12th issue of the journal Neuron. The findings also provide a neural explanation for why altruistic tendencies remain stable over time.

The junction (yellow) between the parietal and the temporal lobes, in which the relative proportion of gray matter is significantly positively correlated with the propensity for altruistic behavior. Credit: University of Zurich
"This is the first study to link both brain anatomy and brain activation to human altruism,” says senior study author Ernst Fehr of the University of Zurich. “The findings suggest that the development of altruism through appropriate training or social practices might occur through changes in the brain structure and the neural activations that we identified in our study.”
Individuals who excel at understanding others’ intents and beliefs are more altruistic than those who struggle at this task. The ability to understand others’ perspectives has previously been associated with activity in a brain region known as the temporoparietal junction (TPJ). Based on these past findings, Fehr and his team reasoned that the size and activation of the TPJ would relate to individual differences in altruism.
In the new study, subjects underwent a brain imaging scan and played a game in which they had to decide how to split money between themselves and anonymous partners. Subjects who made more generous decisions had a larger TPJ in the right hemisphere of the brain compared with subjects who made stingy decisions.
Moreover, activity in the TPJ reflected each subject’s specific cutoff value for the maximal cost the subject was willing to endure to increase the partner’s payoff. Activity in the TPJ was higher during hard decisions—when the personal cost of an altruistic act was just below the cutoff value—than during easy decisions associated with a very low or very high cost.
"The structure of the TPJ strongly predicts an individual’s setpoint for altruistic behavior, while activity in this brain region predicts an individual’s acceptable cost for altruistic actions," says study author Yosuke Morishima of the University of Zurich. "We have elucidated the relationship between the hardware and software of human altruistic behavior."
Provided by Cell Press
Source: medicalxpress.com
July 10th, 2012
Guillain-Barre syndrome (GBS) is usually characterized by rapidly developing motor weakness and areflexia (the absence of reflexes). “The disease is thought to be autoimmune and triggered by a stimulus of external origin.
In 1976-1977, an unusually high rate of GBS was identified in the United States following the administration of inactivated ‘swine’ influenza A(H1N1) vaccines. In 2003, the Institute of Medicine (IOM) concluded that the evidence favored acceptance of a causal relationship between the 1976 swine influenza vaccines and GBS in adults.
Studies of seasonal influenza vaccines administered in subsequent years have found small or no increased risk,” according to background information in the article. “In a more recent assessment of epidemiologic studies on seasonal influenza vaccines, experimental studies in animals, and case reports in humans, the IOM Committee to Review Adverse Effects of Vaccines concluded that the evidence was inadequate to accept or reject a causal relationship.”

Analysis of recent H1N1 vaccination data indicated a small but significant risk of GBS following influenza A(H1N1) vaccinations.
Philippe De Wals, M.D., Ph.D., of Laval University, Quebec City, Canada and colleagues conducted a study to assess the risk of GBS following pandemic influenza vaccine administration. In fall 2009 in Quebec an immunization campaign was launched against the 2009 influenza A(H1N1) pandemic strain. By the end of the year, 4.4 million residents had been vaccinated. The study included follow-up over the 6-month period of October 2009 through March 2010 for suspected and confirmed GBS cases reported by physicians, mostly neurologists, during active surveillance or identified in the provincial hospital summary discharge database. Immunization status was verified.
Over the 6-month period, 83 confirmed GBS cases were identified. Twenty-five confirmed cases had been vaccinated against 2009 influenza A(H1N1) 8 or fewer weeks before disease onset, with most (19/25) vaccinated 4 or fewer weeks before onset. Analysis of data indicated a small but significant risk of GBS following influenza A(H1N1) vaccination. The number of cases attributable to vaccination was approximately 2 per 1 million doses. The excess risk was observed only in persons 50 years of age or older.
“In Quebec, the individual risk of hospitalization following a documented influenza A(H1N1) infection was 1 per 2,500 and the risk of death was 1/73,000. The H1N1 vaccine was very effective in preventing infections and complications. It is likely that the benefits of immunization outweigh the risks,” the authors write.
Source: Neuroscience News
ScienceDaily (July 10, 2012) — Researchers at Georgetown University Medical Center appear to have solved the mystery of why some patients infected with HIV, who are using antiretroviral therapy and show no signs of AIDS, develop serious depression as well as profound problems with memory, learning, and motor function. The finding might also provide a way to test people with HIV to determine their risk for developing dementia.
They say the answer, published in the July 11 issue of the Journal of Neuroscience, may ultimately lead to a therapeutic solution that helps these patients as well as others suffering from brain ailments that appear to develop through the same pathway, including those that occur in the aged.
"We believe we have discovered a general mechanism of neuronal decline that even explains what happens in some elderly folks," says the study’s lead investigator, Italo Mocchetti, Ph.D., professor and vice chair of the department of neuroscience at Georgetown University Medical Center. "The HIV-infected patients who develop this syndrome are usually quite young, but their brains act old."
The research team found that even though HIV does not infect neurons, it tries to stop the brain from producing a protein growth factor — mature brain derived neurotrophic factor (mature BDNF) — that Mocchetti says acts like “food” for brain neurons. Reduced mature BDNF results in the shortening of the axons and their branches that neurons use to connect to each other, and when they lose this communication, the neurons die.
"The loss of neurons and their connections is profound in these patients," Mocchetti says. HIV-associated dementia occurs in two to three percent of HIV-infected patients using retroviral therapies, all of who appear to be otherwise healthy, and in 30 percent of HIV-positive patients who are not on medication.
Mocchetti believes that HIV stops production of mature BDNF because that protein interferes with the ability of the virus to attack other brain cells. It does this through the potent gp120 envelope protein that sticks out from the viral shell — the same protein that hooks on to brain macrophages and microglial cells to infect them. “In earlier experiments, when we dumped gp120 into neuronal tissue culture, there was a 30-40 percent loss of neurons overnight. That makes gp120 a remarkable neurotoxin.”
This study is the product of years of work that has resulted in a string of publications. It began when Mocchetti and his colleagues were given a grant from the National Institutes on Drug Abuse to determine whether there was a connection between the use of cocaine and morphine, and dementia. (A substantial number of HIV-positive patients have been or currently are intravenous drugs users.)
They found that it was the virus that was responsible for the dementia, not the drugs, and so they set out to discover how the virus was altering neuronal function.
Their scientific break came when the researchers were able to study the blood of 130 women who were enrolled in the 17 year-old, nationwide WIHS (Women’s Interagency HIV Study, directed at Georgetown by Mary Young, M.D.), which has focused on the effects of HIV in infected females. In one seminal discovery, Mocchetti and colleagues found that when there was less BDNF in the blood, patients were at risk of developing brain abnormalities. He published this finding in 2011 in the May 15 issue of AIDS.
In this study, Mocchetti, Alessia Bachis, Ph.D., and their colleagues studied the brains of HIV-positive patients who had died, and who had developed HIV-associated dementia. They also found that neurons had shrunk, and that mature BDNF had substantially decreased.
He and his colleagues then worked out the mechanism responsible for this destruction of neurons.
Normally, neurons release a long form of BDNF known as proBDNF, and then certain enzymes, including one called furin, cleave proBDNF to produce mature BDNF, which then nurtures brain neurons. When uncut, proBDNF is toxic, leading to “synaptic simplification,” or the shortening of axons. It does this by binding to a receptor, p75NTR, that contains a death domain.
"HIV interferes with that normal process of cleaving proBDNF, resulting in neurons primarily secreting a toxic form of BDNF," Mocchetti says. The same imbalance between mature BDNF and proBDNF occurs as we age, he says, although no one knows how that happens. "The link between depression and lack of mature BDNF is also known, as is the link to issues of learning and memory. That’s why I say HIV-associated dementia resembles the aging brain."
Loss of mature BDNF has also been suggested to be a risk factor in chronic diseases such as Parkinson’s and Huntington’s diseases, Mocchetti says.
The findings suggest a possible therapeutic intervention, he adds. “One way would be to use a small molecule to block the p75NTR receptor that proBDNF uses to kill neurons. A small molecule like that could get through the blood-brain barrier.
"If this works in HIV-dementia, it may also work in other brain issues caused by proBDNF, such as aging," Mocchetti adds.
The finding also suggests that measuring proBDNF in HIV-positive patients may provide a biomarker of risk for development of dementia, he adds.
"This finding is extremely important for both basic scientists and physicians, because it suggests a new avenue to understand, and treat, a fairly widespread cause of dementia," Mocchetti says.
Source: Science Daily
July 10, 2012
The ability for substances to pass through the blood-brain barrier is increased after adult stroke, but not after neonatal stroke, according to a new study the UCSF that will be published July 11 in the Journal of Neuroscience.
The blood-brain barrier is selectively permeable and blocks unwanted molecules from entering into the brain. The selectivity is achieved through fine coordination in function of many transporting systems in endothelial cells, which line the interior of blood vessels, and communication between endothelial cells and several types of cells in the brain. When blood flow in an artery to the brain is blocked by a blood clot, as occurs in arterial stroke, brain energy metabolism is compromised, and ion and other transporting systems malfunction, leading to blood-brain disruption.
The new finding suggests, the researchers said, that drugs used to treat stroke need to be tailored to the specific makeup of the neonate blood-brain barrier.
"How the blood-brain barrier responds to stroke in adults and neonates currently is poorly understood,” said senior author Zinaida Vexler, PhD, director of research at the Neonatal Brain Disorders Center at the Department of Neurology at UCSF.
"The assumption has been that at birth the blood-brain barrier is immature and thus permeable and that a neonatal brain responds in the same way to injury as an adult brain. This would mean that, after a stroke, the blood-brain barrier is an open gate and different molecules could go in and out, like a floodgate,” she said. “But in neonatal stroke the situation is very different, and this study shows that the neonatal brain has the ability to protect itself by limiting blood-brain barrier permeability.”
In the study, the scientists examined the structural and functional aspects of the blood-brain barrier in live rats that had acute stroke, and found that the blood-brain barrier was markedly more intact in neonatal rats than in adult rats.
The study compared vascular responses to injury in an adult arterial stroke model and an age-appropriate model of neonatal arterial stroke using several blood-brain barrier permeability procedures. Injected molecules that remained in blood vessels under normal conditions leaked into the injured tissue of the adult rats, but the same molecules remained in vessels of neonatal injured rats within 24 hours after injury.
Importantly, the vessels remained intact for molecules of various sizes. The study also showed a different composition of several barrier structural proteins in neonates versus adults, as well as a differential response to stroke at both ages, findings that likely are to contribute to the higher resistance of the neonatal blood-brain barrier after stroke. The study also showed age-related differences in communication between circulating white blood cells and the blood-brain barrier. Neutrophils — a subtype of leukocytes — stuck to injured vasculature and entered the adult brain shortly after stroke, releasing toxic molecules and reactive oxidants and producing damage. In contrast, only a few neutrophils were able to enter the injured neonatal brain. However, pharmacological change – in communication of neutrophils with injured vessels in the neonate made injury worse.
"This study is a very critical step towards developing therapeutics, but these findings are a tip of the iceberg and a lot is still to be learned," said Vexler. "We’re moving to characterize the potential for neonatal repair. Some brain damage can’t be diagnosed early, but might show up later. Now we are experimenting with postponing certain treatments or tweaking some signaling mechanisms to see if we can enhance the capacity of the immature brain to repair itself."
Provided by University of California, San Francisco
Source: medicalxpress.com
ScienceDaily (July 10, 2012) — People who are born deaf process the sense of touch differently than people who are born with normal hearing, according to research funded by the National Institutes of Health. The finding reveals how the early loss of a sense — in this case hearing — affects brain development. It adds to a growing list of discoveries that confirm the impact of experiences and outside influences in molding the developing brain.

People who are born deaf process the sense of touch differently than people who are born with normal hearing, according to research funded by the National Institutes of Health. The finding reveals how the early loss of a sense — in this case hearing — affects brain development. (Credit: © James Steidl / Fotolia)
The study is published in the July 11 online issue of The Journal of Neuroscience.
The researchers, Christina M. Karns, Ph.D., a postdoctoral research associate in the Brain Development Lab at the University of Oregon, Eugene, and her colleagues, show that deaf people use the auditory cortex to process touch stimuli and visual stimuli to a much greater degree than occurs in hearing people. The finding suggests that since the developing auditory cortex of profoundly deaf people is not exposed to sound stimuli, it adapts and takes on additional sensory processing tasks.
"This research shows how the brain is capable of rewiring in dramatic ways," said James F. Battey, Jr., M.D., Ph.D., director of the NIDCD. "This will be of great interest to other researchers who are studying multisensory processing in the brain."
Previous research, including studies performed by the lab director, Helen Neville Ph.D., has shown that people who are born deaf are better at processing peripheral vision and motion. Deaf people may process vision using many different brain regions, especially auditory areas, including the primary auditory cortex. However, no one has tackled whether vision and touch together are processed differently in deaf people, primarily because in experimental settings, it is more difficult to produce the kind of precise tactile stimuli needed to answer this question.
Dr. Karns and her colleagues developed a unique apparatus that could be worn like headphones while subjects were in a magnetic resonance imaging (MRI) scanner. Flexible tubing, connected to a compressor in another room, delivered soundless puffs of air above the right eyebrow and to the cheek below the right eye. Visual stimuli — brief pulses of light — were delivered through fiber optic cables mounted directly below the air-puff nozzle. Functional MRI was used to measure reactions to the stimuli in Heschl’s gyrus, the site of the primary auditory cortex in the human brain’s temporal lobe as well as other brain areas.
The researchers took advantage of an already known perceptual illusion in hearing people known as the auditory induced double flash, in which a single flash of light paired with two or more brief auditory events is perceived as multiple flashes of light. In their experiment, the researchers used a double puff of air as a tactile stimulus to replace the auditory stimulus, but kept the single flash of light. Subjects were also exposed to tactile stimuli and light stimuli separately and time-periods without stimuli to establish a baseline for brain activity.
Hearing people exposed to two puffs of air and one flash of light claimed only to see a single flash. However, when exposed to the same mix of stimuli, the subjects who were deaf saw two flashes. Looking at the brain scans of those who saw the double flash, the scientists observed much greater activity in Heschl’s gyrus, although not all deaf brains responded to the same degree. The deaf individuals with the highest levels of activity in the primary auditory cortex in response to touch also had the strongest response to the illusion.
"We designed this study because we thought that touch and vision might have stronger interactions in the auditory cortices of deaf people," said Dr. Karns." As it turns out, the primary auditory cortex in people who are profoundly deaf focuses on touch, even more than vision, in our experiment."
There are several ways the finding may help deaf people. For example, if touch and vision interact more in the deaf, touch could be used to help deaf students learn math or reading. The finding also has the potential to help clinicians improve the quality of hearing after cochlear implants, especially among congenitally deaf children who are implanted after the ages of 3 or 4. These children, who have lacked auditory input since birth, may struggle with comprehension and speech because their auditory cortex has taken on the processing of other senses, such as touch and vision. These changes may make it more challenging for the auditory cortex to recover auditory processing function after cochlear implantation. Being able to measure how much the auditory cortex has been taken over by other sensory processing could offer doctors insights into the kinds of intervention programs that would help the brain retrain and devote more capacity to auditory processing.
Source: Science Daily
July 10, 2012
A new study published in the July 11 issue of the Journal of Neuroscience details the development of the first mouse model engineered to carry the most common mutation in Usher syndrome III causative gene (Clarin-1) in North America. Further, the research team from Case Western Reserve University School of Medicine used this new model to understand why mutation in Clarin-1 leads to hearing loss.
Usher Syndrome is an incurable genetic disease and it is the most common cause of the dual sensory deficits of deafness and blindness. It affects an estimated 50,000 Americans and many more worldwide. Clinically it is subdivided into types I-III based on the degree of deafness and the presence of balance disorder and each type is associated with distinct genes. While the progression of the disease is different with each type, all patients ultimately arrive at the same consequence. The focus of this study is Usher type III. More than a dozen genetic mutations are associated with Usher III, with ‘N48K’ mutation in Clarin-1 being the most prevalent mutation in Usher III patients in North America. Since N48K mutation originated in Europe, results of this study will be of significance to a subset of Usher III patients in Europe as well.
"With the prospective of designing and exploring therapies for Usher III patients with N48K mutation, this is a significant preclinical finding," says Kumar Alagramam, PhD, associate professor of otolaryngology head & neck surgery, genetics, and neurosciences and senior author of the manuscript. "This key understanding of how deafness occurs in Usher III is based on three years of collaborative work."
This new study reports on the first mouse model that mimicked the N48K mutation in Usher III patients. The genetically engineered mouse developed hearing loss similar to clinical presentations observed in Usher III patients with N48K mutation. This model allowed researchers to understand the pathophysiology in fine detail, as there is no non-invasive way to evaluate soft tissue pathology in the human inner ear.
The new study explains why the mutation in the N48K mutation in Clarin-1 leads to hearing loss – mislocalization of mutant protein in mechanosensory hair cells of the inner ear. Using this new Usher III model, researchers can now explore prospective therapeutics to rescue mutant protein localization and hearing. If successful, this approach could serve as a model to treat Usher I and II associated with missense mutation.
In 2009, Alagramam et al reported on the first mouse model of Usher III. The first mouse model was gene knockout mutation and most recent mouse model is a missense mutation, the first model of its kind for Usher III.
Provided by Case Western Reserve University
Source: medicalxpress.com
July 9, 2012
In the last ten years, a new understanding of pediatric brain injury and recovery has emerged. Professionals now understand that recovery may be a lifelong process for the child’s entire circle of family, friends, and healthcare providers. The latest efforts to advance medical and rehabilitative services to move children from medical care and rehabilitation to community reintegration are discussed by the leading experts in a recently published special issue of NeuroRehabilitation.
“Recovery extends well beyond the technical period of rehabilitation,” say guest editors and noted authorities Peter D. Patrick, PhD, MS, Associate Professor Emeritus of the University of Virginia School of Medicine in Charlottesville, and Ronald C. Savage, EdD, Chairman, North American Brain Injury Society and International Pediatric Brain Injury Society. “Children, adolescents, and families struggle to regain the momentum of their life so as to reduce problems, increase opportunity, and support increased participation in work, play, home, and relationships.”
Neural plasticity introduces unknown challenges in the care of the recovering brain, and the issue addresses the most challenging and demanding medical conditions that children may confront following severe brain injury. However, children do most of their recovery at home, in school, and in the community, beyond medical surveillance. “Family-centered” approaches to developing interventions are emerging. For example, Dr. Damith T. Woods and colleagues report on a novel telephone support program to help parents manage challenging behavior associated with brain injury.
Children and adolescents with brain injuries have difficulty adjusting to their injuries and altered abilities, and frequently suffer from low self-esteem and loss of confidence. A study by Carol A. Hawley finds that children with traumatic brain injury have significantly lower self-esteem than normal children, and recommends that rehabilitation strategies promote a sense of self-worth.
Re-entry into school is a major milestone of recovery and the issue highlights a number of efforts to help children improve and return to a positive developmental trajectory. An article by Beth Wicks describes an innovative program in Britain that looks at “education as rehabilitation,” translating successful adult vocational programs into educational rehabilitation programs for children. Lucia Willadino Braga and colleagues report on a program based on cooperative learning that helped preadolescents with acquired brain injury develop metacognitive strategies and improve self-concept, thereby helping empower the preadolescents in their social relationships.
"Over the years and in multiple places around the world, innovative and creative efforts have slowly revealed effective interventions for recovery," comment Dr. Patrick and Dr. Savage. "Increasingly the interventions are evidence-based. This issue is a contribution to the effort to improve outcomes for children and families."
Provided by IOS Press
Source: medicalxpress.com
ScienceDaily (July 9, 2012) — Scientists showed in mice that disabling a gene linked to a common pediatric tumor disorder, neurofibromatosis type 1 (NF1), made stem cells from one part of the brain proliferate rapidly. But the same genetic deficit had no effect on stem cells from another brain region.
The results can be explained by differences in the way stem cells from these regions of the brain respond to cancer-causing genetic changes.
NF1 is among the world’s most common genetic disorders, occurring in about one of every 3,000 births. It causes a wide range of symptoms, including brain tumors, learning disabilities and attention deficits.
Brain tumors in children with NF1 typically arise in the optic nerve and do not necessarily require treatment. If optic gliomas keep growing, though, they can threaten the child’s vision. By learning more about the many factors that contribute to NF1 tumor formation, scientists hope to develop more effective treatments.
"To improve therapy, we need to develop better ways to identify and group tumors based not just on the way they look under the microscope, but also on innate properties of their stem cell progenitors," says David H. Gutmann, MD, PhD, the Donald O. Schnuck Family Professor of Neurology.
The study appears July 9 in Cancer Cell. Gutmann also is the director of the Washington University Neurofibromatosis Center.
In the new study, researchers compared brain stem cells from two primary sources: the third ventricle, located in the midbrain, and the nearby lateral ventricles. Before birth and for a time afterward, both of these areas in the brain are lined with growing stem cells.
First author Da Yong Lee, PhD, a postdoctoral research associate, showed that the cells lining both ventricles are true stem cells capable of becoming nerve and support cells (glia) in the brain. Next, she conducted a detailed analysis of gene expression in both stem cell types.
"There are night-and-day differences between these two groups of stem cells," Gutmann says. "These results show that stem cells are not the same everywhere in the brain, which has real consequences for human neurologic disease."
The third ventricle is close to the optic chiasm, the point where the optic nerves cross and optic gliomas develop in NF1 patients. Lee and Gutmann postulated that stem cells from this ventricle might be the source of progenitor cells that can become gliomas in children with NF1.
To test the theory, they disabled the Nf1 gene in neural stem cells from the third and lateral ventricles in the mice. This same gene is mutated in patients with NF1, increasing their risk of developing tumors.
Lee found that loss of Nf1 activity had little effect on stem cells from the lateral ventricle, but stem cells from the third ventricle began to divide rapidly, a change that puts them closer to becoming tumors.
The third ventricle usually stops supplying stem cells to the brain shortly after birth. When researchers inactivated the Nf1 gene before the third ventricle closed, the mice developed optic gliomas. When they waited until the third ventricle had closed to inactivate the Nf1 gene, gliomas did not develop.
Gutmann plans further studies to determine whether all NF1-related optic gliomas form in cells descended from the third ventricle. He suspects that additional factors are necessary for optic gliomas to form in cooperation with Nf1 gene loss in third-ventricle stem cells.
"We have to recognize that cancers which appear very similar actually represent a collection of quite different diseases," he says. "Tumors are like us — they’re defined by where they live, what their families are like, the traumas they experience growing up, and a variety of other factors. If we can better understand the interplay of these factors, we’ll be able to develop treatments that are much more likely to succeed, because they’ll target what is unique about a specific patient’s tumor."
Source: Science Daily
ScienceDaily (July 9, 2012) — Nearly a third of adults with the most common type of brain cancer develop recurrent, invasive tumors after being treated with a drug called bevacizumab. The molecular underpinnings behind these detrimental effects have now been published by Cell Press in the July issue of Cancer Cell. The findings reveal a new treatment strategy that could reduce tumor invasiveness and improve survival in these drug-resistant patients.
"Understanding how and why these tumors adopt this invasive behavior is critical to being able to prevent this recurrence pattern and maximizing the benefits of bevacizumab," says study author Kan Lu of the University of California, San Francisco (UCSF).
Glioblastoma multiforme (GBM) is the most aggressive type of tumor originating in the brain. GBM tumors express high levels of vascular endothelial growth factor (VEGF), a protein that promotes the growth of new blood vessels that provide nutrients that allow tumors to expand. In 2009, the Food and Drug Administration approved bevacizumab, a VEGF inhibitor, for GBM patients who don’t respond to first-line therapies. Although the drug is initially effective, up to 30% of patients develop tumors that infiltrate deep into the brain, making surgery and treatment difficult.
To study how bevacizumab can lead to adverse effects, senior study author Gabriele Bergers of UCSF and her collaborators focused on hepatocyte growth factor (HGF), a protein that controls the growth and movement of cells, because they previously found a link between VEGF and HGF in GBM cells. In the new study, they found that VEGF inhibits the migration of GBM cells by decreasing HGF signaling through its receptor MET. Moreover, tumors were much less invasive — and survival improved — in mice with GBM tumors lacking both VEGF and MET rather than just VEGF alone. The results suggest that MET plays a critical role in GBM invasion when VEGF is blocked.
"These findings provide a rationale for therapeutically combining VEGF and MET inhibition so that patients can benefit from bevacizumab without developing more invasive tumors," Lu says. Because the VEGF and HGF/MET signaling pathways are active in a variety of tumors, this combined treatment strategy may also be applied to other types of cancer.
Source: Science Daily