Neuroscience

Articles and news from the latest research reports.

Posts tagged neuroscience

46 notes

Gladstone Scientists Identify Key Biological Mechanism in Multiple Sclerosis

Scientists at the Gladstone Institutes have defined for the first time a key underlying process implicated in multiple sclerosis (MS)—a disease that causes progressive and irreversible damage to nerve cells in the brain and spinal cord. This discovery offers new hope for the millions who suffer from this debilitating disease for which there is no cure.

Researchers in the laboratory of Gladstone Investigator Katerina Akassoglou, PhD, have identified in animal models precisely how a protein that seeps from the blood into the brain sets off a response that, over time, causes the nerve cell damage that is a key indicator of MS. These findings, which are reported in the latest issue of Nature Communications, lay the groundwork for much-needed therapies to treat this disease.

(Source: gladstoneinstitutes.org)

Read more …

Filed under MS nerve cells blood protein in vivo imaging fibrinogen microglia neuroscience science

505 notes


Preschoolers at play show science skills
When kids incessantly ask “Why?,” mess around in the dirt and run their hands over everything within reach, they’re not just being kids. It turns out they’re also being scientists.
Until recently, preschoolers were widely believed to be irrational thinkers. For most of the 20th century, the prevailing theory pioneered by cognitive development expert Jean Piaget held that children roughly ages 2 through 7 cannot understand concrete logic or other people’s perspectives.
Although young children are the only ones who truly know what they ponder, research conducted over the past decade has led many psychologists to see infants and toddlers as, in fact, capable of thinking logically and abstractly.
"The main thing is that they’re drawing conclusions from data and evidence and experiences the same way scientists are - by making hypotheses, testing them, analyzing statistics and even doing experiments, even though when they do experiments, it’s called ‘getting into everything,’ " said Alison Gopnik, a UC Berkeley psychology professor and one of the field’s leading experts.
Better understanding of how children learn about the world could have important implications for their formal schooling, Gopnik argued in a recent paper published in the journal Science, which summarized studies by her and other researchers.

Preschoolers at play show science skills

When kids incessantly ask “Why?,” mess around in the dirt and run their hands over everything within reach, they’re not just being kids. It turns out they’re also being scientists.

Until recently, preschoolers were widely believed to be irrational thinkers. For most of the 20th century, the prevailing theory pioneered by cognitive development expert Jean Piaget held that children roughly ages 2 through 7 cannot understand concrete logic or other people’s perspectives.

Although young children are the only ones who truly know what they ponder, research conducted over the past decade has led many psychologists to see infants and toddlers as, in fact, capable of thinking logically and abstractly.

"The main thing is that they’re drawing conclusions from data and evidence and experiences the same way scientists are - by making hypotheses, testing them, analyzing statistics and even doing experiments, even though when they do experiments, it’s called ‘getting into everything,’ " said Alison Gopnik, a UC Berkeley psychology professor and one of the field’s leading experts.

Better understanding of how children learn about the world could have important implications for their formal schooling, Gopnik argued in a recent paper published in the journal Science, which summarized studies by her and other researchers.

Filed under children thinking learning science probabilistic models neuroscience psychology

102 notes

New biomaterials promote neuroregeneration after a brain injury

Professor José Miguel Soria, a member of the Institute of Biomedical Sciences, Universidad CEU Cardenal Herrera, has co-directed with Professor Manuel Monleón of the Universitat Politècnica de València a study on the compatibility of polymeric biomaterials in the brain and its effectiveness to favour neuroregeneration in areas with some kind of damage or brain injury.  

The research carried out has shown that these types of implants, made of a biocompatible synthetic material, are colonized within two months by neural progenitor cells and irrigated by new blood vessels. This allows the generation, within these structures, of new neurons and glia, capable of repairing injured brain tissue caused by trauma, stroke or neurodegenerative disease, among other causes.

The synthetic structures used in this study are made with a porous and biocompatible polymeric material called acrylate copolymer. In the first phase of the project, the structures have been studied in vitro by implanting them into neural tissue, and subsequently also in vivo, when implanted in two areas of the adult rat brain: the cerebral cortex and the subventricular zone, the most important source of generation of adult neural stem cells.

The study has confirmed the high biocompatibility of polymeric materials, such as acrylate copolymer, with brain tissue and opens new possibilities of the effectiveness of the implementation of these structures in the brain, seeking optimum location for developing regenerative strategies of the central nervous system.

Furthermore, the results are particularly relevant when one considers that in the adult brain neuroregeneration capacity is more limited than in younger individuals and that the main impediment for this is the lack of revascularization of damaged tissue, something that the biomaterial studied has shown to favour.

(Source: alphagalileo.org)

Filed under brain brain injury biomaterials neuroregeneration neuroscience science

48 notes


Brain Cooling to Treat Epilepsy Moves Closer to Human Application
Neuroscientists from Japan’s Yamaguchi University today reported during the 66th annual scientific meeting of the American Epilepsy Society (AES) that chronic focal brain cooling suppresses seizures during wakefulness and achieves the effect without significantly affecting brain function. Their research, and that of others in the field, provides critical evidence that this approach to seizure control has reached a stage where testing in humans will soon be possible.
Focal brain cooling is well established as an effective method for suppressing seizures. But the technology for creating a practical device with potential clinical application has only recently become available and tested in rodents. More evidence from large animals and humans is needed prior to testing in clinical trials for drug-resistant epilepsy.
The Yamaguchi researchers implanted two feline and two non-human primates with a titanium cooling plate, or heat exchanger. The brain cooling device was placed in contact with the brain surface over cortex areas responsible for movement and sensation. Seizures were then induced in the motor cortex. Brain wave recordings to assess seizure activity and temperature recordings were performed under wakefulness.
According to Masami Fujii, M.D.,Ph.D., and Takao Inoue, Ph.D., and Michiyasu Suzuki, M.D., Ph.D., who presented the report, seizure discharges were significantly suppressed at 15˚C (59˚F).
“The results of our study suggest that focal brain cooling has a strong effect to suppress the epileptiform seizures under the awake condition,” Dr. Fujii said. “Moreover, implantation of the device for at least five months did not result in detrimental changes in brain tissue subjected to cooling compared to tissue from a similar site in the opposing hemisphere.”

Brain Cooling to Treat Epilepsy Moves Closer to Human Application

Neuroscientists from Japan’s Yamaguchi University today reported during the 66th annual scientific meeting of the American Epilepsy Society (AES) that chronic focal brain cooling suppresses seizures during wakefulness and achieves the effect without significantly affecting brain function. Their research, and that of others in the field, provides critical evidence that this approach to seizure control has reached a stage where testing in humans will soon be possible.

Focal brain cooling is well established as an effective method for suppressing seizures. But the technology for creating a practical device with potential clinical application has only recently become available and tested in rodents. More evidence from large animals and humans is needed prior to testing in clinical trials for drug-resistant epilepsy.

The Yamaguchi researchers implanted two feline and two non-human primates with a titanium cooling plate, or heat exchanger. The brain cooling device was placed in contact with the brain surface over cortex areas responsible for movement and sensation. Seizures were then induced in the motor cortex. Brain wave recordings to assess seizure activity and temperature recordings were performed under wakefulness.

According to Masami Fujii, M.D.,Ph.D., and Takao Inoue, Ph.D., and Michiyasu Suzuki, M.D., Ph.D., who presented the report, seizure discharges were significantly suppressed at 15˚C (59˚F).

“The results of our study suggest that focal brain cooling has a strong effect to suppress the epileptiform seizures under the awake condition,” Dr. Fujii said. “Moreover, implantation of the device for at least five months did not result in detrimental changes in brain tissue subjected to cooling compared to tissue from a similar site in the opposing hemisphere.”

Filed under brain brain cooling epilepsy seizures neuroscience science

137 notes


Extended sleep reduces pain sensitivity
A new study suggests that extending nightly sleep in mildly sleepy, healthy adults increases daytime alertness and reduces pain sensitivity.
"Our results suggest the importance of adequate sleep in various chronic pain conditions or in preparation for elective surgical procedures," said Timothy Roehrs, PhD, the study’s principal investigator and lead author. "We were surprised by the magnitude of the reduction in pain sensitivity, when compared to the reduction produced by taking codeine."
The study, appearing in the December issue of the journal SLEEP, involved 18 healthy, pain-free, sleepy volunteers. They were randomly assigned to four nights of either maintaining their habitual sleep time or extending their sleep time by spending 10 hours in bed per night. Objective daytime sleepiness was measured using the multiple sleep latency test (MSLT), and pain sensitivity was assessed using a radiant heat stimulus.
Results show that the extended sleep group slept 1.8 hours more per night than the habitual sleep group. This nightly increase in sleep time during the four experimental nights was correlated with increased daytime alertness, which was associated with less pain sensitivity.
In the extended sleep group, the length of time before participants removed their finger from a radiant heat source increased by 25 percent, reflecting a reduction in pain sensitivity. The authors report that the magnitude of this increase in finger withdrawal latency is greater than the effect found in a previous study of 60 mg of codeine.
According to the authors, this is the first study to show that extended sleep in mildly, chronically sleep deprived volunteers reduces their pain sensitivity. The results, combined with data from previous research, suggest that increased pain sensitivity in sleepy individuals is the result of their underlying sleepiness.

Extended sleep reduces pain sensitivity

A new study suggests that extending nightly sleep in mildly sleepy, healthy adults increases daytime alertness and reduces pain sensitivity.

"Our results suggest the importance of adequate sleep in various chronic pain conditions or in preparation for elective surgical procedures," said Timothy Roehrs, PhD, the study’s principal investigator and lead author. "We were surprised by the magnitude of the reduction in pain sensitivity, when compared to the reduction produced by taking codeine."

The study, appearing in the December issue of the journal SLEEP, involved 18 healthy, pain-free, sleepy volunteers. They were randomly assigned to four nights of either maintaining their habitual sleep time or extending their sleep time by spending 10 hours in bed per night. Objective daytime sleepiness was measured using the multiple sleep latency test (MSLT), and pain sensitivity was assessed using a radiant heat stimulus.

Results show that the extended sleep group slept 1.8 hours more per night than the habitual sleep group. This nightly increase in sleep time during the four experimental nights was correlated with increased daytime alertness, which was associated with less pain sensitivity.

In the extended sleep group, the length of time before participants removed their finger from a radiant heat source increased by 25 percent, reflecting a reduction in pain sensitivity. The authors report that the magnitude of this increase in finger withdrawal latency is greater than the effect found in a previous study of 60 mg of codeine.

According to the authors, this is the first study to show that extended sleep in mildly, chronically sleep deprived volunteers reduces their pain sensitivity. The results, combined with data from previous research, suggest that increased pain sensitivity in sleepy individuals is the result of their underlying sleepiness.

Filed under sleep pain sensitivity alertness chronic pain neuroscience psychology science

134 notes


Scientists create road map to metabolic reprogramming for aging
In efforts to understand what influences life span, cancer and aging, scientists are building road maps to navigate and learn about cells at the molecular level.
To survey previously uncharted territory, a team of researchers at UW-Madison has created an “atlas” that maps more than 1,500 unique landmarks within mitochondria that could provide clues to the metabolic connections between caloric restriction and aging.
The map, as well as the techniques used to create it, could lead to a better understanding of how cell metabolism is rewired in some cancers, age-related diseases and metabolic conditions such as diabetes.
"It’s really a dynamic atlas for regulatory points in mitochondrial function — there are many interesting avenues that other scientists can follow up on," says John Denu, University of Wisconsin-Madison professor of biomolecular chemistry and leader of the epigenetics theme at the Wisconsin Institute for Discovery (WID). "It could take years for researchers to understand what it all means, but at least now we have a list of the most important players."

(Image Credit: © Alexander Raths - Fotolia.com)

Scientists create road map to metabolic reprogramming for aging

In efforts to understand what influences life span, cancer and aging, scientists are building road maps to navigate and learn about cells at the molecular level.

To survey previously uncharted territory, a team of researchers at UW-Madison has created an “atlas” that maps more than 1,500 unique landmarks within mitochondria that could provide clues to the metabolic connections between caloric restriction and aging.

The map, as well as the techniques used to create it, could lead to a better understanding of how cell metabolism is rewired in some cancers, age-related diseases and metabolic conditions such as diabetes.

"It’s really a dynamic atlas for regulatory points in mitochondrial function — there are many interesting avenues that other scientists can follow up on," says John Denu, University of Wisconsin-Madison professor of biomolecular chemistry and leader of the epigenetics theme at the Wisconsin Institute for Discovery (WID). "It could take years for researchers to understand what it all means, but at least now we have a list of the most important players."

(Image Credit: © Alexander Raths - Fotolia.com)

Filed under caloric restriction mitochondrial proteins mitochondria metabolism aging neuroscience science

206 notes

Is “Deep Learning” a Revolution in Artificial Intelligence?
Can a new technique known as deep learning revolutionize artificial intelligence as the New York Times suggests?
The technology on which the Times focusses, deep learning, has its roots in a tradition of “neural networks” that goes back to the late nineteen-fifties. At that time, Frank Rosenblatt attempted to build a kind of mechanical brain called the Perceptron, which was billed as “a machine which senses, recognizes, remembers, and responds like the human mind.” The system was capable of categorizing (within certain limits) some basic shapes like triangles and squares. Crowds were amazed by its potential, and even The New Yorker was taken in, suggesting that this “remarkable machine…[was] capable of what amounts to thought.”
But the buzz eventually fizzled; a critical book written in 1969 by Marvin Minsky and his collaborator Seymour Papert showed that Rosenblatt’s original system was painfully limited, literally blind to some simple logical functions like “exclusive-or” (As in, you can have the cake or the pie, but not both). What had become known as the field of “neural networks” all but disappeared.
Read more

Is “Deep Learning” a Revolution in Artificial Intelligence?

Can a new technique known as deep learning revolutionize artificial intelligence as the New York Times suggests?

The technology on which the Times focusses, deep learning, has its roots in a tradition of “neural networks” that goes back to the late nineteen-fifties. At that time, Frank Rosenblatt attempted to build a kind of mechanical brain called the Perceptron, which was billed as “a machine which senses, recognizes, remembers, and responds like the human mind.” The system was capable of categorizing (within certain limits) some basic shapes like triangles and squares. Crowds were amazed by its potential, and even The New Yorker was taken in, suggesting that this “remarkable machine…[was] capable of what amounts to thought.”

But the buzz eventually fizzled; a critical book written in 1969 by Marvin Minsky and his collaborator Seymour Papert showed that Rosenblatt’s original system was painfully limited, literally blind to some simple logical functions like “exclusive-or” (As in, you can have the cake or the pie, but not both). What had become known as the field of “neural networks” all but disappeared.

Read more

Filed under brain neural networks AI deep learning neuroscience science

54 notes


Research shows brain hub activity different in coma patients
A team of French and British researchers has found that brain region activity for coma patients is markedly different than for healthy people. In their paper published in the Proceedings of the National Academy of Sciences, the group describes the differences found when comparing fMRI scans of people in a coma with healthy volunteers.
To gain a better understanding of what goes on in the brain when a person is in a coma, and perhaps the nature of consciousness, the researchers performed fMRI brain scans on 17 people who had recently become comatose due to medical conditions that led to blockage of oxygen to the brain. They then compared those scans to those taken of 20 healthy volunteers.
In analyzing the results the team found that global comparisons between the two groups revealed very few if any differences. Blood continued to flow to all of the parts of the brain. When focusing on the brain as a network however, they found very large differences.
To look at the brain as a network requires looking at its different parts as regions that communicate with one another, forming hubs. In healthy people, certain regions or hubs are busier than others as evidenced by more blood flow. But for the people in a coma, the team found, the normally busy hubs grew less busy, while other hubs grew busier, indicating a major change in the flow of information.
The researchers suggest that the brain scans reveal that the normally busy hubs in healthy people are centers of consciousness and their reduced role in communications in comatose patients suggests that they are most likely not conscious of their existence. They point to prior research that has suggested that being in a coma is more likely closer to the experience of being under anesthesia than being asleep. They add that the their research indicates that regions of the brain that are responsible for conscience thought likely require more oxygen rich blood, and are thus likely to be more sensitive to oxygen deprivation than other areas of the brain, which might explain why people go into a coma when those regions are harmed.

Research shows brain hub activity different in coma patients

A team of French and British researchers has found that brain region activity for coma patients is markedly different than for healthy people. In their paper published in the Proceedings of the National Academy of Sciences, the group describes the differences found when comparing fMRI scans of people in a coma with healthy volunteers.

To gain a better understanding of what goes on in the brain when a person is in a coma, and perhaps the nature of consciousness, the researchers performed fMRI brain scans on 17 people who had recently become comatose due to medical conditions that led to blockage of oxygen to the brain. They then compared those scans to those taken of 20 healthy volunteers.

In analyzing the results the team found that global comparisons between the two groups revealed very few if any differences. Blood continued to flow to all of the parts of the brain. When focusing on the brain as a network however, they found very large differences.

To look at the brain as a network requires looking at its different parts as regions that communicate with one another, forming hubs. In healthy people, certain regions or hubs are busier than others as evidenced by more blood flow. But for the people in a coma, the team found, the normally busy hubs grew less busy, while other hubs grew busier, indicating a major change in the flow of information.

The researchers suggest that the brain scans reveal that the normally busy hubs in healthy people are centers of consciousness and their reduced role in communications in comatose patients suggests that they are most likely not conscious of their existence. They point to prior research that has suggested that being in a coma is more likely closer to the experience of being under anesthesia than being asleep. They add that the their research indicates that regions of the brain that are responsible for conscience thought likely require more oxygen rich blood, and are thus likely to be more sensitive to oxygen deprivation than other areas of the brain, which might explain why people go into a coma when those regions are harmed.

Filed under brain brain activity coma blood flow neuroimaging neuroscience science

159 notes

Research may explain why some people with schizophrenia do not respond to treatment

New research suggests that the molecular mechanism leading to schizophrenia may be different in patients who fail to respond to anti-psychotic medication compared to patients who do respond.

The research, from King’s College London’s Institute of Psychiatry may help explain why up to one third of patients with schizophrenia do not respond to traditional anti-psychotic medication.

Schizophrenia is known to be associated with an overactive dopamine system, meaning that the brain processes abnormally high levels of dopamine. Traditional dopamine-blocking anti-psychotic medication attempts to normalise this process. However, approximately one third of patients with schizophrenia do not respond to this treatment, and until now, no study has examined whether dopamine abnormality is present in patients resistant to antipsychotic treatment.

The study was led by Dr Arsime Demjaha, Dr Oliver Howes, Professor Shitij Kapur, Professor Sir Robin Murray and Professor Philip McGuire from King’s Institute of Psychiatry and published in the American Journal of Psychiatry

Dr Arsime Demjaha and co-authors, say: ‘Despite considerable scientific and therapeutic progress over the last 50 years, we still do not know why some patients with schizophrenia respond to treatment whilst others do not. Treatment resistance in such a disabling condition is one of the greatest clinical and therapeutic challenges to psychiatry, significantly affecting patients, their families and society in general.’

The authors conclude: ‘Our findings suggest that there may be a different molecular mechanism leading to schizophrenia in patients who do not respond to anti-psychotic medication. Identifying the precise molecular pathway particularly in these patients is of utmost importance and will help inform the development of much-needed novel treatments.’

Researchers used PET scan imaging to investigate dopamine synthesis capacity in 12 patients with schizophrenia who did not respond to treatment, 12 who did, and 12 healthy controls. They found that schizophrenia patients whose illness was resistant to antipsychotic treatment have relatively normal levels of dopamine synthesis capacity which would explain why the dopamine blocking anti-psychotic medication was not effective in this group. 

However, the authors add that the findings need to be replicated in larger samples before the research can affect clinical practice. They add that future research will need to focus on long-term prospective studies of patients who have never taken anti-psychotics to determine whether presynaptic dopamine synthesis capacity was normal in patients in the treatment-resistant group at the onset of their illness, and predates antipsychotic exposure.

(Source: kcl.ac.uk)

Filed under schizophrenia dopamine system treatment resistance treatment neuroscience science

27 notes

Biomarker progress offers hope for early autism spectrum disorder detection

Autism spectrum disorders (ASD) are neurodevelopmental disorders typically characterized by difficulties in social interactions and delayed or abnormal language development. Although ASD reportedly affects 1 in 88 people in the United States, to date there have been no distinctive biomarkers to diagnose the disease. In a special themed issue of Disease Markers, investigators report on the current understanding of ASD genetics and the possibilities of translating genetic research toward biomarker development in ASD.

"Although some individuals with ASD are highly functional, many are severely impaired and require permanent care. The significant level of impairment combined with the fact that no specific therapy is yet available for ASD, make ASD a devastating illness for patients and families, and a heavy financial burden for the healthcare system," says guest editor, Irina Voineagu, MD, PhD, RIKEN Omics Science Center, Yokohama, Japan. "The most effective intervention for ASD has proven to be early behavioral therapy. Thus the identification of biological markers for ASD, allowing very early detection, even before the onset of symptoms, would be of tremendous value."

Five articles comprise this comprehensive issue, providing an overview of ASD genetic models, an exploration of several key emerging concepts in understanding ASD’s molecular basis, and discussion of current biomarker development, focusing on genomic data.

Following an introduction by Voineagu, Yuri Bozzi and colleagues review the phenotype characteristics of currently available mouse models of ASD. Carmen Panaitof then discusses the role of the songbird as an experimental model system for investigating the genetic basis of human language and its ASD-related impairments. Michael Bowers and Genevieve Konopka further explore language deficits and provide new evidence for the role of the FOXP gene to regulate language. Alka Saxena, Dave Tang, and Piero Carninci focus on the functional roles of the gene MECP2, which is mutated in most cases of Rett syndrome, one of the ASDs.

A review rounding out the issue is “Subphenotype-Dependent Disease Markers for Diagnosis and Personalized Treatment of Autism Spectrum Disorders,” by Valerie W. Hu, PhD, The George Washington University, School of Medicine and Health Sciences, Washington, DC, PhD, which discusses current progress toward identifying ASD biomarkers based on genome-wide data.

"Without genetic or molecular markers for screening, individuals with ASD are typically not diagnosed before the age of 2, with milder cases diagnosed much later," writes Dr. Hu. "Because early diagnosis is tantamount to early behavioral intervention, which has been shown to improve individual outcomes, an objective biomarker test that can diagnose at-risk children perinatally is a medical imperative."

Hu demonstrates the possibility and importance of developing ASD subtypes to help identify relevant disease markers, which can ultimately aid in developing specific targeted therapies.

Voineagu concludes, “It is exciting times for genetic research and although the phenotypic and genetic heterogeneity of ASD often seem to be a daunting conundrum, well-defined diagnostic criteria, larger cohort sizes for genetic studies and integrative approaches of genomic and epigenomic data already delineate a promising avenue for elucidating the mechanisms of ASD.”

(Source: eurekalert.org)

Filed under autism ASD biomarkers genetic models neuroscience science

free counters