Neuroscience

Month

July 2013

Foraging for thought – new insights into our working memory

We take it for granted that our thoughts are in constant turnover. Metaphors like “stream of consciousness” and “train of thought” imply steady, continuous motion. But is there a mechanism inside our heads that drives this? Is there something compelling our attention to move on to new ideas instead of dwelling in the same spot forever?

image

A research team led by Dr Matthew Johnson in the School of Psychology at The University of Nottingham Malaysia Campus (UNMC) may have discovered part of the answer. They have pinpointed an effect that makes people turn their attention to something new rather than dwelling on their most recent thoughts. The research, which has been published in the academic journal Psychological Science, could have implications for studying disorders like autism and ADHD.

Dr Johnson said: “We have discovered a very promising paradigm. The effect is strong and replicates easily – you could demonstrate it in any psychology lab in the world. The work is still in its early stages but I think this could turn out to be a very important part of our understanding of how and why our thoughts work the way they do.

The paper “Foraging for Thought: An Inhibition-of-Return-Like Effect Resulting From Directing Attention Within Working Memory” sheds new light on what makes us turn our attention to things we haven’t recently thought rather than ones we have. It was carried out in collaboration with Yale University, Princeton University, The Ohio State University, and Manhattanville College.

The “inhibition of return” effect is well-established in visual attention. At certain time scales, people are slower to turn their thoughts back to a location they have just paid attention to. They are much quicker to focus on a new location. Some have interpreted this effect as a “foraging facilitator,” a process that encourages organisms to visit new locations over previously visited ones when exploring a new environment or performing a visual search.

However, in this new study, the researchers weren’t focusing on visual search, but on the process of thought itself. Participants were shown either two words or two pictures, and when the items disappeared, they were instructed to turn their attention briefly to one of the items they were just shown and ignore the other. Immediately afterwards they were asked to identify either the item they had just thought about, or the one they had ignored. For both pictures and words the participants were quicker to react to the item they had ignored.

Dr Johnson said: “The effect was shocking. When we began we expected to find the exact opposite – that thinking about something will make it easier to identify. We were initially disappointed – but when the effect was replicated over multiple experiments we realised we were onto something new and exciting.”

Critically, the effect is temporary; on a later memory test participants remembered attended items better than ignored ones.

Dr Johnson said: “That’s important. If thinking about things made us worse at remembering them long-term, it would make no sense for real-world survival. That’s why we think we’ve tapped into something fundamental about how we think in the moment – a possible mechanism keeping our thoughts moving onto new things, and not getting stuck.”

The researchers have more experiments planned to explore this effect. They say the new task could have implications for studying disorders like autism and ADHD, where attention may persist too long or move on too easily, as well as conditions with more general cognitive impairments, such as schizophrenia and ageing-related dementia.

Future studies planned also include applying cognitive neuroscience techniques to determine the effect’s underlying neural foundations.

Jul 15, 201388 notes
#working memory #autism #ADHD #attention #psychology #neuroscience #science
Jul 14, 2013664 notes
#science #flatworms #regeneration #memory RNA #memory #epigenetics #neuroscience
Jul 14, 2013256 notes
#science #nostalgia #southampton nostalgia scale #music #memories #psychology #neuroscience
Jul 14, 2013117 notes
#Neandertals #evolution #language #modern language #linguistics #mitochondrial DNA #science
People in their 90s are getting smarter

Ninety-somethings seem to be getting smarter. Today’s oldest people are surviving longer, and thankfully appear to have sharper minds than the people reaching their 90s 10 years ago.

image

Kaare Christensen, head of the Danish Aging Research Center at the University of Southern Denmark in Odense, and colleagues found Danish people born in 1915 were about a third more likely to live to their 90s than those born in 1905, and were smarter too.

During research, which spanned 12 years and involved more than 5000 people, the team gave nonagenarians born in 1905 and 1915 a standard test called a “mini-mental state examination”, and cognitive tests designed to pick up age-related changes. Not only did those born in 1915 do better at both sets of tests, more of them also scored top marks in the mini-mental state exam.

It’s a landmark study, says Marcel Olde Rikkert, head of the Alzheimer’s centre at Radboud University Nijmegen Medical Centre in the Netherlands. It is scientifically rigorous, it invited all over 90-year-olds in Denmark to participate, and it also overturns our ingrained views of old age, he says.

Getting better all the time

"The outcome underlines that ageing is malleable," Olde Rikkert says, adding that cognitive function can actually be a lot better than people would assume until a very high age.

"It’s motivating that people, their lifestyles, and their environments can contribute a lot to the way they age," he says, though he cautions that not everything is in our own hands and help is still needed for those with dementia or those who do experience cognitive decline as they age.

Improved education played a part in the changes, says Christensen. But the study does not disentangle the individual effects of the numerous things that could be responsible for the improvements. “The 1915 cohort had a number of factors on their side – they experienced better living and working conditions, they had radio, TV and newspapers earlier in their lives than those born 10 years before,” he says.

Tellingly, there was no difference in the physical test results between the two groups. The authors say this “suggests changes in the intellectual environment rather than in the physical environment are the basis for the improvement”.

Jul 14, 2013152 notes
#aging #cognitive functioning #performance #cognitive tests #psychology #neuroscience #science
New insight into the human genome through the lens of evolution

By comparing the human genome to the genomes of 34 other mammals, Australian scientists have described an unexpectedly high proportion of functional elements conserved through evolution.

Less than 1.5% of the human genome is devoted to conventional genes, that is, encodes for proteins.  The rest has been considered to be largely junk.  However, while other studies have shown that around 5-8% of the genome is conserved at the level of DNA sequence, indicating that it is functional, the new study shows that in addition much more, possibly up to 30%, is also conserved at the level of RNA structure.

DNA is a biological blueprint that must be copied into another form before it can be actualised. Through a process known as ‘transcription’, DNA is copied into RNA, some of which ‘encodes’ the proteins that carry out the biological tasks within our cells. Most RNA molecules do not code for protein, but instead perform regulatory functions, such as determining the ways in which genes are expressed.

Like infinitesimally small Lego blocks, the nucleic acids that make up RNA connect to each other in very specific ways, which force RNA molecules to twist and loop into a variety of complicated 3D structures.

Dr Martin Smith and Professor John Mattick, from Sydney’s Garvan Institute of Medical Research, devised a method for predicting these complex RNA structures – more accurate than those used in the past – and applied it to the genomes of 35 different mammals, including bats, mice, pigs, cows, dolphins and humans. At the same time, they matched mutations found in the genomes with consistent RNA structures, inferring conserved function. Their findings are published in Nucleic Acids Research, now online.

“Genomes accumulate mutations over time, some of which don’t change the structure of associated RNAs. If the sequence changes during evolution, yet the RNA structure stays the same, then the principles of natural selection suggest that the structure is functional and is required for the organism,” explained Dr Martin Smith.

“Our hypothesis is that structures conserved in RNA are like a common template for regulating gene expression in mammals – and that this could even be extrapolated to vertebrates and less complex organisms.”

“We believe that RNA structures probably operate in a similar way to proteins, which are composed of structural domains that assemble together to give the protein a function.”

“We suspect that many RNA structures recruit specific molecules, such as proteins or other RNAs, helping these recruited elements to bond with each other. That’s the general hypothesis at the moment – that non-coding RNAs serve as scaffolds, tethering various complexes together, especially those that control genome organization and expression during development.”

“We know that many RNA transcripts are associated with diseases and developmental conditions, and that they are differentially expressed in distinct cells.”

“Our structural predictions can serve as an annotative tool to help researchers understand the function of these RNA transcripts.”

“That is the first step – the next is to describe the structures in more detail, figure out exactly what they do in the cell, then work out how they relate to our normal development and to disease.”

Jul 14, 2013106 notes
#mammals #human genome #evolution #mutations #gene expression #science
The brain is alive, will new MRI diffusion techniques let us see it move and shake?

Pioneering experiments back in 1982 by Tasaki and Iwasa at the NIH revealed that action potentials in neurons are more than just the electrical blips that physiologists readily amplify and record. These so-called “spikes” are in fact multi-modal signalling packages that include mechanical and thermal disturbances propagating down the axon at their own rates. Nobel Laureate Francis Crick published a paper that same year, in which he postulated potential mechanisms that would explain twitching in dendritic spines, adding to an emerging picture of a brain more vibrant and motile than had been previously imagined. More recently, researchers have developed diffusion-based MRI methods, like diffusion tensor imaging (DTI), to trace the trajectories of axons, and perhaps more intriguingly, determine their directional polarity. Working at the EPFL in Switzerland, Denis Le Bihan and his co-workers have been using diffusional MRI in slightly different way. They now appear to be able to directly measure neuronal activity from the subtle movements of membranes, the water within them, and in the extracellular space around them. Their work, just published in PNAS, provides a much needed conceptual shift away from currently established, but typically nebulous, ideas regarding neurovascular coupling of brain activity to blood flow.

image

Present-day imaging methods, like blood oxygen level-dependent (BOLD) MRI, are only indirectly and remotely related to the cortical activity they often claim to measure. In 2006, Le Bihan reported a water “phase transition” response that preceded the neurovascular response normally detected by functional MRI. He attributed the changes in water diffusion to previously established effects involving membrane expansion and cell swelling secondary to activity. At the biophysical level, interpreting action potentials as phase transitions is a little off the beaten path from traditional neurobiology, but it can be an informative approach when to trying to understand what might be going on when cells fire.

As biophysicist Gerald Pollack has previously pointed out, spikes may involve the propagation of the line of transition of water from the ordered phase, (as patterned by hydrophic interactions nucleated at the surfaces of membranes and proteins) to a disordered phase.
Traditionally, the so-called bound surface water only extends out a only a couple of molecules from the surface of nondiffusable features. That idea may need to be revisited in light of more recent understanding when attempting to account for the diffusion of water in axons. A decrease in water diffusion as measured by MRI may be in part explained by a decrease in extracellular space, and that has been suggested from experiments measuring intrinsic optical effects. The larger picture of water diffusion, however, is likely a bit more complicated than this.

In his new study, Le Bihan stimulated the forepaw of a rat and looked at responses in the somatosensory cortex. The key experiment was to infuse nitroprusside in attempt to inhibit neurovascular coupling. It is a tricky alteration because nitroprusside apparently has many diffuse effects. It can induce potent vasodilation, particularly on the vascular end (mainly the smaller venules), after it breaks down to produce nitric oxide. It is also a diamagnetic molecule, and each molecule releases five cyanide ions, which are presumably detoxified by the mitochondrial enzyme rhodanese. The experiments were done under isoflurane anesthesia, which also introduces a few uncertainties, particularly with regard to responses to different frequencies of forepaw stimulation.

If nitroprusside is indeed a realistic experimental proxy for neurovascular uncoupling, then the results of Le Bihan appear to show that the diffusion response is not of vascular origin, and that it is closely linked to neural activation. He found that the standard BOLD MRI responses were completely quenched under nitroprusside, whereas the diffusion MRI responses were only slightly suppressed. Local field potentials were also simultaneously measured and suggested at least, that the neuronal responses were also intact.

The work of Le Bihan indicates that diffusion-based MRI can be used to infer neural activity directly from the structural changes that affect the molecular displacements of water. The ability to use shape changes in neurons, astrocytes, or even spines, raises the question of whether these kinds of techniques might eventually be of use in creating larger scale, and more detailed, Brain Activity Maps (BAMs). I asked Konrad Kording, author on the recent theoretical paper which discussed the theoretical limits to MRI and other activity recording methods, whether methods that probe water movements might be applied to this end.

Kording observed that the spatial resolution of standard MRI is ultimately limited by the diffusion of water, but more importantly perhaps, the temporal resolution of all known MRI methods is nowhere near that required to create spike maps. None-the-less, detecting mechanical responses in the brain could provide many unique insights into function. For example, experiments using agents that dissolve the extracellular matrix, like the clot-busting drug TPA, result in more twitching, or vibration if you will, in dendritic spines. Other studies have shown that the greater the electrical drive on a spine, the less it tends to twitch or change size, particularly during periods of rapid development.

Similarly, sensory deprivations appear to increase these kinds of movements as neurons grow and reorganize connections. While these effects are far below that which could be detected by any large external method of MRI, new tools may permit us to access these newly-revealed activities. Diffusional MRI in particular, can be done with a little modification of the standard MRI procedure. For example, to determine directional diffusion parameters, or diffusion tensors, typically six gradients are used to measure three directional vectors. As these capabilities become more common, hopefully the results of Le Bihan can be further explored and verified.

Jul 13, 201399 notes
#brain activity #blood flow #neuroimaging #diffusion tensor imaging #cortical activity #neuroscience #science
Jul 13, 2013138 notes
#science #alzheimer's disease #beta amyloid #dementia #protein misfolding #fibrils #neuroscience
Jul 13, 201366 notes
#zebrafish #brain activity #telencephalon #memory #LTM #neuroimaging #neurons #neuroscience #science
Jul 12, 2013106 notes
#vitamin B #choline #memory #attention #dietary supplements #animal model #neuroscience #science
The brain processes complex stimuli more cumulatively than we thought

A new study reveals that the representation of complex features in the brain may begin earlier—and play out in a more cumulative manner—than previously thought.

The finding represents a new view of how the brain creates internal representations of the visual world. “We are excited to see if this novel view will dominate the wider consensus” said senior author Dr. Miyashita, who is also Professor of Physiology at the University of Tokyo’s School of Medicine, “and also about the potential impact of our new computational principle on a wide range of views on human cognitive abilities.”

The brain recalls the patterns and objects we observe by developing distinct neuronal representations that go along with them (this is the same way it recalls memories). Scientists have long hypothesized that these neuronal representations emerge in a hierarchical process limited to the same cortical region in which the representations are first processed.

Because the brain perceives and recognizes the external world through these internal images, any new information about the process by which this takes place has the power to inform our understanding of related functions, including knowledge acquisition and memory.

However, studies attempting to uncover the functional hierarchy involved in the cortical process of visual stimuli have tried to characterize this hierarchy by analyzing the activity of single nerve cells, which are not necessarily correlated with neurons nearby, thus leaving these analyses lacking.

In a new study appearing in the 12 July issue of the journal Science, lead author Toshiyuki Hirabayashi and colleagues focus not on single neurons but instead on the relationship between neuron pairs, testing the possibility that the representation of an object in a single brain region emerges in a hierarchically lower brain area.

"I became interested in this work," said Dr. Hirabayashi, "because I was impressed by the elaborate neuronal circuitry in the early visual system, which is well-studied, and I wanted to explore the circuitry underlying higher-order visual processing, which is not yet fully understood."

Hirabayashi and colleagues analyzed nerve cell pairs in cortical areas TE and 36, the latter of which is hierarchically higher, in two adult macaques. After these animals looked at six sets of paired stimuli for several months to learn to associate related objects (a process that can lead to pair-coding neurons in the brain), the researchers recorded neuron responses in areas TE and 36 of both animals as they again performed this task.

The neurons exhibited pair association, but not where the researchers would have thought. “The most surprising result,” said senior author Dr. Yasushi Miyashita “was that the neuronal circuit that generated pair-association was found only in area TE, not in area 36.” Indeed, based on previous studies, which indicated that the number of pair-coding neurons in area TE is much smaller, the researchers would have expected the opposite.

During their study, Miyashita and other team members observed that in region TE of the macaque cortex, unit 1 neurons (or source neurons) provided input to unit 2 neurons (or target neurons), which—unlike unit 1 neurons—responded to both members of a stimulus pair.

"The representations generated in area TE did not reflect a mere random fluctuation of response patterns," explained Dr. Miyashita, "but rather, they emerged as a result of circuit processing inherent to that area of the brain."

In area 36, meanwhile, members of neuron pairs behaved differently; on average, unit 1 as well as unit 2 neurons responded to both members of a stimulus pair. Neurons in area 36 received input from area TE, but only from its unit 2 neurons.

Taken together, these findings lead the authors to hypothesize the existence of a hierarchical relationship between regions TE and 36, in which paired associations first established in the former region are propagated to the latter one. Here, area 36 represents the next level of a so-called feed forward hierarchy.

The work by Hirabayashi and colleagues suggests that the detailed representations of objects commonly observed in the brain are attained not by buildup of representations in a single area, but by emergence of these representations in a hierarchically prior area and their subsequent transfer to the brain region that follows. There, they become sufficiently prevalent for the brain to register. The work also reveals that the brain activity involved in recreating visual stimuli emerges in a hierarchically lower brain area than previously thought.

Moving forward, the Japanese research team has plans to expand upon this research, thus continuing to contribute to studies worldwide that aim to give scientists the best possible tools with which to obtain a dynamic picture of the brain. As a next step, the team hopes to further elucidate interactions between the various cortical microcircuits that operate in memory encoding. Dr. Miyashita has conjectured that these microcircuits are manipulated by a global brain network. Using the results of this latest study, he and colleagues are poised to further evaluate this assumption.

"It will also be important to weave the neuronal circuit mechanisms into a unified framework," said Dr. Hirabayashi," and to examine the effects of learning on these circuit organizations."

Equipped with their new view of cortical processing, the team also hopes to trace the causal chain of memory retrieval across different areas of the cortex. “I am excited by the recent development of genetic tools that will allow us to do this,” said Dr. Miyashita. A better understanding of object representations from one area of the brain to the next will shed even greater light on elusive aspects of this hierarchical organ.

Jul 12, 201377 notes
#object representations #neural circuitry #temporal cortex #neurons #primates #neuroscience #science
Jul 12, 2013167 notes
#science #daydreaming #brain activity #brain networks #AI #memory #cognitive impairment #neuroscience
Jul 12, 2013251 notes
#science #extroverts #dopamine #reward system #associative conditioning #neuroscience
Study Finds Factors That May Cause Fluctuations in Deep Brain Stimulation Levels Over Time

Deep brain stimulation therapy blocks or modulates electrical signals in the brain to improve symptoms in patients suffering from movement disorders such as Parkinson’s disease, essential tremor and dystonia, but a new study suggests that several factors may cause electrical current to vary over time.

Led by Michele Tagliati, MD, director of Cedars-Sinai Medical Center’s Movement Disorders Program, the study identified variables that affect impedance – resistance in circuits that affect intensity and wavelength of electrical current. Doctors who specialize in programming DBS devices fine-tune voltage, frequency and other parameters for each patient; deviations from these settings may have the potential to alter patient outcomes.

“Deep brain stimulation devices are currently designed to deliver constant, steady voltage, and we believe consistency and reliability are critical in providing therapeutic stimulation. But we found that we cannot take impedance stability for granted over the long term,” said Tagliati, the senior author of a journal article that reveals the study’s findings.

“Doctors with experience in DBS management can easily make adjustments to compensate for these fluctuations, and future devices may do so automatically,” he added. “Although our study was not designed to link changes in impedance and voltage with clinical outcomes, we believe it is important for patients to have regular, ongoing clinic visits to be sure they receive a steady level of stimulation to prevent the emergence of side effects or the re-emergence of symptoms.”

Findings of the study – one of the largest of its kind and possibly the first to follow patients for up to five years – were published online ahead of print in Brain Stimulation. Researchers collected 2,851 impedance measurements in 94 patients over a period of six months to five years, evaluating fluctuations in individual patients and in individual electrodes. They looked at a variety of factors, including how long a patient had undergone treatment, the position of the implanted electrode, the side of the brain where the electrode was implanted, and even placement and function of contact positions along electrodes.

Medications usually are the first line of treatment for movement disorders, but if drugs fail to provide adequate relief or side effects are excessive, neurologists and neurosurgeons may supplement them with deep brain stimulation. Electrical leads are implanted in the brain, and an electrical pulse generator is placed near the collarbone. The device is then programmed with a remote, hand-held controller.

Jul 12, 201347 notes
#deep brain stimulation #parkinson's disease #movement disorders #neuroscience #science
Jul 12, 201372 notes
#brain development #cry analyzer #infants #cri du chat #down syndrome #neuroscience #science
Autism Speaks Collaborative Releases First Full Genome Sequencing for Autism Spectrum Disorders

A collaborative formed by Autism Speaks, the world’s leading autism science and advocacy organization, has found full genome sequencing examining the entire DNA code of individuals with autism spectrum disorder (ASD) and their family members to provide the definitive look at the wide ranging genetic variations associated with ASD. The study published online today in American Journal of Human Genetics, reports on full genome sequencing on 32 unrelated Canadian individuals with autism and their families, participants in the Autism Speaks Autism Genetic Resource Exchange (AGRE). The results include both inherited as well as spontaneous or de novo, genetic alterations found in one half of the affected families sequenced.

This dramatic finding of genetic risk variants associated with clinical manifestation of ASD or accompanying symptoms in 50 percent of the participants tested is promising, as current diagnostic technology has only been able to determine a genetic basis in about 20 percent of individuals with ASD tested. The large proportion of families identified with genetic alterations of concern is in part due to the comprehensive and uniform ability to examine regions of the genome possible with whole genome sequencing missed in other lower resolution genome scanning approaches.

"From diagnosis to treatment to prevention, whole genome sequencing efforts like these hold the potential to fundamentally transform the future of medical care for people with autism," stated Autism Speaks Chief Science Officer and study co-author Robert Ring, Ph.D.

The study identified genetic variations associated with risk for ASD including de novo, X-linked and other inherited DNA lesions in four genes not previously recognized for ASD; nine genes previously determined to be associated with ASD risk; and eight candidate ASD risk genes. Some families had a combination of genes involved. In addition, risk alterations were found in genes associated with fragile X or related syndromes (CAPRIN1 and AFF2), social-cognitive deficits (VIP), epilepsy (SCN2A and KCNQ2) as well as NRXN1 and CHD7, which causes ASD-associated CHARGE syndrome.

“Whole genome sequencing offers the ultimate tool to advance the understanding of the genetic architecture of autism,” added lead author Dr. Stephen Scherer, senior scientist and director of the Centre for Applied Genomics at The Hospital for Sick Children (SickKids) and director of the McLaughlin Centre at the University of Toronto. “In the future, results from whole genome sequencing could highlight potential molecular targets for pharmacological intervention, and pave the way for individualized therapy in autism. It will also allow for earlier diagnosis of some forms of autism, particularly among siblings of children with autism where recurrence is approximately 18 per cent.”  

This $1 million collaboration of Autism Speaks, SickKids, BGI and Duke University piloted Autism Speaks’ initiative to generate the world’s largest library of sequenced genomes of individuals with ASD announced in late 2011. “As we continue to test more individuals and their family members from the AGRE cohort, we expect to discover and study additional genetic variants associated with autism. This collaboration will accelerate basic and translational research in autism and related developmental disabilities,” concluded Autism Speaks Vice President for Scientific Affairs Andy Shih, Ph.D. who oversees the collaboration, “and this collection of sequenced genomes will facilitate new collaborations engaging researchers around the world, and enable public and private entities to pursue pivotal research.”

In this pilot effort, a total of 99 individuals were tested, including the 32 individuals with ASD (25 males and seven females) and their two parents, as well as three members of one control family not on the autism spectrum.  Using families in the Autism Speaks AGRE collection, this Autism Speaks initiative will ultimately perform whole genome sequencing on more than 2,000 participating families who have two or more children on the autism spectrum. The data from the 10,000 AGRE participants will enable new research in the genomics of ASD, and significantly enhance the science and technology networks of Autism Speaks and its collaborators.

Jul 12, 201369 notes
#autism #ASD #fragile X #genomics #genetics #neuroscience #science
Jul 11, 2013207 notes
#prefrontal cortex #neural plasticity #stress #aging #neuroscience #psychology #science
Females Respond Better to Stress Because of Estrogen in the Brain, Animal Study Finds

The idea that females are more resilient than males in responding to stress is a popular view, and now University at Buffalo researchers have found a scientific explanation. The paper describing their embargoed study will be published July 9 online, in the high-impact journal, Molecular Psychiatry.

“We have examined the molecular mechanism underlying gender-specific effects of stress,” says senior author Zhen Yan, PhD, a professor in the Department of Physiology and Biophysics in the UB School of Medicine and Biomedical Sciences. “Previous studies have found that females are more resilient to chronic stress and now our research has found the reason why.”

The research shows that in rats exposed to repeated episodes of stress, females respond better than males because of the protective effect of estrogen.

In the UB study, young female rats exposed to one week of periodic physical restraint stress showed no impairment in their ability to remember and recognize objects they had previously been shown. In contrast, young males exposed to the same stress were impaired in their short-term memory.

An impairment in the ability to correctly remember a familiar object signifies some disturbance in the signaling ability of the glutamate receptor in the prefrontal cortex, the brain region that controls working memory, attention, decision-making, emotion and other high-level “executive” processes.

Last year, Yan and UB colleagues published in Neuron a paper showing that repeated stress results in loss of the glutamate receptor in the prefrontal cortex of young males.

The current paper shows that the glutamate receptor in the prefrontal cortex of stressed females is intact. The findings provide more support for a growing body of research demonstrating that the glutamate receptor is the molecular target of stress, which mediates the stress response.

The stressors used in the experiments mimic challenging and stressful, but not dangerous, experiences that humans face, such as those causing frustration and feelings of being under pressure, Yan says.

By manipulating the amount of estrogen produced in the brain, the UB researchers were able to make the males respond to stress more like females and the females respond more like males.

“When estrogen signaling in the brains of females was blocked, stress exhibited detrimental effects on them,” explains Yan. “When estrogen signaling was activated in males, the detrimental effects of stress were blocked.

“We still found the protective effect of estrogen in female rats whose ovaries were removed,” says Yan. “It suggests that it might be estrogen produced in the brain that protects against the detrimental effects of stress.”

In the current study, Yan and her colleagues found that the enzyme aromatase, which produces estradiol, an estrogen hormone, in the brain, is responsible for female stress resilience. They found that aromatase levels are significantly higher in the prefrontal cortex of female rats.

“If we could find compounds similar to estrogen that could be administered without causing hormonal side effects, they could prove to be a very effective treatment for stress-related problems in males,” she says.

She notes that while stress itself is not a psychiatric disorder, it can be a trigger for the development of psychiatric disorders in vulnerable individuals.

Jul 11, 2013225 notes
#stress #women #estrogen #glutamate receptor #neuroscience #science
Jul 11, 2013107 notes
#autism #ASD #biomarkers #brain size #cerebral spinal fluid #infancy #neuroscience #science
The brain's response to sweets may indicate risk for development of alcoholism

Several human and animal studies have shown a relationship between a preference for highly sweet tastes and alcohol use disorders. Furthermore, the brain mechanisms of sweet-taste responses may share common neural pathways with responses to alcohol and other drugs. A new study using functional magnetic resonance imaging (fMRI) has found that recent drinking is related to the orbitofrontal-region brain response to an intensely sweet stimulus, a brain response that may serve as an important phenotype, or observable characteristic, of alcoholism risk.

Results will be published in the December 2013 issue of Alcoholism: Clinical & Experimental Research and are currently available at Early View.

"It has long-been known that animals bred to prefer alcohol also drink considerably greater quantities of sweetened water than do animals without this selective breeding for alcohol preference," explained David A. Kareken, deputy director of the Indiana Alcohol Research Center, a professor in the department of neurology at Indiana University School of Medicine, and corresponding author for the study. "More recently, it has become clear that animals bred to prefer the artificial sweetener, saccharin, also drink more alcohol. Although the data in humans are somewhat more variable, some studies do show that alcoholics, or even non-alcoholics with a family history of alcoholism, have a preference for unusually sweet tastes. Thus, while the precise reasons remain unclear, there does seem to be significant evidence suggesting some link between the rewarding properties of both sweet tastes and alcohol."

Kareken added that this is the first study to examine the extent to which regions of the brain’s reward system, as they respond to an intensely sweet taste, are related to human drinking patterns.

Kareken and his colleagues recruited 16 (12 males, 4 females) right-handed, non-treatment seeking, healthy volunteers with a mean age of 26 years from the community. All participants underwent a taste test using a range of sucrose concentrations, and their blood oxygen dependent (BOLD) activation was measured during an fMRI scan while receiving small squirts of either water or an intensely sweet mixture of sugar in water. All were asked about their drinking patterns.

"Our study was designed to determine which brain areas responded to sweet taste – as compared to plain water – and the extent to which these brain responses were related to subjects’ binge-drinking patterns, the number of alcoholic drinks subjects consumed per day when drinking," explained Kareken.

"In addition to ‘activating’ the brain’s gustatory or taste circuits, the sugared water also activated key elements of what neuroscientists consider to be part of the brain’s reward system, including the ventral striatum, amygdala, and parts of the orbitofrontal cortex – the inferior frontal lobe surface just above the eyes – that respond to ingested rewards," Kareken said. "We refer to these as ‘primary’ rewards, being distinct from secondary rewards, like money, which can be used to obtain primary rewards."

What the researchers found was that the response to this intensely sweet taste in the left orbitofrontal area correlated significantly with subjects’ drinking patterns.

"Specifically, the trend was such that those who drank more alcohol on drinking days had stronger left orbitofrontal responses to the intensely sweet water," said Kareken. "Subjects’ subjectively rated liking of the sweetened water also contributed to this relationship, so that both the brain response itself, as well as liking of the sugared water, collectively correlated with drinking behavior."

While previous human and animal research has noted this association between preferences for both sweet tastes and alcohol intoxication, Kareken believes that this is the first study to examine the human brain mechanism behind this association.

"While much more research needs to be done to truly understand the commonalities between sweet-liking and alcoholism, and while alcoholism itself is likely the product of several mechanisms, our findings may implicate a particular brain region that is more generally involved in coding for the value of ‘primary’ rewards such as pleasures," he said. "In a more practical sense, the findings are compelling evidence that the brain response to an intensely sweet taste may be used in future research to test for differences in the reward circuits of those at risk for alcoholism. This may be particularly useful since alcohol itself is not an easy drug to work with in this kind of human imaging, and since alcohol exposure is not ethically appropriate for use in all at-risk subjects, or in subjects trying to abstain from drinking."

Jul 11, 2013109 notes
#alcoholism #brain response #sweet taste #reward system #orbitofrontal cortex #neuroscience #science
Jul 11, 201396 notes
Jul 11, 201357 notes
#science #stroke #stroke survivors #rehabilitation #robotic arm #robotics #neuroscience
One More Homo Species?

A recent 3D-comparative analysis confirms the status of Homo floresiensis as a fossil human species

image

Ever since the discovery of the remains in 2003, scientists have been debating whether Homo floresiensis represents a distinct Homo species, possibly originating from a dwarfed island Homo erectus population, or a pathological modern human. The small size of its brain has been argued to result from a number of diseases, most importantly from the condition known as microcephaly.

Based on the analysis of 3-D landmark data from skull surfaces, scientists from Stony Brook University New York, the Senckenberg Center for Human Evolution and Palaeoenvironment, Eberhard-Karls Universität Tübingen, and the University of Minnesota provide compelling support for the hypothesis that Homo floresiensis was a distinct Homo species.

The study, titled “Homo floresiensis contextualized: a geometric morphometric comparative analysis of fossil and pathological human samples,” is published in the July 10 edition of PLOS ONE.

The ancestry of the Homo floresiensis remains is much disputed.
The critical questions are: Did it represent an extinct hominin species? Could it be a Homo erectus population, whose small stature was caused by island dwarfism?

Or, did the LB1 skull belong to a modern human with a disorder that resulted in an abnormally small brain and skull? Proposed possible explanations include microcephaly, Laron Syndrome or endemic hypothyroidism (“cretinism”).

The scientists applied the powerful methods of 3-D geometric morphometrics to compare the shape of the LB1 cranium (the skull minus the lower jaw) to many fossil humans, as well as a large sample of modern human crania suffering from microcephaly and other pathological conditions. Geometric morphometrics methods use 3D coordinates of cranial surface anatomical landmarks, computer imaging, and statistics to achieve a detailed analysis of shape.

This was the most comprehensive study to date to simultaneously evaluate the two competing hypotheses about the status of Homo floresiensis.

The study found that the LB1 cranium shows greater affinities to the fossil human sample than it does to pathological modern humans. Although some superficial similarities were found between fossil, LB1, and pathological modern human crania, additional features linked LB1exclusively with fossil Homo. The team could therefore refute the hypothesis of pathology.

“Our findings provide the most comprehensive evidence to date linking the Homo floresiensis skull with extinct fossil human species rather than with pathological modern humans. Our study therefore refutes the hypothesis that this specimen represents a modern human with a pathological condition, such as microcephaly,” stated the scientists.

Jul 11, 2013104 notes
#homo floresiensis #hominin species #geometric morphometrics #microcephaly #evolution #science
Jul 11, 2013154 notes
#brain mapping #neuroimaging #brain activity #cognitive subtraction #neuroscience #science
Jul 11, 2013103 notes
#stem cells #inner ear #hair cells #embryonic development #hearing loss #neuroscience #science
Double-barreled attack on obesity in no way a no-brainer

In the constant cross talk between our brain and our gut, two gut hormones are already known to tell the brain when we have had enough to eat. New research suggests that boosting levels of these hormones simultaneously may be an effective new weapon in the fight against obesity.

Dr Shu Lin, Dr Yan-Chuan Shi and Professor Herbert Herzog, from Sydney’s Garvan Institute of Medical Research, have shown that when mice are injected with PYY3-36 and PP, they eat less, gain less fat, and tend not to develop insulin-resistance, a precursor to diabetes. At the same time, the researchers have shown that the hormones stimulate different nerve pathways, ultimately, however, affecting complementary brain regions. Their findings are now published online in the journal Obesity.

While the double-barreled approach may seem like a no-brainer, the strongly enhanced effect seen was by no means inevitable. In the complex world of neuroscience, two plus two does not always make four.

Drug companies are in the process of conducting pre-clinical trials to examine the separate effects of boosting the hormones PYY3-36 and PP. Until now, there is no research to indicate the detailed molecular interactions that might occur when they are boosted in tandem.

When used together, the hormones independently, yet with combined force, reduce the amount of neuropeptide Y (NPY) produced by the brain, a powerful neurotransmitter that affects a variety of things including appetite, mood, heart rate, temperature and energy levels.

Each hormone also communicates with a different part of the arcuate nucleus in the hypothalamus, a region of the brain where signals can cross the normally impermeable blood / brain barrier. The stimulated regions then produce other neuronal signals deep within the hypothalamus, bringing about a powerful combined effect.

“There are many factors that influence appetite control – and we now realise that there won’t be a single molecular target, or a single drug, that will be effective,” said Dr Yan-Chuan Shi.

“It will be important for drug companies to try different combinations of targets, to see which combinations are most potent, and at the same time have no side effects, or at least minimal side effects.”

“At the moment, the only effective tool against obesity is surgery. Drug companies have so far failed to produce an effective drug without unacceptable side effects, such as mood disorders, nausea or cardiovascular problems.”

Jul 10, 201345 notes
#obesity #hormones #neuropeptide Y #hypothalamic nuclei #hypothalamus #neuroscience #science
Newly Identified Bone Marrow Stem Cells Reveal Markers for ALS

Amyotrophic Lateral Sclerosis (ALS) is a devastating motor neuron disease that rapidly atrophies the muscles, leading to complete paralysis. Despite its high profile — established when it afflicted the New York Yankees’ Lou Gehrig — ALS remains a disease that scientists are unable to predict, prevent, or cure.

image

Although several genetic ALS mutations have been identified, they only apply to a small number of cases. The ongoing challenge is to identify the mechanisms behind the non-genetic form of the disease and draw useful comparisons with the genetic forms.

Now, using samples of stem cells derived from the bone marrow of non-genetic ALS patients, Prof. Miguel Weil of Tel Aviv University’s Laboratory for Neurodegenerative Diseases and Personalized Medicine in the Department of Cell Research and Immunology and his team of researchers have uncovered four different biomarkers that characterize the non-genetic form of the disease. Each sample shows similar biological abnormalities to four specific genes, and further research could reveal additional commonalities. “Because these genes and their functions are already known, they give us a specific direction for research into non-genetic ALS diagnostics and therapeutics,” Prof. Weil says. His initial findings were reported in the journal Disease Markers.

Giving in to stress

To hunt for these biomarkers, Prof. Weil and his colleagues turned to samples of bone marrow collected from ALS patients. Though more difficult to collect than blood, bone marrow’s stem cells are easy to isolate and grow in a consistent manner. In the lab, he used these cells as cellular models for the disease. He ultimately discovered that cells from different ALS patients shared the same abnormal characteristics of four different genes that may act as biomarkers of the disease. And because the characteristics appear in tissues that are related to ALS — including in muscle, brain, and spinal cord tissues in mouse models of genetic ALS — they may well be connected to the degenerative process of the disease in humans, he believes.

Searching for the biological significance of these abnormalities, Prof. Weil put the cells under stress, applying toxins to induce the cells’ defense mechanisms. Healthy cells will try to fight off threats and often prove quite resilient, but ALS cells were found to be overwhelmingly sensitive to stress, with the vast majority choosing to die rather than fight. Because this is such an ingrained response, it can be used as a feature for drug screening for the disease, he adds.

The hunt for therapeutics

Whether these biomarkers are a cause or consequence of ALS is still unknown. However, this finding remains an important step towards uncovering the mechanisms of the disease. Because these genes have already been identified, it gives scientists a clear direction for future research. In addition, these biomarkers could lead to earlier and more accurate diagnostics.

Next, Prof. Weil plans to use his lab’s high-throughput screening facility — which can test thousands of compounds’ effects on diseased cells every day — to search for drug candidates with the potential to affect the abnormal expression of these genes or the stress response of ALS cells. A compound that has an impact on these indicators of ALS could be meaningful for treating the disease, he says.

Prof. Weil is the director of the new Cell Screening Facility for Personalized Medicine at TAU. The facility is dedicated to finding potential drugs for rare and Jewish hereditary diseases.

Jul 10, 201358 notes
#ALS #motor neuron disease #neurodegenerative diseases #genetics #medicine #biomarkers #science
Jul 10, 2013162 notes
#child development #cognitive development #irregular bedtimes #performance #childhood #neuroscience #science
Study identifies brain circuits involved in learning and decision making

Finding has implications for alcoholism and other patterns of addictive behavior

Research from the National Institutes of Health has identified neural circuits in mice that are involved in the ability to learn and alter behaviors. The findings help to explain the brain processes that govern choice and the ability to adapt behavior based on the end results.

Researchers think this might provide insight into patterns of compulsive behavior such as alcoholism and other addictions.

“Much remains to be understood about exactly how the brain strikes the balance between learning a behavioral response that is consistently rewarded, versus retaining the flexibility to switch to a new, better response,” said Kenneth R. Warren, Ph.D., acting director of the National Institute on Alcohol Abuse and Alcoholism. “These findings give new insight into the process and how it can go awry.”

The study, published online in Nature Neuroscience, indicates that specific circuits in the forebrain play a critical role in choice and adaptive learning.

Like other addictions, alcoholism is a disease in which voluntary control of behavior progressively diminishes and unwanted actions eventually become compulsive. It is thought that the normal brain processes involved in completing everyday activities become redirected toward finding and abusing alcohol.

The research, conducted by investigators from NIAAA, with support from the National Institute of Mental Health and the University of Cambridge, England, used a variety of approaches to study choice.

Researchers used a simple choice task in which mice viewed images on a computer touchscreen and learned to touch a specific image with their nose to get a food reward. Using various techniques to visualize and record neural activity, researchers found that as the mice learned to consistently make a choice, the brain’s dorsal striatum was activated. The dorsal striatum is thought to play an important role in motivation, decision-making, and reward.

Conversely, when the mice later had to shift to a new choice to receive a reward, the dorsal striatum quieted while regions in the prefrontal cortex, an area involved in decision-making and complex cognitive processes, became active.

Building upon these findings, the authors next deleted or pharmacologically blocked a component of nerve cells which normally binds the neurochemical glutamate (specifically, the GluN2B subunit of the NMDA receptor) within two different areas of the brain, the striatum and the frontal cortex. Previous studies have shown that GluN2B plays a role in memory, spatial reference, and attention. Researchers found that making dorsal striatal GluN2B inactive markedly slowed learning, while shutting down GluN2B in the prefrontal cortex made the mice less able to relearn the touchscreen reward task after the reward image was changed.

“These data add to what we understand about the neural control of behavioral flexibility and striatal learning by identifying GluN2B as a critical molecular substrate to both processes,” said the study’s senior author, Andrew Holmes, Ph.D., Laboratory Chief and Principal Investigator of the NIAAA Laboratory of Behavioral and Genomic Neuroscience.

“This is particularly intriguing for future studies because NMDA receptors are a major target for alcohol and contribute to important features of alcoholism, such as withdrawal. These new findings suggest that GluN2B in corticostriatal circuits may also play a key role in driving the transition from controlled drinking to compulsive abuse that characterizes alcoholism.”

Jul 10, 201379 notes
#addiction #alcoholism #prefrontal cortex #NMDA receptors #neural circuits #learning #neuroscience #science
Jul 10, 2013100 notes
#brain cancer #pilocytic astrocytoma #brain tumor #genes #mutations #genetics #neuroscience #science
Jul 10, 2013134 notes
#autism #ADHD #neurodevelopmental disorders #white matter #neuroimaging #neuroscience #science
New Research Points to Biomarker that Could Track Huntington’s Disease Progression

A hallmark of neurodegenerative diseases such as Alzheimer’s, Parkinson’s and Huntington’s is that by the time symptoms appear, significant brain damage has already occurred—and currently there are no treatments that can reverse it. A team of SRI International researchers has demonstrated that measurements of electrical activity in the brains of mouse models of Huntington’s disease could indicate the presence of disease before the onset of major symptoms. The findings, “Longitudinal Analysis of the Electroencephalogram and Sleep Phenotype in the R6/2 Mouse Model of Huntington’s Disease,” are published in the July 2013 issue of the neurology journal Brain, published by Oxford University Press.

image

SRI researchers led by Stephen Morairty, Ph.D., a director in the Center for Neuroscience in SRI Biosciences, and Simon Fisher, Ph.D., a postdoctoral fellow at SRI, used electroencephalography (EEG), a noninvasive method commonly used in humans, to measure changes in neuronal electrical activity in a mouse model of Huntington’s disease. Identification of significant changes in the EEG prior to the onset of symptoms would add to evidence that the EEG can be used to identify biomarkers to screen for the presence of a neurodegenerative disease. Further research on such potential biomarkers might one day enable the tracking of disease progression in clinical trials and could facilitate drug development.

“EEG signals are composed of different frequency bands such as delta, theta and gamma, much as light is composed of different frequencies that result in the colors we call red, green and blue,” explained Thomas Kilduff, Ph.D., senior director, Center for Neuroscience, SRI Biosciences. “Our research identified abnormalities in all three of these bands in Huntington’s disease mice. Importantly, the activity in the theta and gamma bands slowed as the disease progressed, indicating that we may be tracking the underlying disease process.”

EEG has shown promise as an indicator of underlying brain dysfunction in neurodegenerative diseases, which otherwise occurs surreptitiously until symptoms appear. Until now, most investigations of EEG in patients with neurodegenerative diseases and in animal models of neurodegenerative diseases have shown significant changes in EEG patterns only after disease symptoms occurred.

“Our breakthrough is that we have found an EEG signature that appears to be a biomarker for the presence of disease in this mouse model of Huntington’s disease that can identify early changes in the brain prior to the onset of behavioral symptoms,” said Morairty, the paper’s senior author. “While the current study focused on Huntington’s disease, many neurodegenerative diseases produce changes in the EEG that are associated with the degenerative process. This is the first step in being able to use the EEG to predict both the presence and progression of neurodegenerative diseases.”

Although previous studies have shown there are distinct and extensive changes in EEG patterns in Alzheimer’s and Huntington’s disease patients, researchers are looking for changes that may occur decades before disease onset.

Huntington’s disease is an inherited disorder that causes certain nerve cells in the brain to die, resulting in motor dysfunction, cognitive decline and psychiatric symptoms. It is the only major neurodegenerative disease where the cause is known with certainty: a genetic mutation that produces a change in a protein that is toxic to neurons.

Jul 9, 201370 notes
#science #neurodegenerative diseases #huntington's disease #neuronal activity #biomarkers #animal model #neuroscience
Brain and eye combined monitoring breakthrough could lead to fewer road accidents

Latest advances in capturing data on brain activity and eye movement are being combined to open up a host of ‘mindreading’ possibilities for the future. These include the potential development of a system that can detect when drivers are in danger of falling asleep at the wheel.

image

The research has been undertaken at the University of Leicester with funding from the Engineering and Physical Sciences Research Council (EPSRC), and in collaboration with the University of Buenos Aires in Argentina.

The breakthrough involves bringing two recent developments in the world of technology together: high-speed eye tracking that records eye movements in unprecedented detail using cutting-edge infra-red cameras*; and high-density electroencephalograph** (EEG) technology that measures electrical brain activity with millisecond precision through electrodes placed on the scalp.

The research has overcome previous technological challenges which made it difficult to monitor eye movement and brain activity simultaneously. The team has done this by developing novel signal processing techniques.

This could be the first step towards a system that combines brain and eye monitoring to automatically alert drivers who are showing signs of drowsiness. The system would be built into the vehicle and connected unobtrusively to the driver, with the EEG looking out for brain signals that only occur in the early stages of sleepiness. The eye tracker would reinforce this by looking for erratic gaze patterns symptomatic of someone starting to feel drowsy and different from those characteristic of someone driving who is constantly looking out for hazards. Fatigue has been estimated to account for around 20 per cent of traffic accidents on the UK’s motorways.***

The breakthrough achieved by the University of Leicester could also ultimately be built on to deliver many other everyday applications in the years ahead. For example:

  • Computer games of the future could dispense with the need for the player to physically interact with any type of console, mouse or other hand-operated system. Instead, eye movement and brain activity data would be collected and processed to indicate what action the player wants to take. By distinguishing the tiny differences in various types of brain activity, the EEG would identify the precise action the player desires (e.g. run, jump or throw), while the eye movement data would show exactly where on the screen the player was looking when they had this thought. This information could be combined to enable the correct action to occur. An unobtrusive headset would be all that would be required to capture the necessary data.
  • People who have no arm functionality could move their wheelchairs simply through their eye movements. These movements could be tracked and the corresponding brain activity analysed to identify when these indicate a desire to move in a certain direction. This would then automatically activate a steering and propulsion mechanism that would drive the wheelchair to that place.
  • The breakthrough could also provide the basis for improved tests to diagnose dyslexia and other reading disorders. Current tests revolve around a rapid succession of single words flashed onto a computer screen, with the resulting brain activity monitored by EEG. The new technique could enable the person being tested to move their eyes and read longer passages of text in a natural way, making the tests much more realistic and revealing.
  • With the basic concept now demonstrated successfully, the team aim to continue their work and eventually develop software that, in real time, automatically monitors both eye movement and brain activity.
  • Dr Matias Ison, who has led the research, says: “Historically, eye-tracking and EEG have evolved as independent fields. We have managed to overcome the challenges that were standing in the way of integrating these technologies. This is already leading to a much better understanding of how the brain responds when the eyes are moving. Monitoring the alertness of drivers is just one of many potential applications for this work. Building on the foundation provided by our EPSRC-funded project, we hope to see the first of these starting to become feasible within the next three to five years.”
Jul 9, 201344 notes
#brain activity #eye movement #eye tracking #EEG #neuroscience #science
Jul 9, 2013107 notes
#vOICe #sensory substitution #blindness #sensory substitution device #acuity #neuroscience #science
New tissue engineering breakthrough encourages nerve repair

A new combination of tissue engineering techniques could reduce the need for nerve grafts, according to new research by The Open University. Regeneration of nerves is challenging when the damaged area is extensive, and surgeons currently have to take a nerve graft from elsewhere in the body, leaving a second site of damage. Nerve grafts contain aligned tissue structures and Schwann cells that support and guide neuron growth through the damaged area, encouraging function to be restored. The research, published in Biomaterials, reported a way to manufacture artificial nerve tissue with the potential to be used as an alternative to nerve grafts.

Pieces of Engineered Neural Tissue (EngNT) are formed by controlling natural Schwann cell behaviour in a three-dimensional collagen gel so that the cells elongate and align, then a stabilisation process removes excess fluid to leave robust artificial tissues. These living biomaterials contain aligned Schwann cells in an aligned collagen environment, recreating key features of normal nerve tissue.

Incorrect orientation of regenerating nerve cells can lead to delays in repair, scarring and poor restoration of nerve function. Much research has taken place into how support cells (Schwann cells) can be combined with materials to guide nerve regeneration. The new technology from The Open University avoids the use of synthetic materials by building neural tissue from collagen, a protein that is abundant in normal nerve tissue. Building the artificial tissue from natural proteins and directing the cellular alignment using normal cell-material interactions means the EngNT can integrate effectively at the repair site.

Dr James Phillips, Lecturer in Health Sciences at The Open University, said: “We previously reported how self-alignment of Schwann cells could be achieved by using a tethered collagen hydrogel, which exploited cells’ natural ability to orientate in the appropriate direction by using their internal contraction forces. Our current research shows that cell-alignment in the hydrogel can be stabilised using plastic compression. The compression removes fluid from the gels, leaving a strong and stable aligned structure that has many features in common with nerve tissue.”

The team incorporated Schwann cells within the aligned material to form artificial neural tissue that could potentially be used in peripheral nerve repair. The technique could be applied to other regenerative medicine scenarios, where a stable artificial tissue containing aligned cellular architecture would be of benefit.

Jul 9, 201390 notes
#science #tissue engineering #artificial nerve tissue #schwann cells #nerve cells #neuroscience
Jul 9, 201388 notes
#science #alzheimer's disease #brain inflammation #dementia #amyloid deposition #neuroscience
Jul 9, 2013102 notes
#alzheimer's disease #cognitive impairment #grey matter #memory #parahippocampal gyrus #neuroscience #science
Jul 9, 201350 notes
#human-robot interaction #AI #robotics #robots #psychology #neuroscience #science
Jul 9, 201391 notes
#science #mitochondria #membrane potential #neurodegenerative diseases #synaptic plasticity #neuroscience
Jul 9, 2013171 notes
#memory #LTM #calcium #cell nucleus #neurons #memory proteins #neuroscience #science
Brain Structural Deficits May Contribute to Increased Functional Connections Between Brain Regions Implicated in Depression

Major depressive disorder is associated with a dysregulation of brain regions including the prefrontal cortex and limbic system. The relationship between structural and functional abnormalities in these brain regions in depressed patients is far from clear. However, both types of changes are assumed to underlie the symptoms of this disorder.

This lack of understanding prompted Dr. Bart de Kwaasteniet at the Academic Medical Center in Amsterdam and his colleagues to use a multimodal neuroimaging approach to further investigate this relationship.

The researchers, led by Professor Damiaan Denys, recruited 18 patients with major depressive disorder and 24 healthy individuals, all of whom underwent multiple neuroimaging scans. They specifically focused on the structural and functional connectivity between the subgenual anterior cingulate cortex (ACC) and the medial temporal lobe, two regions that are connected by a white matter tract called the uncinate fasciculus. These regions are known to be involved in the regulation of emotion and memory.

de Kwaasteniet explained their findings: “We identified decreased structural integrity of the uncinate fasciculus connecting the medial temporal lobe and the subgenual ACC. Furthermore, we identified an increased functional connection between these regions in major depression relative to controls. Importantly, we identified a negative correlation between the integrity of this white matter tract and the functional connection between the subgenual ACC and bilateral hippocampus in major depression.”

These results suggest that structural disturbances in the uncinate fasciculus contribute to abnormally high functional interactions among brain circuits associated with the symptoms of depression. “This leads to the hypothesis that abnormalities in brain structure lead to differences in connectivity between brain areas in depressive disorder,” added de Kwaasteniet.

However, they also hypothesized that the reverse may be true as well. In other words, that the increased functional connectivity among these brain regions leads to structural changes in the brain’s white matter fibers by means of an abnormally increased signal transduction. This hypothesis is supported by recent studies in schizophrenia which suggest that circuit hyperactivity may be a predictor of subsequent cortical atrophy.

"This interesting study suggests that abnormalities in the structural connections between brain regions, the white matter, are associated with abnormal activity within a brain circuit implicated in the symptoms of depression. This observation raises an important question about the implications of treating the circuit functional abnormalities without fixing the underlying brain structure," commented Dr. John Krystal, Editor of Biological Psychiatry. “Perhaps the structural abnormalities contribute to the risk for the relapse of depression among individuals whose brain circuit activity has responded to antidepressant medications.”

More research will be necessary to test the theories generated from the findings of this study.

Jul 9, 201379 notes
#prefrontal cortex #depression #white matter #brain circuitry #uncinate fasciculus #neuroscience #science
Jul 9, 2013549 notes
#science #stress #brain development #CRF-1 gene #genetics #psychology #neuroscience
…treating neurological diseases and computers that see!

Some 165 million Europeans are likely to experience some form of brain-related disease during their life. As the population ages, Alzheimer’s and other neurodegenerative or age-related mental disorders are affecting more people and contributing to higher health costs. Finding better ways of preventing and treating brain diseases is therefore becoming urgent, and understanding how our brains work is important to keep our economies at the forefront of new information technologies and services. EU-funded research is answering these challenges.

image

As mentioned in the first part of this article, this May the European Commission announced EUR 150 million of funding for 20 new ICT research projects expected to deliver new insights and innovations relating to traumatic brain injury, mental disorders, pain, epilepsy and paediatric conduct disorders.

The European Commissioner for Research, Innovation and Science, Máire Geoghegan-Quinn has said, ”Treating those affected (by brain-related disease) is already costing us EUR 1.5 million every minute […] Brain research could help alleviate the suffering of millions of patients and those that care for them. Unlocking the secrets of how the brain works could also open up a whole new universe of services and products for our economies.”

Treating neurological diseases

Stroke is the most common neurological disease to afflict people, causing cognitive problems - such as difficulties with attention, memory or language - or severe physical disability. The incidence increases with age, making it the most frequent cause of life-long impairment in adulthood.

These effects tend to increase patients” dependence on other people, and this lost autonomy can then lead to depression. The CONTRAST project seeks to bridge the gap between institutional rehabilitation and monitoring of the patient at home.

The project is developing an adaptive ”human-computer interface” (HCI) to improve cognitive functioning, offering training modules that improve the recovery of attention and memory. Patients will be able to go through an individually tailored rehabilitation process at home at the computer, while their doctor provides home-based training and monitors their progress from the clinic.

A third of stroke patients will experience long-term physiological or cognitive disabilities - preventing them from maintaining independent lives. COGWATCH aims to enhance the rehabilitation of stroke patients with symptoms of ”apraxia and action disorganisation syndrome” (AADS). Such patients retain their motor capabilities but commit cognitive errors during every-day goal-oriented tasks.

The project is developing intelligent tools and objects, portable and wearable devices, and ambient systems to provide personalised cognitive rehabilitation at home for stroke patients with AADS symptoms. By providing persistent feedback, the system will help to re-train patients on how to carry out the everyday activities they need to be independent.

Parkinson’s disease is another neurodegenerative disorder that is growing in incidence as our population ages - it particularly affects areas of the brain that are involved in movement control. The CUPID project aims to develop innovative, personalised rehabilitation at home for people with Parkinson”s disease, based on the patient”s needs.

The CUPID service will employ wearable sensors, audio biofeedback, virtual reality and external cueing to provide intensive motivating training that is suited to the patient and monitored remotely - decreasing the need for travel to a rehabilitation centre.

By the end of its first year, in December 2012, the project had designed the rehabilitation exercises and developed prototype virtual games for these exercises, as well as the telemedicine infrastructure needed for remote supervision.

Epilepsy is another common neurological disorder that, despite progress in treatment, is still incurable. Nowadays, pharmaceutical treatment can reduce or remove the symptoms, but this needs life-long continuous adjustment in order to be effective. The condition therefore requires monitoring of multiple parameters for accurate diagnosis, prediction, alerting and prevention, as well as treatment follow-up and presurgical evaluation.

The ARMOR project is designing a more holistic, personalised, medically efficient and economical monitoring system to analyse brain and body data from epilepsy patients. This portable system will provide more accurate diagnosis for individual patients, and allow better understanding and prediction of the time and type of their seizures - helping to give a warning and ensure the availability of medical assistance and advice if necessary.

Amputation of a limb is not just a traumatic physical experience. It can also lead to sensations - usually accompanied by pain - that seem to come from the missing body part, called a ”phantom limb”. The TIME project is developing an alternative treatment for phantom limb pain based on a new ”human-machine interface” (HMI) and selective, electrical stimulation of the peripheral nerves.

Using an implantable electrode placed inside the nerve, and electrical stimulators placed outside the body, the system will provide electrical micro stimulation to help reduce painful sensations - and may even have applications such as enabling amputees to sense virtual environments by touch.

Seeing things

The potential of such techniques doesn’t stop at monitoring, diagnosis and managing chronic conditions. The OPTONEURO project could ultimately help return functional sight to blind people.

”Optogenetics” is an exciting new gene therapy technique that makes nerve cells sensitive to particular colours of light. Simple pulses of intense light cause these photosensitised nerve cells to fire ”action potentials”, the carriers of information in the nervous system. To activate the nerve cells, however, the new therapy depends on high illumination densities - bright light shining on very small areas.

The OPTONEURO project therefore aims to develop the complementary optoelectronics needed to stimulate these photosensitised neurons. The system would be scalable for applications both in basic neuroscience research and in ”neuroprosthesis”. In particular, the optoelectronics should be used in a future optogenetic-optoelectronic retinal prosthesis - an artificial eye - for those blinded by the ”retinitis pigmentosa” disease.

The project requires a team of specialists in photonics, micro-optics and neurobiology to develop an array of ultra-bright electronically controlled micro-LEDs, which could also provide a new research tool for the neuroscience and neurotechnology community.

The SEEBETTER project is also looking to develop artificial vision prosthetics for the blind. Conventional image sensors have severe limitations, but ”silicon retina” vision sensors aim to mimic the biological retina”s information processing - computing both spatial and temporal aspects of the visual input. To date, these silicon retinas suffer from low quantum efficiency - meaning low light sensitivity - and an inability to combine both spatial and temporal processing on the same chip.

SEEBETTER’s team of experts - from biology and biophysics, as well as biomedical, electrical and semiconductor engineering - aim to use genetic and physiological techniques to understand better the function of the retina and model the retina’s vision processing. They will then design and build the first high-performance silicon retina, implemented on a single silicon wafer, specialised for both spatial and temporal visual processing.

Understand the neurobiological principles of seeing - beyond the functioning of the retina alone - may help us to replicate the success of human vision for computers and robots. The RENVISION project aims to achieve a comprehensive understanding of how the retina encodes visual information through the different cellular layers and to use such insights to develop a retina-inspired computational approach to computer vision.

Using high-resolution 3D microscopy will allow the researchers to make images of the inner retinal layers at near-cellular resolution. This new knowledge on retinal processing will help develop advanced pattern recognition and machine-learning technologies. The project could therefore solve some of the most difficult tasks in computer vision - such as automated scene categorisation and human action recognition - so that robots and computers can see and perceive what is happening in the images they receive.

These are just some of the EU-funded ICT projects using electronics and computing technologies to understand, augment and improve the human brain and its functioning. The results have the potential to reduce the impact of disability and disease, and improve our computing power, IT infrastructure and economy.

Jul 8, 201380 notes
#neurological diseases #cognitive functioning #brain research #optogenetics #neuroscience #science
Jul 8, 2013143 notes
#autism #gut microbiome #gut bacteria #medicine #science
Dissecting the Distinctive Walk of Disease

Pitt multidisciplinary research team proposes mathematical model that examines multiple walking patterns and movements in adults older than 65

Older adults diagnosed with brain disorders such as Parkinson’s disease often feel a loss of independence because of their lack of mobility and difficulty walking. To better understand and improve these mobility issues—and detect them sooner—a University of Pittsburgh multidisciplinary research team is working toward building a more advanced motion test that addresses a wider range of walking patterns and movements.

image

In a recent issue of IEEE Transactions on Neural Systems and Rehabilitation Engineering, researchers from Pitt’s Swanson School of Engineering, School of Health and Rehabilitation Sciences, and School of Medicine propose a mathematical model that can examine multiple walking, or gait-related, features in healthy and clinical populations. To date, no study has brought together such a team to examine such a high number of movement features comparing healthy and clinical older adults. Previous studies have typically only measured one or two types of movement features in just one population. 

“Right away, you can tell whether an older individual has difficulties walking by conducting a simple gait test,” said Ervin Sejdic, lead author of the paper and an assistant professor of engineering in the Swanson School. “But can we quantify these changes and document them earlier? That’s the biggest issue here and what we’re trying to model.”

Thirty-five adults older than 65 were recruited for the study, including 14 healthy participants, 10 individuals with Parkinson’s disease, and 11 adults who had impaired feeling in their legs owing to peripheral neuropathy (nerve damage). Walking trials were performed using a computer-controlled treadmill, and participants wore an accelerometer—a small box attached with a belt—and a set of reflective markers on their lower body that allowed for tracking of the participants’ movements through a camera-based, motion-analysis system. These two systems allowed the team to examine the torso and lower body movements of patients as they walked. Participants completed three walking trials on the treadmill—one at a usual walking pace, another while walking slowly, and another that included working on a task while walking (i.e. pushing a button in response to a sound). 

The accelerometer signals were used to examine three aspects of movement: participants moving forward and backward, side to side, and up and down. The researchers then used advanced mathematical computations to extract data from these signals. 

The results—integrated into the mathematical models—showed significant differences between the healthy and clinical populations. These metrics were able to discriminate between the three groups, identifying critical features in how the participants walked. 

The Pitt team is now looking to conduct this type of study on a larger scale—evaluating the gait patterns of older adults residing within independent living facilities. 

“Our results indicate that we can potentially develop these mathematical models as biomarkers to predict changes in walking due to diseases like Parkinson’s disease,” said Sejdic. “Now, we want to take it further. We’re especially hoping to help those individuals in independent living facilities by predicting the declines in their walking even earlier.”  

“What also makes this study unique is the multidisciplinary team approach we used,” said Jennifer S. Brach (SHRS ’94G, ’00G) coprincipal investigator of the study and associate professor in Pitt’s Department of Physical Therapy. “Here we brought together a research team that included engineers, physical therapists, and experts in geriatrics to work on an important problem in older adults—changes in mobility.”

Jul 7, 201348 notes
#parkinson's disease #walking patterns #mathematical model #peripheral neuropathy #neuroscience #science
Jul 7, 2013226 notes
#humor #humorous creativity #creativity #terror management #mortality #psychology #neuroscience #science
Semantics on the basis of words’ connectivity

It is now possible to identify the meaning of words with multiple meanings, without using their semantic context

image

Two Brazilian physicists have now devised a method to automatically elucidate the meaning of words with several senses, based solely on their patterns of connectivity with nearby words in a given sentence – and not on semantics. Thiago Silva and Diego Amancio from the University of São Paulo, Brazil, reveal, in a paper about to be published in EPJ B, how they modelled classics texts as complex networks in order to derive their meaning. This type of model plays a key role in several natural processing language tasks such as machine translation, information retrieval, content analysis and text processing.

In this study, the authors chose a set of ten so-called polysemous words—words with multiple meanings—such as bear, jam, just, rock or present. They then verified their patterns of connectivity with nearby words in the text of literary classics such as Jane Austen’s Pride and Prejudice. Specifically, they established a model that consisted of a set of nodes representing words connected by their “edges,” if they are adjacent in a text.

The authors then compared the results of their disambiguation exercise with the traditional semantic-based approach. They observed significant accuracy rates in identifying the suitable meanings when using both techniques. The approach described in this study, based on a so-called deterministic tourist walk characterisation, can therefore be considered a complementary methodology for distinguishing between word senses.In future works, the authors are planning to devise new measures to connect not only adjacent words, but also words within a given interval in order to enhance the ability of the model to grasp semantic factors. This approach is supported by another recent study by the same authors, showing that traditional complex network measures mainly depend on the syntax.

Jul 7, 2013135 notes
#language #semantics #complex networks #learning techniques #tourist walk #neuroscience #science
Jul 6, 2013167 notes
#brain mapping #dorsomedial prefrontal cortex #temporoparietal junction #psychology #neuroscience #science
Jul 5, 201382 notes
#primates #great apes #evolution #genetic variation #genetics #genomics #science
Jul 5, 2013428 notes
#brain #hearing #inflammation #lip movements #McGurk illusion #neuroscience #science
Next page →
20132014
  • January
  • February
  • March
  • April
  • May
  • June
  • July
  • August
  • September
  • October
  • November
  • December
201220132014
  • January
  • February
  • March
  • April
  • May
  • June
  • July
  • August
  • September
  • October
  • November
  • December
20122013
  • January
  • February
  • March
  • April
  • May
  • June
  • July
  • August
  • September
  • October
  • November
  • December