Robustness and fragility in neural development

So many things can go wrong in the development of the human brain it is amazing that it ever goes right. The fact that it usually does – that the majority of people do not suffer from a neurodevelopmental disorder – is due to the property engineers call robustness. This property has important implications for understanding the genetic architecture of neurodevelopmental disorders – what kinds of insults will the system be able to tolerate and what kind will it be vulnerable to?

The development of the brain involves many thousands of different gene products acting in hundreds of distinct molecular and cellular processes, all tightly coordinated in space and time – from patterning and proliferation to cell migration, axon guidance, synapse formation and many others. Large numbers of proteins are involved in the biochemical pathways and networks underlying each cell biological process. Each of these systems has evolved not just to do a particular job, but to do it robustly – to make sure this process happens even in the face of diverse challenges.

Robustness is an emergent and highly adaptive property of complex systems that can be selected for in response to particular pressures. These include extrinsic factors, such as variability in temperature, supply of nutrients, etc., but also intrinsic factors. A major source of intrinsic variation is noise in gene expression – random fluctuations in the levels of all proteins in all cells. These fluctuations arise due to the probabilistic nature of gene transcription – whether a messenger RNA is actively being made from a gene at any particular moment. The system must be able to deal with these fluctuations and it can be argued that the noise in the system actually acts as a buffer. If the system only worked within a narrow operating range for each component then it would be very vulnerable to failure of any single part.

Natural selection will therefore favour system architectures that are more robust to environmental and intrinsic variation. In the process, such systems also indirectly become robust to the other major source of variation – mutations.

Many individual components can be deleted entirely with no discernible effect on the system (which is why looking exhaustively for a phenotype in mouse mutants can be so frustrating – many gene knockouts are irritatingly normal). You could say that if the knockout of a gene does not affect a particular process, that that means the gene product is not actually involved in that process, but that is not always the case. One can often show that a protein is involved biochemically and even that the system is sensitive to changes in the level of that protein – increased expression can often cause a phenotype even when loss-of-function manipulations do not.

Direct evidence for robustness of neurodevelopmental systems comes from examples of genetic background effects on phenotypes caused by specific mutations. While many components of the system can be deleted without effect, others do cause a clear phenotype when mutated. However, such phenotypes are often modified by the genetic background. This is commonly seen in mouse experiments, for example, where the effect of a mutation may vary widely when it is crossed into various inbred strains. The implication is that there are some genetic differences between strains that by themselves have no effect on the phenotype, but that are clearly involved in the system or process, as they strongly modify the effect of another mutation.

How is this relevant to understanding so-called complex disorders? There are two schools of thought on the genetic architecture of these conditions. One considers the symptoms of, say, autism or schizophrenia or epilepsy as the consequence of mutation in any one of a very large number of distinct genes. This is the scenario for intellectual disability, for example, and also for many other conditions like inherited blindness or deafness. There are hundreds of distinct mutations that can result in these symptoms. The mutations in these cases are almost always ones that have a dramatic effect on the level or function of the encoded protein.

The other model is that complex disorders arise, in many cases, due to the combined effects of a very large number of common polymorphisms – these are bases in the genome where the sequence is variable in the population (e.g., there might be an “A” in some people but a “G” in others). The human genome contains millions of such sites and many consider the specific combination of variants that each person inherits at these sites to be the most important determinant of their phenotype. (I disagree, especially when it comes to disease). The idea for disorders such as schizophrenia is that at many of these sites (perhaps thousands of them), one of the variants may predispose slightly to the illness. Each one has an almost negligible effect alone, but if you are unlucky enough to inherit a lot of them, then the system might be pushed over the level of burden that it can tolerate, into a pathogenic state.

These are the two most extreme positions – there are also many models that incorporate effects of both rare mutations and common polymorphisms. Models incorporating common variants as modifiers of the effects of rare mutations make a lot of biological sense. What I want to consider here is the model that the disease is caused in some individuals purely by the combined effects of hundreds or thousands of common variants (without what I call a “proper mutation”).

Ironically, robustness has been invoked by both proponents and opponents of this idea. I have argued that neurodevelopmental systems should be robust to the combined effects of many variants that have only very tiny effects on protein expression or function (which is the case for most common variants). This is precisely because the system has evolved to buffer fluctuations in many components all the time. In addition to being an intrinsic, passive property of the architecture of developmental networks, robustness is also actively promoted through homeostatic feedback loops, which can maintain optimal performance in the face of variations, by regulating the levels of other components to compensate. The effects of such variants should therefore NOT be cumulative – they should be absorbed by the system. (In fact, you could argue that a certain level of noise in the system is a “design feature” because it enables this buffering).

Others have argued precisely the opposite – that robustness permits cryptic genetic variation to accumulate in populations. Cryptic genetic variation has no effect in the context in which it arises (allowing it to escape selection) but, in another context – say in a different environment, or a different genetic background – can have a large effect. This is exactly what robustness allows to happen – indeed, the fact that cryptic genetic variation exists provides some of the best evidence that we have that the systems are robust as it shows directly that mutations in some components are tolerated in most contexts. But is there any evidence that such cryptic variation comprises hundreds or thousands of common variants?

To be fair, proving that is the case would be very difficult. You could argue from animal breeding experiments that the continuing response to selection of many traits means that there must be a vast pool of genetic variation that can affect them, which can be cumulatively enriched by selective breeding, almost ad infinitum. However, new mutations are known to make at least some contribution to this continued response to selection. In addition, in most cases where the genetics of such continuously distributed traits have been unpicked (by identifying the specific factors contributing to strain differences for example) they come down to perhaps tens of loci showing very strong and complex epistatic interactions (1, 2, 3). Thus, just because variation in a trait is multigenic, does not mean it is affected by mutations of small individual effect – an effectively continuous distribution can emerge due to very complex epistatic interactions between a fairly small number of mutations which have surprisingly large effects in isolation.

(I would be keen to hear of any examples showing real polygenicity on the level of hundreds or thousands of variants).

In the case of genetic modifiers of specific mutations – say, where a mutation causes a very different phenotype in different mouse strains – most of the effects that have been identified have been mapped to one or a small number of mutations which have no effect by themselves, but which strongly modify the phenotype caused by another mutation.

These and other findings suggest that (i) cryptic genetic variation relevant to disease is certainly likely to exist and to have important effects on phenotype, but that (ii) such genetic background effects can most likely be ascribed to one, several, or perhaps tens of mutations, as opposed to hundreds or thousands of common polymorphisms.

This is already too long, but it begs the question: if neurodevelopmental systems are so robust, then why do we ever get neurodevelopmental disease? The paradox of systems that are generally robust is that they may be quite vulnerable to large variation in a specific subset of components. Why specific types of genes are in this set, while others can be completely deleted without effect, is the big question. More on that in a subsequent post…

America: as if it is 1970

I noticed that The Washington Post had an article up, Number of biracial babies soars over past decade, based on 2010 Census data. I was immediately curious if my expectations were correct in this case, because the term “biracial” has a very specific connotation. That is, there are two races, and in America that is black and white. If you want to break out of this old dichotomy you usually say multiracial. This paradigm has a historical valence, because the “race issue” in America has traditionally been in black and white, with a minor secondary role for native populations. I say traditionally, because by any measure the minority of America’s minorities are now black.

And sure enough the article does focus on the black-white dimension, with honorable mention for a woman of Asian heritage. But it is notionally based on the Census, right? It was easy to find the press release on the Census website. Here is the table accompanying it:

 

Read More

The last days of Grendel

A new paper in Science has just been published which in its broad outlines has been described in conference presentations. When examining the autosomal genetic variation of three individuals of the hunter-gatherer Pitted Ware Culture (PWC), and one of the agriculturalist Funnel Beaker Culture (TRB), the authors found that the two groups were sharply differentiated. The number of SNPs was on the order of 10,000 or so if I read the methods correctly. This is rather thin for studying contemporary within European population differences (~100,000 or more seems to be safe), in particular using hypothesis based clustering algorithms (it seems more manageable for PCA). But the findings are strong enough that I think we shouldn’t discount them. The most fascinating aspect of the results is that while the PWC seem to exhibit affinities with Northern and Northeastern Europeans, the TRB individual seems more similar to extant Southern Europeans!

Others have already commented extensively on the results. Keeping in mind the small sample sizes, limitation of comparisons, and the relatively thin marker set, I think the primary result we can take away from these findings is that old models of pure cultural and demographic diffusion are false. By this, I mean that prior debates which culminated in the early aughts on the “Paleolithic vs. Neolithic” contribution to the ancestry of modern Europeans were fundamentally premised on a demographic diffusion dynamic, whereby genes and ideas exhibited a continuous flow across a flat and featureless landscape. On the contrary, the basic outlines we are seeing here is that the human past exhibited spatial and temporal discontinuity. And why should this surprise us? There is no dialect continuum between Spanish and Chinese across Eurasia. Rather, broad language families are sharply differentiated from each other at zones of contact. Though there are theoretical reasons why the variation in genes should be more clinal, the reality remains that cultural parameters are going to shape the outlines of genetic variation, and those parameters are discontinuous.

Read More

Types of genetics

  • Molecular genetics
  • Developmental genetics
  • Population genetics
  • Quantitative genetics
  • Phylogenetics

Thoughts? Recently had a discussion whether phylogeneticists considered themselves geneticists (qualified “no”). Quantitative genetics really evolved out of biometrics, which actually opposed Mendelian genetics. You can construct quantitative genetics from Mendelian first principles, but it is not necessary. As for population vs. molecular, ask each group what they mean by “gene.” Modern developmental geneticists seem to be closely aligned with molecular geneticists.

Leaning the wrong way?

Many of the people I socialize with in “real life” have a biological sciences background. That being said, a relatively deep understanding of ncRNA does not give you any better sense of behavior genetics than the person off the street. And of course when you have a small child conversation often goes in the direction of how you want to raise the child so as to maximize their outcomes. Setting aside the particular normative valence of those outcomes, I am always struck by the power people think parents have over their child’s life path. This is not to say parents don’t have power. There are many young people who have college degrees because of parental expectations. Or, perhaps more precisely the social expectations which the parents set in motion by selecting the milieu of one’s children. Yet so many times I’ve been in a conversation where the phrase “I lean toward nurture” has come up. These are not dogmatic “blank slate” individuals. Rather, they are simply falling back upon the null or default of our age.

But for me here is the irony: I think it is arguably the case today we live in a world where nurture matters far less in variation in outcome of exactly the people who ‘lean toward nurture.’ Let me repeat: when you remove environmental variation by providing a modicum of comfort , you are left with genetic variation! There were times in the past when ‘nurture,’ in other words the hand that environment dealt, was much more influential. And yet during those periods it was nature which was ascendant.

In 2004 the General Social Survey asked a question where respondents were asked to decide between “genes play major role in determining personality” and “experience determines personality.” For various reasons I do not think that the question was good, but, the responses are illustrative of the unanimity we’ve achieved in American society on some questions.

Read More

One baby, alone on a PCA island

A week ago I reported that according to 23andMe I’m 40% Asian, and she is 8% Asian (in the future if I say “she” without explanation, you know of whom I speak). Obviously something is off here. The situation resolved itself when I tuned my parameters and increased my sampled populations in Interpretome. By now I’ve already done the estimates of recombination on the chromosomes which came together to produce her, and the realized value of 8 percent instead of 20 percent “Asian” simply can not be due to a particular set of unlikely crossing over events. From what I can gather it seems like ancestry painting should be viewed as a qualitative rather than a quantitative assessment. This sounds really strange when you are given percentages, but the results are strange, and obviously wrong too often in terms of the specific values.

Here’s an admixture plot which shows more realistically informative values:

Read More

The world is as it should be in personal genomics

I’ve been having some fun with my daughter’s personal genomics. You see, she has her whole pedigree out to r = 1/4. So, for example, contributions from her grandparents seem to be about on this order:

Paternal grandfather = 0.28
Paternal grandmother = 0.22
Maternal grandfather = 0.23
Maternal grandmother = 0.27

I’ve also calculated the number of recombinations which occurred leading up to the gametes which fused to create her. That will be for a future post. But here let’s confirm that she is not inbred. I used plink for this. Here is the description of the command:

Given a large number of SNPs, in a homogeneous sample, it is possible to calculate inbreeding coefficients (i.e. based on the observed versus expected number of homozygous genotypes).

The estimate of F can sometimes be negative. Often this will just reflect random sampling error, but a result that is strongly negative (i.e. an individual has fewer homozygotes than one would expect by chance at the genome-wide level) can reflect other factors, e.g. sample contamination events perhaps.

My main confusion here was which population I should select? Should I select GIH (HapMap Gujaratis?) or CEU (Utah whites)? I ended up on the TSI sample (Tuscans) as a fake compromise. And of course, because she is mixed-race the results came out very negative, as she had way less homozygosity than would be “expected” from the population wide statistic. I also added an inbred friend (his parents are first cousins) as a “control.” Below are two plots which show the result.

Read More

An algorithm is just an algorithm

In the comments below:

You should include a Moroccan or otherwise native North African sample. Without a North African sample West Africans act as proxy for some of that North African ancestry that does exist in Iberia, specially the Western third (Portugal, Galicia, Extremadura, León, etc.) Doing that your analysis would become more precise and you could make better informed claims.

I was reading through all the entry and there was no mention to the rather surprising notable West African component in Iberians other than Basques. For my somewhat trained eye it is clear that this is a proxy for North African ancestry and not directly West African ancestry. This is demonstratedly also the case in Canary Islands, at least to a large extent, and, by extension in Cuba (which is nearly identical to your average Canarian), at least Cuba-1. Cuba-2 seems actually admixed at low levels and both seem to have some Amerindian ancestry not existent in Spain.

This is a fair point. I switched computers recently, and the Behar et al. data set I had seems to have become corrupted. So I snatched the Mozabites from the HGDP, and removed the Gujaratis from the previous run. I also added Russians, Druze, and some extra Amerindian groups. At K = 7 this pattern jumped out:

Read More

Paternity most assured

The myth that 10 percent of children the product of ‘non-paternity events’ is rather persistent. I have no idea why, but I do know that even biologists accept it. But how we can we continue to accept this when surnames can provide population genetic information 400 years after the fact? The population of Belgium is famously divided between Latinate Walloons and Germanic Flemings. But is notable that a substantial number of Flemings carry surnames of clear Romance origin. This is in large part due to acculturation. Nevertheless, even 400 years after the largest of the migration and assimilation events males with Romance-origin surnames reflect their genetic background:

 

If non-paternity events occurred at a rate of 1 out of 10 the correlation between surnames and genetic lineage would have been decoupled long ago. These results have been confirmed in other societies. I predict that low non-paternity rates will also be confirmed in China; as that nation has a long history of surnames. Of course, one might posit a scenario where males who are the products of non-paternity events tend to be less fit than those who are not, so over the long term these estimates based on present day Y chromosomal lineages may not be appropriate reflective of the frequency of events at some point in the past.

The culture that is Microsoft

Frustration, Disappointment And Apathy: My Years At Microsoft:

Large companies have overheads, a necessary evil, you say. Overheads need to be managed. And managed they are: Group Managers, Program managers, General managers, together with ‘Senior’ flavours of those and a whole new breed of directors, stakeholders, business owners, relationship leads coupled with their own countless derivatives.

All those meeting-goers are not making anything. Deciding upon and making something is hard. And if this onerous activity has to be done, then hire external consultants for it. It’s easier and less risky.

There is no creative tension, no vision these days. Left to Microsoft’s hands we’d still be toiling on overheating Vista desktops.

This company is becoming the McDonalds of computing. Cheap, mass products, available everywhere. No nutrients, no ideas, no culture. Windows 8 is a fine example. The new Metro interface displays nonstop, trivial updates from Facebook, Twitter, news sites and stock tickers. Streams of raw noise distract users from the moment they login.

The rise of sclerotic bureaucratic intermediaries isn’t just a problem with Microsoft. Remember when parasitic squid were generating 40 percent of the economy’s profits? It’s no better in academia: