Race debate

John Wilkins started a debate on race. I responded here & here. John responds to the response. Interesting discussion threads on Pharyngula, here & here. The Contingency Table adds his 2 cents. You all know about this topic, but I figured you might want pointers to the “discussion.” I’ve stated at least twice that I’m an instrumentalist on this issue, here is the definition from Wiki:

In the philosophy of science, instrumentalism is the view that concepts and theories are merely useful instruments whose worth is measured not by whether the concepts and theories are true or false (or correctly depict reality), but by how effective they are in explaining and predicting phenomena.

For the record, I’m close to an instrumentalist on the “species concept” debate too.

Update: My last response to John’s last response. Evolgen also offers his 2 cents.

I say “race” and you say “population substructure”

John responds to the “race” response from Matt & I. I’m not interested in making a point-by-point response to the response because I don’t think the “objective” difference in opinion is that great, rather, it seems to be that we are clashing in the turbulent waters of nominalism.
First, I will respond to what I believe is the perception by John that I am conceiving of race as an essential and fundamental taxonomical unit. I don’t hold to that. I’ve rejected the Platonic conception of race before. The problem that “race based public policy” often has is that the legal system is deterministic, and probablistic entities like subpopulations don’t fit into neat categories. I don’t deny that human populations exhibit spatial and temporal gradients. I stated before that my attitude toward population substructure is instrumentalist. I am happy to treat human beings as fundamentally the same, having equal rights before God & Nature. This the basal starting point, but, we often just stop there.

Read More

Haldane's Dilemma: should we worry?

In previous posts and comments I have mentioned Haldane’s Dilemma, and I thought it would be worth looking into the subject more closely.

The short answer to the question, ‘Should we worry?’, is ‘No’. But when was I ever satisfied with a short answer? The following is not intended as a casual read but (hopefully) a resource for those who are following up the subject via search engines. As usual, I claim no technical expertise in genetics, so don’t take my assertions on trust. Links and references are provided for those who want to explore the subject further.

In a classic 1957 paper, J. B. S. Haldane estimated the number of generations of natural selection required to take an advantageous gene from rarity to near-fixation in a population. His headline conclusion was that it would normally take about 300 generations. More accurately, on average about one gene would be fixed in every 300 generations, so that if several genes were under selection simultaneously, the time required to fix each would be longer, but the number of genes fixed per unit of time would be the same. [Note 1] The total length of time required to fix T genes would therefore be approximately 300T generations. [Added: To avoid misunderstanding, I should have emphasised this last point. Haldane does not claim that every gene takes 300 generations to go from rarity to fixation. Suppose there is a long-established process of selection with a stream of 100 advantageous genes under selection at any time. Suppose each gene on average takes 30,000 generations to fix. If we take one of these genes at random, it therefore has a probability of 1 in 30,000 of being fixed in any given generation. But there are 100 genes to choose from, so the expected number of genes to be fixed in any given generation is 100 x 1/30,000 = 1/300. Equivalently, we expect on average 1 gene to be fixed in any 300 generations. This is what Haldane claims, based on his various assumptions.]

Based on the information available in 1957, Haldane estimated the number of gene differences between closely related species (not to be confused with the number of gene changes required for speciation) as typically about 1,000. It would therefore take about 300,000 generations for this amount of divergence to evolve by natural selection. Haldane thought this was broadly consistent with the fossil record, so he saw no ‘dilemma’.

In the 1960s new evidence showed more genetic diversity within and between species than was previously assumed. It therefore became problematic that natural selection seemed either too slow to explain the observed diversity or too costly in mortality for species to survive. Hence the ‘dilemma’. Motoo Kimura used Haldane’s figures to support his own theory of molecular evolution, maintaining that a large proportion of change at the molecular level is not due to selection but to genetic drift. In the resulting debate geneticists picked holes in Haldane’s analysis, and showed that in some circumstances natural selection could be much quicker than Haldane had claimed. (Haldane died in 1964, so he took no part in this debate.) By the mid-1970s it was generally accepted that Haldane’s Dilemma was not a serious problem. [Note 2]

The issue was reopened in 1992 by George C. Williams in his book Natural Selection: Domains, Levels and Challenges, where he argued that Haldane’s Dilemma had not been solved, but ‘merely faded away, because people got interested in other things’. The subject has also taken a bizarre twist in recent years as Haldane’s Dilemma has been seized on by Creationists as an objection to evolution. (Searching the Web for ‘Haldane’s Dilemma’, a large proportion of the results will be Creationist websites. A key Creationist example is here, and an evolutionist response is here.) This raises a dilemma of a different kind, as there is a danger that anyone who takes Haldane’s Dilemma seriously will be misrepresented as a crypto-Creationist, or misquoted to give support to Creationism. So just to be clear: I do not think Haldane’s Dilemma is a major problem for evolutionary theory. On the other hand, I do think that George C. Williams raised some interesting points. The real value of the Dilemma is now not so much to set any firm limit to the rate of evolution, as to focus attention on important questions about how natural selection works…

Haldane’s 1957 paper itself is available here. The arguments and calculations are not easy to follow, as the paper requires some prior knowledge of population genetics, and makes a lot of simplifying assumptions and approximations. There is also at least one serious misprint [note 3]. I will give my own outline of Haldane’s argument, not always using Haldane’s own notation.

First Haldane assumes that a population is in equilibrium with its environment, and that most members of the population have the best available genes (alleles) at most loci. Disadvantageous alleles are present at low frequencies (around 1 in 20,000 for dominants and 1 in 100 for recessives) due to a balance between selection and mutation.

The environment then changes so that some previously advantageous alleles are disadvantageous and vice versa. The newly disadvantageous alleles will each have a selective fitness detriment, s, relative to its advantageous alternative, where the s’s may be different for each allele. Any individual lucky enough to have the best alleles at all loci has fitness 1 (by stipulation), while the bearers of disadvantageous alleles have fitness (1 – s)(1 – s)(1 – s), etc, with the appropriate s’s for these alleles. Note that this assumes that the fitness effects of the alleles at different loci are independent, so that fitness is simply multiplicative. (If all the s’s are small, fitness is also approximately additive, as the product-terms involving more than one s can be neglected.) This assumption of independent fitness effects is not discussed by Haldane.

Haldane assumes for simplicity that all selection occurs by way of mortality in the juvenile stage, i.e. before reproduction. If the population at the time of selection is N, and the frequency of a disadvantageous allele is q, there will be a certain number of selective deaths attributable to the allele in that generation, depending on the size of N, q, s, and the genetic characteristics of the allele (e.g. dominance). The selective deaths reduce the population size, but we assume that it is restored to N in the next generation, an assumption not discussed by Haldane [note 4]. In the new generation the frequency of the disadvantageous allele is reduced, while that of the advantageous alternative is increased. By iterating the process of selection, and assuming that the selective advantage remains constant, we can calculate D, the total number of selective deaths, expressed as a multiple of N, required to increase the frequency of the advantageous allele from rarity to near-fixation. Provided s is small, the problem can be approximately solved by integration. The technical part of Haldane’s paper is concerned with performing these integrations for a variety of scenarios: haploid, diploid, partly recessive, sex-linked, and so on. The important conclusion is that over a wide range of parameters the resulting value of D does not vary drastically, though it is more sensitive to variations in the initial frequency q than in the fitness detriment s. For diploids D is usually between 10 and 100, with 30 as a representative value. This means that the total number of selective deaths required to take a single advantageous allele from rarity to near-fixation is about 30 times the population size. The process will therefore normally take at least 30 generations, but probably more: the best
estimate remaining to be determined.

This problem is taken up in the Discussion section of Haldane’s paper. The figure to be determined is the number of generations required for a single gene substitution. Call this K. As the total number of selective deaths required is about 30 times the population size, the number of selective deaths required per generation is about 30/K times the population size. However, usually more than one allele will be under selection at the same time. If each such allele involves d selective deaths per generation, then provided all the d’s are small, so that there is no significant overlap in the selective deaths attributable to different alleles, their sum, Sd, is approximately the total of selective deaths per generation. (Strictly the d’s and Sd are proportions of the population, rather than absolute numbers, but it is tedious to keep repeating this.) Since each gene substitution requires selective deaths totalling about 30 times the population size, Sd selective deaths can therefore be regarded as ‘paying for’ Sd/30 gene substitutions per generation. Equivalently, one gene substitution will on average occur every 30/Sd generations. But by definition this is also the desired number K, so we have K = 30/Sd. We can therefore estimate K if we know Sd. Here Haldane makes the assumption that the proportion of all mortality that is due to selection will usually be about 1/10, though occasionally as high as 1/2. (In most species under natural conditions, there are a lot of random, non-selective deaths, so Haldane’s assumption is not unreasonable.) Taking 1/10 as typical, we therefore have K = 30/[1/10] = 300. Hence we get the headline figure of 300 generations required for a gene substitution.

Given the assumption that fitness effects are independent, the selective deaths which ‘pay for’ an increase in the frequency of one advantageous allele are neutral with respect to all other alleles. Provided the selective advantage for each allele is not very large (say, not more than 10%), there will be little overlap in the selective deaths which pay for different advantageous alleles, so that the cost of each allele must be paid in full by 300 generations worth of selective deaths. [Note 5] (Haldane does not spell this out, but it is worth making it explicit.)

There are obviously a lot of approximations and simplifications in all this, and Haldane himself mentions cases, such as colonisation of a new territory, where selection might be quicker than average. But given Haldane’s key assumptions, his reasoning is generally accepted as sound. Criticism has therefore focused on the validity of the assumptions. A wide variety of objections and alternatives have been raised, some of which seem to me persuasive, others less so. A useful review in 1974 by V. Grant and R. Flake is available here. My own layman’s assessment is as follows:

1. The weakest of Haldane’s assumptions is that fitness effects are independent. This is crucial for his overall conclusions, so it is unfortunate that Haldane does not discuss it! A simple example shows how selection can be much faster with non-independent fitness. (I found this example in an article by Nick Barton and Linda Partridge, but they do not claim it as original.) Most of the cost of selection is incurred when an advantageous allele is rare. Suppose then that there are a number of rare advantageous alleles in a population. Suppose also that a proportion W of the population survives and breeds in each generation, and that all those individuals bearing any of the advantageous alleles are among the survivors. In this model all of the advantageous alleles will simultaneously increase to a frequency 1/W times their previous frequency in a single generation, so, for example, if W = 2/3, the frequency of each allele will jump to 1/[2/3] = 1.5 times its former frequency. There is no limit to the number of alleles that can be selected at this rate simultaneously, provided their total frequency in the population does not exceed W. In this simple model every selective death simultaneously helps pay for all the advantageous alleles, contrary to Haldane’s assumption of independent fitness.

A somewhat more complicated model was examined by John Maynard Smith in a 1968 paper. In this model JMS assumes that all those individuals with more than an average number of advantageous alleles survive and breed, while all those below this threshold die. JMS finds that with this assumption, of the order of 1,000 times as many alleles can be selected simultaneously, for a given overall intensity of selection, as in Haldane’s model. This implies that instead of on average 1 gene being fixed every 300 generations, there could be several genes fixed in each generation.

These examples suffice to show that selection much faster than Haldane supposed is possible. However, to show that something is possible is not the same as showing that it actually happens. One of George C. Williams’s main arguments in reviving Haldane’s Dilemma was that in practice the assumption of independent fitness effects was often a ‘realistic approximation’, and that more empirical research was needed into the actual patterns of selection. There may be some circumstances where Haldane’s assumption of independent fitness is reasonable. There may be others where fitness is not independent but the pattern of selection is not as favourable to rapid evolution as in the simple models described above. In a critique of Maynard Smith’s paper, P. O’Donald argued that a simple threshold model was unrealistic, and that it was more plausible that the fitness of a polygenic trait should be proportional to the square of its deviation from the optimum value. With a model based on this assumption, O’Donald found that selection could work somewhat faster than in Haldane’s model, but nowhere near as fast as in Maynard Smith’s.

2. In discussing Haldane’s Dilemma a distinction is often made between hard and soft selection. As introduced and defined by Bruce Wallace, a hard selective disadvantage applies regardless of the density of population. Soft selection, on the other hand, does not apply until the population exceeds a certain threshold of density. It is argued that Haldane’s model presuppose hard selection, and that it does not apply where soft selection predominates.

It is true that Haldane’s illustrative scenario assumes hard selection, but his calculations would apply to any circumstances where the pattern of fitness satisfies his assumptions. Frequency-dependent selection might in principle do so. The real point of the soft-selection argument is that, according to Wallace, in soft selection the fitness of individuals is decided by their place in a rank order of fitness. At any given time there is a fixed number N of ‘places’ available for the population, determined by ecological factors, so that the first N individuals in order of fitness (allowing for random mortality) will survive, and the rest will die. It is evident that this is another version of threshold selection, as in Maynard Smith’s model, and the same pros and cons apply.

3. It is sometimes suggested that Haldane’s arguments do not apply to new advantageous mutations. A new advantageous mutation does not cause any deaths except to the extent that its bearers gradually squeeze out non-bearers (assuming a fixed population size). The maximum cumulative number of deaths caused by this ‘squeezing out’ can be no more than the total population size, and not 30 times the population size, as in Haldane’s model.

I think this objection is largely misconceived. It is true that a new advantageous mutation cannot cause more individuals to die than the total population size, but it does require many more selective deaths than this.

To understand this paradoxical assertion, consider the following example. Suppose that in every generation 2% of a population dies from
spells of extreme cold, against which the species has no defence. These deaths will be counted as part of random, non-selective mortality. Suppose now that a new dominant mutation gives its bearers partial protection against cold spells, so that only 1% instead of 2% of the bearers die from this cause. The bearers will therefore have a 1% fitness advantage over the non-bearers, and will gradually increase in frequency. During this process a large number of selective deaths will be attributable to the new mutation – usually more than 30 times the population size, because a new mutation probably starts from a greater rarity in the population than in Haldane’s model. There are no more deaths from cold than before – in fact there are fewer – but deaths previously counted as non-selective are now selective.

So what? Does it matter if deaths are transferred from the ‘non-selective’ to the ‘selective’ category? If we are only considering a single gene, it probably does not matter, but if we are considering the overall scope for natural selection, it does. The problem is that there are only a certain number of deaths available to pay for natural selection. The number cannot exceed the total population size in any generation, and it will normally be much less than this, as there is bound to be a large amount of random mortality. The more mutations there are at any time, the fewer the deaths available to pay for each (assuming independent fitness). There is no free lunch: advantageous mutations still have to be paid for, and the ‘funds’ available, in terms of the scope for selective deaths, are limited. The problem is not that new advantageous mutations somehow make things worse – obviously not – but what they imply about the previous state of the species. If a large number of new advantageous mutations are under rapid selection, without any change in the environment, this suggests that the species was very badly adapted to its environment before. So how did it survive? One answer to this may be that the ‘environment’ of a species consists largely of other organisms – competitors, predators, prey, parasites, and so on – and they are all evolving together. At any given time a species may be adequately adopted to its organic environment, but if it does not keep evolving it will fall behind.

This may be true, but it does not seem particularly relevant to the case of strong selection of a large number of new mutations simultaneously. This would imply that the other species (competitors, etc) had somehow obtained a major advantage, so that the species under selection had to catch up. It isn’t clear how this situation would arise.

4. An initially attractive answer to Haldane’s Dilemma is that it only applies to selection due to differential mortality, and not to differential fertility. If some genotypes have more offspring than others, there is no need for selective deaths at all. The worst that can happen is that if there is a fixed population size, some individuals will be forced to have fewer or no offspring. But the cumulative total of individuals forced to be childless will not exceed the total population size. Selection by differential fertility is especially attractive when a gene is rare, because in these circumstances it is much less wasteful than selection by differential mortality. When an advantageous gene is rare, selection by mortality requires a fraction 1/x of the entire population to die just to raise the frequency of the small number of advantageous genes by roughly the same proportion (if selection is not very strong). In contrast, selection by fertility only requires the rare bearers of the advantageous gene each to have on average 1/x more offspring to achieve the same effect. This is much more efficient, from the point of view of the species a whole. Unfortunately, there is no reason to expect the evolutionary process to be efficient for the species as a whole.

Strictly speaking, Haldane’s model does allow for selection by differential fertility. Haldane several times refers to diminished fertility as being equivalent to selective mortality. There is however an important difference. In the case of selective mortality, a certain number of actual deaths must occur. In contrast, ‘diminished fertility’ is measured relative to the fertility of the optimum genotype, which may be entirely hypothetical. If several advantageous genes are still rare, there will probably be no actual individual who has all of them. Whereas real deaths in any generation cannot exceed the actual population size, there is no upper limit to a hypothetical loss of fertility. Each new advantageous allele increases the fertility of the optimum genotype, against which ‘diminished fertility’ is measured. The ‘cost’ of selection may be largely theoretical. It does therefore seem that if we allow selection to be largely by differential fertility, there is no real constraint on the speed of evolution.

I think there is something to this argument, but that we should not be too optimistic about the scope for selection by differential fertility. The reasons for differences in fertility (number of offspring) of individuals in a species may be divided broadly into two categories:

a) differences in the resources the individual can devote to reproduction. This is turn depends on its general condition, and especially on its health and nutritional status;

b) differences in the number of offspring produced from a given amount of resources.

Differences of the first kind depend on much the same factors as differences in mortality, such as the ability to find food, to avoid predators, and to resist disease. They are therefore not really an alternative to selection by mortality, and differences in fertility of this kind, in so far as they have a genetic basis, are likely to be accompanied by selective deaths on much the same scale as in Haldane’s model.

The second category of differences falls within the field of ‘life history strategy’. There is a trade-off between the number of offspring produced and their average survival rate. We expect there to be an optimum fertility level for each species, determined by its ecology and behaviour. Any mutations which cause an individual to produce more offspring than the current optimum will in general be selected against, except in the special circumstance that the mutation itself changes the optimum, for example by increasing the efficiency of parental care. To expect it to be easy to increase fertility is to expect another free lunch.

5. A more technical criticism is of Haldane’s measure of fitness. Haldane stipulates that the optimum genotype that can be formed from the available genes has fitness 1. But if some advantageous alleles are still rare, in a finite population the optimum genotype is unlikely to have any representatives, and the fitness of all existing individuals may appear very low by Haldane’s measure. It has therefore been argued that fitness should be measured by reference to the best genotype available in practice. It is further argued that with this approach, the cost of selection is much lower, and that several alleles can be fixed in each generation, even if all the other assumptions of Haldane’s model are adopted.

I have not followed all the details of this argument, but it has been used by some distinguished geneticists, such as Warren Ewens, so I am sure they have proved something. But I do not see how it can solve Haldane’s Dilemma. Haldane’s argument does not depend crucially on how fitness is measured. His key assumptions are:

a) selection is by means of differential mortality

b) selection at different loci is independent

c) selection at each locus is not very strong

d) there is a substantial amount of non-selective mortality.

From assumption (a) it seems inevitable that fixation of any single allele will require selective deaths cumulatively totalling several times the population
size, and assumptions (b) – (d) entail that the cost of fixing more than one allele will be approximately the sum of the costs of fixing each. Setting a different benchmark for fitness would not help. I suspect that any findings to the contrary involve some de facto relaxation of the ‘independence’ assumption.

Overall, it seems to me that Haldane’s Dilemma has more force than is sometimes assumed, and that it, or rather the issues it raises, should be taken seriously. Whenever it is suggested that a large number of genes have been under selection simultaneously, we need to look at what this implies for the intensity and mode of selection. This applies also when the selection of entities other than genes, such as social groups or cultural traits, is under consideration. For example, if it is suggested that cultural traits have evolved by the differential survival of societies possessing those traits, we need to take a hard look at what this implies for the number of group ‘extinctions’ and the number of cultural traits that might feasibly evolve in the way suggested.

At the same time, the Dilemma does not set a rigid limit to the speed of evolution. If we give up the assumption of independent fitness, evolution can go at any speed we like. Selection by means of differential fertility may also offer a way out of the Dilemma, though I have argued above that it is not a panacea.

It remains a matter for empirical research whether rapid multi-locus selection is a common evolutionary phenomenon, and if so what is the selective mechanism that permits it. This also has implications for other problems in evolutionary biology. For example, some theories of the advantages of sexual reproduction (e.g. Kondrashov’s) require patterns of synergistic fitness effects of a kind that might also solve Haldane’s Dilemma. The Dilemma also seems relevant to the theory of punctuated equilibrium, where it is claimed that most evolutionary change occurs in relatively short bursts in between long periods of stasis. Yet Haldane’s 1957 paper is not mentioned in S. J. Gould’s vast book on The Structure of Evolutionary Theory. Maybe the time is right for some joined-up thinking?

Note 1: some commentators have misinterpreted Haldane as arguing that only one gene can be under selection at a time, at intervals of 300 generations! (see for example this Wikipedia entry). This is complete nonsense, and a travesty of Haldane’s paper.

Note 2: not all biologists dismissed the problem. For example, John Maynard Smith, in his 1978 book The Evolution of Sex, gave it serious attention.

Note 3: on page 520 there is a formula which may be written as n = 30 Sd, where n is the number of generations required for a gene substitution (my K), S is a summation sign (Haldane’s large sigma), and d is the average fitness effect of each gene under selection (Haldane’s small delta). I think this formula must be a misprint for n = 30/Sd. The formula as printed makes no sense, as it implies that substitution takes longer when selection is more intense! The correction n = 30/Sd is consistent with the rest of Haldane’s working, and also with the exposition by John Maynard Smith, who gives a formula equivalent to Sd = 30/n.

Note 4: the assumption is not explicit in the 1957 paper, but it seems to be required for Haldane’s calculations, and in a paper of 1960, describing his 1957 conclusions, Haldane referred explicitly to ‘a population of constant size’. This requirement can be satisfied if we assume that there is always a surplus of offspring sufficient to cover ‘selective’ mortality, with the surplus being reduced by random mortality to a constant carrying capacity. Some commentators have seen the assumption of constant population size as an inconsistency or major flaw in Haldane’s analysis, but R. Flake and V. Grant showed in 1974 that Haldane’s main conclusions still hold good for varying population size and varying intensity of selection.

Note 5: If selection on each allele is strong, overlap of selection for different alleles may become significant, and reduce the overall number of selective deaths required. To take an extreme example, if all individuals who lack any of the rare advantageous alleles die, then the only survivors will be those who bear all of those alleles, and in a single generation they will all be fixed, at a cost of slightly less than 1 times the population size. The snag is that there probably will not be any individuals with all those alleles, so the species will go extinct! To take a less extreme example, if there are two rare advantageous alleles, A and B, each with a 50% selective advantage, then in each generation 50% of non-As and 50% of non-Bs will be killed, but around 25% of the population will be ‘killed’ by both factors of selection, so the total selective mortality will be around 75%, not 100%. Even if all the s’s are individually small, overlap will become significant if there are enough alleles under selection, but this will not occur unless the sum of the s’s is relatively large, which again implies likely extinction.

References:

J. B. S. Haldane: The cost of natural selection, Journal of Genetics, 55, 1957, 511-24.
J. Maynard Smith: ‘Haldane’s Dilemma’ and the rate of evolution, Nature, 219, 1968, 1114-6.
P. O.Donald: ‘Haldane’s Dilemma’ and the rate of natural selection, Nature, 221, 1969, 815-6.
B. Wallace: Genetic Load, 1970.

Posted in Uncategorized

I was wrong about skin color in Europeans and East Asians

In December I posted a hypothesis about skin color evolution in East Asians and Europeans in light of new research which suggested that Europeans were fixed for a particular locus which explained a great deal of the variation between themselves and Africans. The facts are like so:

1) There are probably around 5 loci of major effect that explain skin color variation in human beings (older pedigree analysis as well as new genomic data seems to be focusing on a number in this range).

2) One of the major loci, MC1R, is highly polymorphic in Europeans, constrained in dark-skinned populations to the ancestral consensus sequence (Africans, to a lesser extent in South Asians), and being driven toward fixation on a derived allele in East Asians.

3) A new locus in Europeans seems fixed and is responsible for the skin color depigmentation, but in both East Asians and Africans this locus is in an ancestral condition.

From these data I hypothesized that East Asians and Europeans were genetically isolated so they reached the same fitness optimums (assuming light skin has selective benefits at high latitudes) via different genetic architectures. In Europeans MC1R is diversified (either through frequency dependent selection or relaxation) because SLC24A5 handles the phenotype. In East Asians SLC24A5 is in an ancestral state, so MC1R has been coopted via the derived allele to generate the same end via different means.

I think I was wrong, and the evil Finns are to blame…they exhibit a rather high frequency of the MC1R allele which confers light skin in East Asians, but they’re in Europe! The idea that I threw out there is really contingent on powerful genetic barriers, beneficial alleles have a way of sweeping everywhere that they confer fitness all things being equal. Lactose tolerance is spread throughout Eurasia, and not because of massive population movements, while various malarial resistance alleles are pretty common across regions just where malaria is endemic. Perhaps there was a massive ice sheet in the middle of Eurasia, and Finns are the byproduct of an admixture event, but my understanding is that MC1R has been under positive selection for around ~10,000 years in East Asians. That’s pretty recent, and I don’t think that the genetic barriers across Eurasia were that strong for that period, they should been able to “borrow” SLC24A5 instead of throwing up their own trick de novo. Other shit is going down…and I’m sure that pleiotropy has its dirty hand involved here.

But I was wrong, though I did give it a good college try 🙂

Stupidity about religious people

Recently I’ve been saying that it is important to distinguish between what people believe, what they say they believe and what they do. The three do not always integrate well together. One can posit many reasons for this, though I believe some form of cognitive modularity and various social psychological pressures can probably account for much of it. It seems a trivial and obvious point, but in discussions about human motives and predictions about their behavior these issues tend to be ignored. The problem is that people have a hard time treating other human beings as natural objects. There’s a reason for this, we’re damn complicated and hard to figure out. Additionally, our intuitive psychology gives us a lot of “free information” concealed within our heads which we take for granted, we don’t have to make recourse to explicit models because we know much of what needs to be known implicitly. We trust introspection because it works well in day to day interactions, we come preloaded with a lot of people interpretation software.

I bring this up because I was remembering a stupid thing I used to say about religious people: if the religious believe in an afterlife why do they get so scared when people are about to kill them? This was a dumb “gotcha” for showing that religious people weren’t really that religious, that they didn’t really believe that there was an afterlife because they didn’t behave as if there was an afterlife when this life was going to get snuffed out. I’ve seen this come up on the message boards here too, I’m not the only person who has naively expressed this truth as if it’s clever.

The problem is that we assume that people are cognitively integrated, and that there is a tight fidelity between their various beliefs and how they inform one’s actions. And, reflectively, if one accepts that evolution has shaped our behavorial suite to some extent, reflexive fear of getting killed would seem to be one of the most adaptive handy mental biases! I was always a big fan of evolutionary psychology, but I was too much a moron to consider this. That is, even if someone sincerely believed that there was a life after, their “instinct” for fear would simply overwhelm their reflective beliefs in the “heat of the moment.” Obviously if you didn’t have this “flight or fear” sort of instinct, not matter your prior beliefs about the afterlife (most peoples seem to have some expectation of post-mortem sentience, though it usually isn’t as cheery as modern religions make it out to be), your long term fitness will decrease. Ultimately you could probably think it through and come to the conclusion that it is best to run and hide, but by that time you might be dead.

Addendum: You can invert this with the “there are no atheists in a foxhole” saying, just because someone makes recourse to divine aid in a moment of duress doesn’t mean they rationally believe said being exists. Many unbelievers I know have a strong bias toward wishing there was a divine being, they simply can’t reflectively convince themselves that its existence is plausible. It makes sense that native cognitive biases can shunt aside reflection in situations where cost of philosophical inconsistency seems trivial. This sort of in-the-moment apostacy is well known in in religious circles, the Donatist movement in ancient Rome arose because so many Christians, including bishops and church leaders, had apostasized under threat of persecution and death. The orthodox position was that apostates had to be forgiven because of considerations of human nature, while the Donatists rejected such compromises. The reality is that the apostates were almost certainly sincere Christians, they often reverted back the first chance they got, but they also had a strong instinctive fear of death, so they compromised their principles.

Posted in Uncategorized

Living Neandertals & another idea about the Hobbits

Just a note for those of you who don’t read my other weblog, Greg Cochran has two new ideas that we’re trying to guess at. First, he’s got a new theory about the origin/evolution about the Hobbits of Flores. Second, he thinks it might be a possibility that there are living Neandertals. Anyway, I’m throwing this out there because I won’t be too surprised if Greg can convince Nick Wade to write a story in the Grey Lady about this stuff at some point in the near future….

In Search of Good Metaphors

When trying to get across the gist of complex technical subjects to a layperson, sometimes a good metaphor can do more pound-for-pound explanatory work than hurling large amounts of jargon-laden details.

Fore example, Gary Marcus provides a lucid explanation of gene expression by analogizing each gene to a conditional statement in a software program. Each gene has a set of conditions under which it’s expressed (IF/WHILE), and a specific protein it generates when activated (THEN/DO). While this is most readily intelligible to people with experience in computer progamming, it can be understood by intelligent non-programmers with a little coaxing and can clear up several confusions and help order one’s thoughts:

  • So-called “junk” DNA is just dead code — code whose execution conditions are never met. Piles of this stuff can accumulate during the development of sufficiently large and complex software programs, and that’s with intelligent programmers watching over it. Small wonder if orders of magnitude more happen to accumulate under the guidance of a completely blind process.
  • The oft-quoted trope that we’re 98% similar to chimpanzees at the genetic level loses significance; even a 2% change in the source code of a program, particularly on the conditionals, can have massive effects on how the program executes.
  • The sillyness of opposing genes versus environment or “innateness” versus “plasticity” becomes evident — these things are orthogonal to eachother rather than in conflict. A program can be specified to execute multiple different ways given different inputs, and even if the brain starts with a highly specified innate structure, it could be programmed to rewire itself based on environmental inputs.

And so forth. Another good example for the numerate is the definition of race that got Steve Hsu’s comments deleted by Brad DeLong: represent each individual’s genome as a point in a space of extremely high dimension, and define a race as a set of points whose distance from each other is less than some radius. These clusters map onto intuitive self-identified race with a very high degree of accuracy.

A third example (which I first encountered from Henry Harpending [PDF]) is analogizing intelligence to size. Many opponents of IQ testing start from the correct point that there is no single definition, measurement or task that completely captures everything we consider “intelligence,” but then slide from there into saying that IQ is meaningless and intelligence can’t be measured.

If you substitute “size” for “intelligence” it becomes apparent just how silly the argument is. Height is one dimension of a person’s size, as is shoulder breadth, weight, the length and circumference of one’s limbs and so forth. There are short guys who are built like tanks and tall guys who’d blow away in a strong wind, but nobody takes exception to the concept of “size” on their account or argues that size can’t be measured.

This cuts both ways: it also suggests that attempting to reify g is a category mistake. Like the concept of a center of gravity, g is an abstractum, a theorist’s fiction — but one that is well-behaved and has a causal “reality” to it all the same. We can predict that someone with a high g will perform well on all kinds of cognitively challenging tasks just as we can predict that a chair tipped on its back legs past a certain angle will fall over.

Since GNXPers are used to dealing in complex subjects which are very misunderstood and commonly regarded with perplexity (if not suspicion or outright hostility) by laymen, I thought it might be worthwhile to solicit whatever other metaphors y’all have found useful in both understanding and explaining technical concepts. They don’t have to be biology-related, though I expect most will be. Discuss amongst yourselves.

Update: John Wilkins responds thoughtfully to the brief treatment of race here, but I think he and PZ Myers, both professed Lewontinites, ultimately miss the boat. John gives it away thusly: “So, do I think there are races in biology as well as culture? No.”

Asking whether race exists in this way is a category mistake, albeit an all too common one. Race is another abstractum, like the general intelligence factor. How we define it will be wholly a matter of convention, though not totally arbitrary because some definitions obviously have more utility than others. Returning to the initial representation I used, you could shrink the race-radius right down to the point where each individual (plus his twin, if he had one) was his own distinct “race” if you wanted to, but this wouldn’t be interesting. If all that one means by “race is a social construct” is that one can can twiddle the granularity of racial categories virtually however one likes, then this is perfectly true and perfectly beside the point.

Because at bottom, all this abstraction and definition is based on a real molecular substrate. Piles of rock and dirt don’t change their height based on whether one choses to call them hills or mountains. Hypertension doesn’t suddenly stop being more frequent in African men based on how one decides to classify them. Genetic variation between populations doesn’t become any less real. Eppur si muove.

Update the 2nd: GMTA. Razib goes on at greater length and depth, as usual.

Posted in Uncategorized