NotPoliticallyCorrect

Home » Refutations

Category Archives: Refutations

The Myth of “General Intelligence”

5000 words

Introduction

“General Intelligence” or g is championed as the hallmark “discovery” of psychology. First “discovered” by Charles Spearman in 1904, noting that schoolchildren who scored highly on one test scored highly on others and vice versa for lower-scoring children, he assumed that due to the correlation between tests, that there must be an underlying physiological basis to the correlation, which he posited to be some kind of “mental energy”, stating that the central nervous system (CNS) explained the correlation. He proclaimed that g really existed and that he had verified Galton’s claim of a unitary general ability (Richardson, 2017: 82-83). Psychometricians then claim, from these intercorrelations of scores, that what is hidden from us is then revealed, and that the correlations show that something exists and is driving the correlation in question. That’s the goal of psychometrics/psychology—to quantify and then measure psychological traits/mental abilities. However, I have argued at length that it is a conceptual impossibility—the goal of psychometrics is an impossibility since psychometrics isn’t measurement. Therefore, claims that IQ tests measure g is false.

First, I will discuss the reification of g and it’s relation to brain properties. I will argue that if g is a thing then it must have a biological basis, that is it must be a brain property. Reductionists like Jensen have said as much. But it’s due to the reification of g as a concrete, physical thing that has people hold such beliefs. Second, I will discuss Geary’s theory that g is identical with mitochondrial functioning. I will describe what mitochondria does, and what powers it, and then discuss the theory. I will have a negative view of it, due to the fact that he is attempting to co-opt real, actual functions of a bodily process and attempt to weave g theory into it. Third, I will discuss whether or not psychological traits are indeed quantifiable and measurable, and whether or not there is a definition psychometricians can use to ground their empirical investigations. I will argue negatively for all three. Fourth, I will discuss Herrnstein and Murray’s 6 claims in The Bell Curve about IQ and provide a response to each in turn. Fifth, I will discuss the real cause of score variation, which isn’t reduction to a so-called assumed existence of a biological process/mechanism, but which is due to affective factors and exposure to the specific type of knowledge items on the test. Lastly, I will conclude and give an argument for why g isn’t a thing and is therefore immeasurable.

On reifications and brain properties

Contrary to protestations from psychometricians, they in fact do reifiy correlations and then claim that there exists some unitary, general factor that pervades all mental tests. If reification is treating the abstract as something physical, and if psychometrics treat g as something physical, then they are reifying g based on mere intercorrelations between tests. I am aware that, try as they might, they do attempt to show that there is an underlying biology to g, but these claims are defeated by the myriad arguments I’ve raised against the reducibility of the mental to the physical. Another thing that Gould gets at is that psychometricians claim that they can rank people—this is where the psychometric assumption that because we can put a number to their reified thing, that there is something being measured.

Reification is “the propensity to convert an abstract concept (like intelligence) into a hard entity (like an amount of quantifiable brain stuff)” (Gould, 1996: 27). So g theorists treat g as a concrete, physical, thing, which then guides their empirical investigations. They basically posit that the mental has a material basis, and they claim that they can, by using correlations between different test batteries, we can elucidate the causal biological mechanisms/brain properties responsible for the correlation.

Spearman’s theory—and IQ—is a faculty theory (Nash, 1990). It is a theory in which it is claimed that the mind is separated into different faculties, where mental entities cause the intellectual performance. Such a theory needs to keep up the claim that a cognitive faculty is causally efficacious for information processing. But the claim that the mind is “separated” into different faculties fails, and it fails since the mind is a single sphere of consciousness, it is not a complicated arrangement of mental parts. Physicalists like Jensen and Spearman don’t even have a sound philosophical basis on which to ground their theories. Their psychology is inherently materialist/physicalist, but materialism/physicalism is false and so it follows that their claims do not hold any water. The fact of the matter is, Spearman saw what he wanted to see in his data (Schlinger, 2003).

I have already proven that since dualism is true, then the mental is irreducible to the physical and since psychometrics isn’t measurement, then what psychometricians claim to do just isn’t possible. I have further argued that science can’t study first-personal subjective states since science is third-personal and objective. The fact is the matter is, hereditarian psychologists are physicalist, but it is impossible for a purely physical thing to be able to think. Claims from psychometricians about their “mental tests” basically reduce to one singular claim: that g is a brain property. I have been saying this for years—if g exists, it has to be a brain property. But for it to be a brain property, one needs to provide defeaters for my arguments against the irreducibility of the mental and they also need to argue against the arguments that psychometrics isn’t measurement and that psychology isn’t quantifiable. They can assume all they want that it is quantifiable and that since they are giving tests, questionnaires, likert scales, and other kinds of “assessments” to people that they are really measuring something; but, ultimately, if they are actually measuring something, then that thing has to be physical.

Jensen (1999) made a suite of claims trying to argue for a physical basis for g,—to reduce g to biology—though, upon conceptual examination (which I have provided above) these claims outright fail:

g…[is] a biological [property], a property of the brain

The ultimate arbiter among various “theories of intelligence” must be the physical properties of the brain itself. The current frontier of g research is the investigation of the anatomical and physiological features of the brain that cause g.

…psychometric g has many physical correlates…[and it] is a biological phenomenon.

As can be seen, Jensen is quite obviously claiming that g is a biological brain property—and this is what I’ve been saying to IQ-ists for years: If g exists, then it MUST be a property of the brain. That is, it MUST have a physical basis. But for g proponents to show this is in fact reality, they need to attempt to discredit the arguments for dualism, that is, they need to show that the mental is reducible to the physical. Jensen is quite obviously saying that a form of mind-brain identity is true, and so my claim that it was inevitable for hereditarianism to become a form of mind-brain identity theory is quite obviously true. The fact of the matter is, Jensen’s beliefs are reliant upon an outmoded concept of the gene, and indeed even a biologically implausible heritability (Richardson, 1999; Burt and Simons, 2014, 2015).

But Jensen (1969) contradicted himself when it comes to g. On page 9, he writes that “We should not reify g as an entity, of course, since it is only a hypothetical construct intended to explain covariation am ong tests. It is a hypothetical source of variance (individual differences) in test scores.” But then 10 pages later on pages 19-20 he completely contradicts himself, writing that g is “a biological reality and not just a figment of social conventions.” That’s quite the contradiction: “Don’t reifiy X, but X is real.” Jensen then spent the rest of his career trying to reduce g to biology/the brain (brain properties), as we see above.

But we are now in the year 2023, and so of course there are new theoretical developments which attempt to show that Spearman’s hypothesized mental energy really does exist, and that it is the cause of variations in scores and of the positive manifold. This is now where we will turn.

g and mitochondrial functioning

In a series of papers, David Geary (2018, 2019, 2020, 2021) tries to argue that mitochondriaal functioning is the core component in g. At last, Spearman’s hypothetical construct has been found in the biology of our cells—or has it?

One of the main functions of mitochondria is to oxidative phosphorylation to produce adenosine triphosphate (ATP). All living cells use ATP as fuel, it acts as a signaling molecule, it is also involved in cellular differentiation and cell death (Khakh and Burnstock, 2009). The role of mitochondrial functioning in spurring disease states has been known for a while, such as with cardiovascular diseases such as cardiomyopathy (Murphy et al, 2016, Ramaccini at al, 2021).

So due to the positive manifold, where performance in one thing is correlated with a performance in another, Geary assumes—as Spearman and Jensen did before him—that there must be some underlying biological mechanism which then explains the correlation. Geary then uses established outcomes of irregular mitochondrial functioning to then argue that the mental energy that Spearman was looking for could be found in mitochondrial functioning. Basically, this mental energy is ATP. I don’t deny that mitochondriaal functioning plays a role in the acquisition of disease states, indeed this has been well known (eg, Gonzales et al, 2022). What I deny is Gary’s claim that mitochondrial functioning has identity with Spearman’s g.

His theory is, like all other hereditarian-type theories, merely correlative—just like g theory. He hasn’t shown any direct, causal, evidence of mitochondrial functioning in “intelligence” differences (nor for a given “chronological age). That as people age their bodies change which then has an effect on their functioning doesn’t mean that the powerhouse of the cell—ATP—is causing said individual differences and the intercorrelations between tests (Sternberg, 2020). Indeed, environmental pollutants affect mitochondrial functioning (Byun and Baccarelli, 2014; Lambertini and Byun, 2016). Indeed, most—if not all—of Geary’s hypotheses do not pass empirical investigation (Schubert and Hagemann, 2020). So while Geary’s theory is interesting and certainly novel, it fails in explaining what he set out to.

Quantifiable, measurable, definable, g?

The way that g is conceptualized is that there is a quantity of it—where one has “more of it” than other people, and this, then, explains how “intelligent” they are in comparison to others—so implicit in so-called psychometric theory is that whatever it is their tests are tests of, something is being quantified. But what does it mean to quantify something? Basically, what is quantification? Simply, it’s the act of giving a numerical value to a thing that is measured. Now we have come to an impasse—if it isn’t possible to measure what is immaterial, how can we quantify it? That’s the thing, we can’t. The g approach is inherently a biologically reductionist one. Biological reductionism is false. So the g approach is false.

Both Gottfredson (1998) and Plomin (1999) make similar claims to Jensen, where they talks about the “biology of g” and the “genetics of g“. Plomin (1999) claims that studies of twins show that g has a substantial heritability, while Gottfredson (1998) claims that heritability of IQ increases to up until adulthood where it “rises to 60 percent in adolescence and to 80 percent by late adulthood“, citing Bouchard’s MISTRA (Minnesota Study of Twins Reared Apart). (See Joseph 2022 for critique and for the claim that the heritability of IQ in that study is 0 percent.) They, being IQ-ists, of course assume a genetic component to this mystical g. However, there arguments are based on numerous false assumptions and studies with bad designs (and hidden results), and so they must be rejected.

If X is quantitative, then X is measurable. If X is measurable, then X has a physical basis. Psychological traits don’t have a physical basis. So psychological traits aren’t quantitative and therefore not measurable. Geary’s attempt at arguing for identity between g and mitochondrial functioning is an attempt at a specified measured object for g, though his theory just doesn’t hold. Stating truisms about a biological process and then attempting to liken the process with the construct g just doesn’t work; it’s just a post-hoc rationalization to attempt to liken g with an actual biological process.

Furthermore, if X is quantitative, then there is a specified measured object, object of measurement and measurement unit for X. But this is where things get rocky for g theorists and psychometricians. Psychometry is merely pseudo-measurement. Psychometricians cannot give a specified measured object, and if they can’t give a specified measured object they cannot give an object of measurement. They thusly also cannot construct a measurement unit. Therfore, “the necessary conditions for metrication do not exist” (Nash, 1990: 141). Even Haier (2014, 2018) admits that IQ test scores don’t have a unit that is like inches, liters, or grams. This is because those are ratio scales and IQ is ordinal. That is, there is no “0-point” for IQ, like there is for other actual, real measures like temperature. That’s the thing—if you have a thing to be measured, then you have a physical object and consequently a measument unit. But this is just not possible for psychometry. I then wonder why Haier doesn’t follow what he wrote to its logical conclusion—that the project of psychometrics is just not possible. Of course the concept of intelligence doesn’t have a referent, that is, it doesn’t name a property like height, weight, or temperature (Midgley, 2018:100-101). Even the most-cited definition of intelligence—Gottfredson’s—still fails, since she contradicts herself in her very definition.

Of course IQ “ranks” people by their performance—some people perform better on the test than others (which is an outcome of prior experience). So g theorists and IQ-ists assume that the IQ test is measuring some property that varies between groups which then leads to score differences on their psychometric tests. But as Roy Nash (1990: 134) wrote:

It is impossible to provide a satisfactory, that is non-circular, definition of the supposed ‘general cognitive ability’ IQ tests attempt to measure and without that definition IQ theory fails to meet the minimal conditions of measurement.

But Boeck and Kovas (2020) try to sidestep this issue with an extraordinary claim, “Perhaps we do not need a definition of intelligence to investigate intelligence.” How can we investigate something sans a definition of the object of investigation? How can we claim that a thing is measured if we have no definition, and no specified measured object, object of measurement and measurement unit, as IQ-ists seem to agree with? Again, IQ-ists don’t take these conclusions to their further logical conclusion—that we simply just cannot measure and quantify psychological traits.

Haier claims that PGS and “DNA profiles” may lead to “new definitions of intelligence” (however ridiculous a claim). He also, in 2009, had a negative outlook on identifying a “neuro g” since “g-scores derived from different test batteries do not necessarily have equivalent neuro-anatomical substrates, suggesting that identifying a “neuro-g” will be difficult” (Haier, 2009). But one more important reason exists, and it won’t just make it “difficult” to identify a neuro g, it makes it conceptually impossible. That is the fact that cognitive localizations are not possible, and that we reify a kind of average in brain activations when we look at brain scans using fMRI. The fact of the matter is, neuroreduction just isn’t possible, empirically (Uttal, 2001, 2014, 2012), nor is it possible conceptually.

Herrnstein and Murray’s 6 claims

Herrnstein and Murray (1994) make six claims about IQ (and also g):

(1) There is such a thing as a general factor of cognitive on which human beings differ.

Of course implicit in this claim is that it’s a brain property, and that people have this in different quantities. However, the discussion above puts this claim to bed since psychological traits aren’t quantitative. This, of course comes from the intercorrelations of test scores. But we will see that most of the source of variation isn’t even entirely cognitive and is largely affective and due to one’s life experiences (due to the nature of the item content).

(2) All standardized tests of academic aptitude or achievement measure this general factor to some degree, but IQ tests expressly designed for that purpose measure it most accurately.

Of course Herrnstein and Murray are married to the idea that these tests are measures of something, that since they give different numbers due to one’s performance, there must be an underlying biology behind the differences. But of course, psychometry isn’t true measurement.

(3) IQ scores match, to a first degree, whatever it is that people mean when they use the word intelligent or smart in ordinary language.

That’s because the tests are constructed to agree with prior assumptions on who is or is not “intelligent.” As Terman constructed his Stanford-Binet to agree with his own preconceived notions of who is or is not “intelligent”: “By developing  an exclusion-inclusion criteria that favored the  aforementioned groups, test developers created a norm “intelligent” (Gersh, 1987, p.166) population “to differentiate subjects of known superiority from subjects of known inferiority” (Terman, 1922, p. 656)” (Bazemoore-James, Shinaprayoon, and Martin 2017). Of course, since newer tests are “validated”(that is, correlated with) older, tests (Richardson, 199120002002, 2017Howe, 1997), this assumption is still alive today.

(4) IQ scores are stable, although not perfectly so, over much of a person’s life.

IQ test scores are malleable, and this of course would be due to the experience one has in their lives which would then have them ready to take a test. Even so, if this claim were true, it wouldn’t speak to the “biology” of g.

(5) Properly administered IQ tests are not demonstrably biased against social, economic, ethnic, or racial groups.

This claim is outright false and can be known quite simply: the items on IQ tests derive from specific classes, mainly the white middle-class. Since this is true, it would then follow that people who are not exposed to the item content and test structures wouldn’t be as prepared as those who are. Thus, IQ tests are biased against different groups, and if they are biased against different groups it also follows that they are biased for certain groups, mainly white Americans. (See here for considerations on Asians.)

(6) Cognitive ability is substantially heritable, apparently no less than 40 percent and no more than 80 percent.

It’s nonsense to claim that one can apportion heritability into genetic and environmental causes, due to the interaction between the two. IQ-ists may claim that twin, family, and adoption studies show that IQ is X amount heritable so there must thusly be a genetic component to differences in test scores. But the issue with heritability has been noted for decades (see Charney, 2012, 2016, 2022; Joseph, 2014, Moore and Shenk, 2016, Richardson, 2017) so this claim also fails. There is also the fact that behavioral genetics doesn’t have any “laws.” It’s simply fallacious to believe that nature and nurture, genes and environment, contribute additively to the phenotype, and that their relative contributions to the phenotype can be apportioned. But hereditarians need to keep that facade up, since it’s the only way their ideas can have a chance at working.

What explains the intercorrelations?

We still need an explanation of the intercorrelations between test scores. I have exhaustively argued that the usual explanations from hereditarianism outright fail—g isn’t a biological reality and IQ tests aren’t a measure at all because psychometrics isn’t measurement. So what explains the intercorrelations? We know that IQ tests are comprised of different items, whether knowledge items or more “abstract” items like the Raven. Therefore, we need to look to the fact that people aren’t exposed to certain things, and so if one comes across something novel that they’ve never been exposed to, they thusly won’t know how to answer it and their score will then be affected due to their ignorance of the relationship between the question and answer on the test. But there are other things irrespective of the relationship between one’s social class and the knowledge they’re exposed to, but social class would still then have an effect on the outcome.

IQ is, merely, numerical surrogates for class affiliation (Richardson, 1999; 2002; 2022). The fact of the matter is, all human cognizing takes place in specific cultural contexts in which cultural and psychological tools are used. This means, quite simply, that culture-fair tests are impossible and, therefore, that such tests are necessarily biased against certain groups, and so they are biased for certain groups. Lev Vygotsky’s sociocultural theory of cognitive development and his concepts of psychological and cultural tools is apt here. This is wonderfully noted by Richardson (2002: 288):

IQ tests, the items of which are designed by members of a rather narrow social class, will tend to test for the acquisition of a rather particular set of cultural tools: in effect, to test, or screen, for individuals’ psychological proximity to that set per se, regardless of intellectual complexity or superiority as such.

Thinking is culturally embedded and contextually-specific (although irreducible to physical things), mediated by specific cultural tools (Richardson, 2002). This is because one is immersed in culture immediately from birth. But what is a cultural tool? Cultural tools include language (Weitzman, 2013) (it’s also a psychological tool), along with “different kinds of numbering and counting, writing schemes, mnemonic technical aids, algebraic symbol systems, art works, diagrams, maps, drawings, and all sorts of signs (John-Steiner & Mahn, 1996; Stetsenko, 1999)” (Robbins, 2005). Children are born into cultural environments, and also linguistically-mediated environments (Vasileva and Balyasnikova, 2019). But what are psychological tools? One psychological tool (which would also of course be cultural tools) would be words and symbols (Vallotton and Ayoub, 2012).

Vygotsky wrote: “In human behavior, we can observe a number of artificial means aimed at mastering one’s own psychological processes. These means can be conditionally called psychological tools or instruments… Psychological tools are artificial and intrinsically social, rather than natural and individual. They are aimed at controlling human behavior, no matter someone else’s or one’s own, just as technologies are aimed at controlling nature” (Vygotsky, 1982, vol. 1, p. 103, my translation). (Falikman, 2021).

The source of variation in IQ tests, after having argued that social class is a compound of the cultural tools one is exposed to. Furthermore, it has been shown that the language and numerical skills used on IQ tests are class-dependent (Brito, 2017). Thus, the compounded cultural tools of different classes and racial groups then coalesce to explain how and why they score the way they do. Richardson (2002: 287-288) writes

that the basic source of variation in IQ test scores is not entirely (or even mainly) cognitive, and what is cognitive is not general or unitary. It arises from a nexus of sociocognitive-affective factors determining individuals’ relative preparedness for the demands of the IQ test. These factors include (a) the extent to which people of different social classes and cultures have acquired a specific form of intelligence (or forms of knowledge and reasoning); (b) related variation in ‘academic orientation’ and ‘self-efficacy beliefs’; and (c) related variation in test anxiety, self-confidence, and so on, which affect performance in testing situations irrespective of actual ability.

Basically, what explains the intercorrelations of test scores—so-called g—are affective, non-cognitive factors (Richardson and Norgate, 2015). Being prepared for the tests, being exposed to the items on the tests (from which are drawn from the white middle-class) explains IQ score differences, not a mystical g that some have more of than others. That is, what explains IQ score variation is one’s “distance” from the middle-class—this follows due to the item content on the test. At the end of the day, IQ tests don’t measure the ability for complex cognition. (Richardson and Norgate, 2014). So one can see that differing acquisition of cultural tools by different cultures and classes would then explain how and why individuals of those groups then attain different knowledge. This, then, would license the claim that one’s IQ score is a mere outcome of their proximity to the certain cultural tools in use in the tests in question (Richardson, 2012).

The fact of the matter is, children do not enter school with the same degree of readiness (Richardson, 2022), and this is due to their social class and the types of things they are exposed to in virtue of their class membership (Richardson and Jones, 2019). Therefore, the explanation for these differences in scores need not be some kind of energy that people have in different quantities, it’s only the fact that from birth we are exposed to different cultures and therefore different cultural and psychological tools which then causes differences in the readiness of children for school. We don’t need to posit any supposed biological mechanism for that, when the answer is clear as day.

Conclusion

As can be seen from this discussion, it is clear that IQ-ist claims of g as a biological brain property fail. They fail because psychometrics isn’t measurement. They fail because psychometricians assume that what they are “measuring” (supposedly psychological traits) have a physical basis and have the necessary components for metrication. They fail because the proposed biology to back up g theory don’t work, and claiming identity between g and a biological process doesn’t mean that g has identity between that biological process. Merely describing facts about physiology and then attempting to liken it to g doesn’t work.

Psychologists try so very hard for psychology to be a respected science, even when what they are studying bares absolutely no relationship to the objects of scientific study. Their constructs are claimed to be natural kinds, but they are merely historically contingent. Due to the way these tests are constructed, is it any wonder why such score differences arise?

The so-called g factor is also an outcome of the way tests are constructed:

Subtests within a battery of intelligence tests are included n the basis of them showing a substantial correlation with the test as a whole, and tests which do not show such correlations are excluded. (Tyson, Jones, and Elcock, 2011: 67)

This is why there is a correlation between all subtests that comprise a test. Because it is an artificial creation of the test constructors, just like their normal curve. Of course if you pick and choose what you want in your battery or test, you can then coax it to get the results you want and then proclaim that what explains the correlations are some sort of unobserved, hidden variable that individuals have different quantities of. But the assumption that there is a quantity of course assumes that there is a physical basis to that thing. Physicalists like Jensen, Spearman, and then Haier of course presume that intelligence has a physical basis and is either driven by genes or can be reduced to neurophysiology. These claims don’t pass empirical and conceptual analysis. For these reasons and more, we should reject claims from hereditarian psychologists when they claim that they have discovered a genetic or neurophysiological underpinning to “intelligence.”

At the end of the day, the goal of psychometrics is clearly impossible. Try as they might, psychometricians will always fail. Their “science” will never be on the level of physics or chemistry, and that’s because they have no definition of intelligence, nor a specified measured object, object of measurement and measurement unit. They know this, and they attempt to construct arguments to argue their way out of the logical conclusions of those facts, but it just doesn’t work. “General intelligence” doesn’t exist. It’s a mere creation of psychologists and how they make their tests, so it’s basically just like the bell curve. Intelligence as an essence or quality is a myth; just because we have a noun “intelligence” doesn’t mean that there really exists a thing called “intelligence” (Schlinger, 2003). The fact is the matter is, intelligence is simply not an explanatory concept (Howe, 1997).

IQ-ist ideas have been subject to an all-out conceptual and empirical assault for decades. The model of the gene they use is false, (DNA sequences have no privileged causal role in development), heritability estimates can’t do what they need them to do, how the estimates are derived rest on highly environmentally-confounded studies, the so-called “laws” of behavioral genetics are anything but, they lack definitions and specified measured objects, objects of measurement and measurement units. It is quite simply clear that hereditarian ideas are not only empirically false, but they are conceptually false too. They don’t even have their concepts in order nor have they articulated exactly WHAT it is they are doing, and it clearly shows. The reification of what they claim to be measuring is paramount to that claim.

This is yet another arrow in the quiver of the anti-hereditarian—their supposed mental energy, their brain property, simply does not, nor can it, exist. And if it doesn’t exist, then they aren’t measuring what they think they’re measuring. If they’re not measuring what they think they’re measuring, then they’re showing relationships between score outcomes and something else, which would be social class membership along with everything else that is related with social class, like exposure to the test items, along with other affective variables.

Now here is the argument (hypothetical syllogism):

P1: If g doesn’t exist, then psychometricians are showing other sources of variation for differences in test scores.

P2: If psychometricians are showing other sources of variation for differences in test scores and we know that the items on the tests are class-dependent, then IQ score differences are mere surrogates for social class.

C: Therefore, if g doesn’t exist, then IQ score differences are mere surrogates for social class.

Advertisement

On the So-Called “Laws of Behavioral Genetics”

2400 words

In the year 2000, psychologist Erik Turkheimer proposed three “laws of behavioral genetics” (LoBG hereafter):

● First Law. All human behavioral traits are heritable.
● Second Law. The effect of being raised in the same family is smaller than the effect of genes.
● Third Law. A substantial portion of the variation in complex human behavioral traits is not accounted for by the effects of genes or families. (Turkheimer, 2000: 160)

In March of 2021, I asked Turkheimer how he defined “law.” He responded: “With tongue in cheek. In fact, it’s a null hypothesis: an expected result when nothing in particular is going on.

In 2015, Chabris et al (2015) proposed a 4th “law”, that a typical behavioral trait is associated with many variants which each explain a small amount of behavioral variability. They state that the “4th law” explains the failure of candidate gene studies and also the need for higher sample sizes in GWA studies. (It seems they are not aware that larger sample sizes increase the probability of spurious correlations—which is all GWA studies are; Claude and Longo, 2016; Richardson, 2017; Richardson and Jones, 2019) Nice ad hoc hypothesis to save their thinking.

One huge proponent of the LoBG is JayMan, who has been on a crusade for years pushing this nonsense. He added a “5th law” proposed by Emil Kirkegaard, which states that “All phenotypic relationships are to some degree genetically mediated or confounded.”

But what is a “law” and are these “laws of behavioral genetics” laws in the actual sense? First I will describe what a “law” is and if there even are biological laws. Then I will address each “law” in turn. I will then conclude that the LoBG aren’t real “laws”, they are derived from faulty thinking about the relationship between genes, traits, environment and the system and how the “laws” were derived rest on false assumptions.

What is a law? Are there biological laws?

Laws are “true generalizations that are “purely quantitative” … They have counterfactual force” (Sober, 1993: 458). Philosopher of mind Donald Davidson argued that laws are strict and exceptionless (Davidson, 1970; David-Hillel, 2003). That is, there must be no exceptions for that law. Sober (1993) discusses Rosenberg’s and Beatty’s arguments against laws of biology—where Rosenberg states that the only law in biology is “natural selection.” (See Fodor, 2008 and Fodor and Piattelli-Palmarini, 2009, 2010 for the argument against that claim and for arguments against the existence of laws of selection that can distinguish between causes and correlates of causes.) It has even been remarked that there are “so few” laws in biology (Dhar and Giuliani, 2010; also see Ruse, 1970).

Biology isn’t reducible to chemistry or physics (Marshal, 2021), since there are certain things about biology that neither chemistry or physics have. If there are laws of biology, then they will be found at the level of the organism or its ecology (Rull, 2022). In fact, it seems that although three laws of biology have been proposed (Trevors and Sailer Jr., 2008), they appear to be mere regularities, including McShea and Brandon’s (2010) first law of biology; all “laws of biology” seem to be mere laws of physics (Wayne, 2020). The “special sciences”, it seems, “are not fit for laws” (Kim, 2010). There seem to be, though, no uncontroversial laws or regularities in biology (Hamilton, 2007).

Now that I have described what laws are and have argued that there probably aren’t any biological laws, what does that mean for the LoBG? I will take each “law” in turn.

“Laws” of behavioral genetics

(1) All human behavioral traits are heritable.

JayMan gives derivations for the “laws”, and (1) and (2) have their bases in twin studies. We know that the equal environments assumption is false (Charney, 2012; Joseph, 2014; Joseph et al, 2015), and so if the EEA is false then we must reject genetic claims from twin study proponents. Nevertheless, the claim that these “laws” have any meaning gets pushed around a lot.

When it comes to the first law, the claim is that “ALL human behavioral traits are heritable”—note the emphasis on “ALL.” So this means that if we find only ONE behavioral trait that isn’t heritable, then the first law is false.

Reimann, Schilke, and cook (2017) used a sample of MZ and DZ twins and asked questions related to trust and distrust. They, of course, claim that “MZ and DZ twins share comparable environments in their upbringing“—which is false since MZ twins have more comparable environments. Nevertheless, they conclude that while trust has a heritability or 30%, “ACE analyses revealed that the estimated heritability [for] distrust is 0%.” This,therefore, means, that the “1st law” is false.

This “first law”, the basis of which is twin, family, and adoption studies, is why we have poured countless dollars into this research, and of course people have their careers (in what is clear pseudoscience) to worry about, so they won’t stop these clearly futile attempts in their search for “genes for” behavior.

(2) The effect of being raised in the same family is smaller than genes.

This claim is clearly nonsense, and one reason why is that the first “law” is false. In any case, there is one huge effect on, children’s outcomes due to birth order and how, then, parental attitudes–particularly mothers—affect child outcomes (Lehmann, Nuevo-Chiquero, and Vidal-Fernandez, 2018).

Why would birth order have an effect? Quite simply, the first-born child will get more care and attention than children who are born after, and so variations in parental behavior due to birth order can explain differences in education and life outcomes. They conclude that “broad shifts in parental behavior appear to set later-born children on a lower path for cognitive development and academic achievement, with lasting impact on adult outcomes.” Thus, Murray’s (2002) claim that birth order doesn’t matter and JayMan’s claim that “that the family/rearing environment has no effect on eventual outcomes” is clearly false. Thus, along with this and the falsity of the “1st law”, the “2nd law” is false, too.

(3) A substantial portion of the variation in complex human behavioral traits is not accounted for by the effects of genes or families.

This “law” covers the rest of the variance not covered in the first two “laws.” It was coined due to the fact that the first two “laws” had variance left that wasn’t “explained” by them. So this is basically unique experience. This is what behavioral genetics call “non-shared environment.” Of course, unique experiences (that is, subjective experiences) would definitely “shape who we are”, and part of our unique experiences are cultural. We know that cultural differences can have an impact on psychological traits (Prinz, 2014: 67). So the overall culture would explain why these differences aren’t “accounted for” in the first two “laws.”

Yet, we didn’t need the LoBG for us to know that individual differences are difference-makers for differences in behavior and psychology. So this means that what we choose to do can affect our propensities and then, of course, our behavior. Non-shared environmental effects are specific to the individual, and can include differing life events. That is, they are random. Non-shared environment, then, is parts of the environment that aren’t shared. Going back to Lehmann, Nuevo-Chiquero, and Vidal-Fernandez (2018) above, although children to grow up in the same family under the same household, they are different ages and so they also experience different life events. They also experience the same things differently, due to the subjectivity of experience.

In any case, the dichotomy between shared and non-shared environment is a dichotomy that upholds the behavioral geneticists main tool—the heritability estimate—from which these “laws” derive (from studies of twins, adoptees, and families). So, due to how the law was formulated (since there were still portions “unaccounted for” by the first two “laws”), it doesn’t really matter and since it rests on the first two false “laws”, therefore the third “law” is also false.

(4) Human behavioral traits are associated with many genes of small effect which contribute to a small amount of behavioral variability.

This “law” was formulated by Chabris et al (2015) due to the failure of molecular genetic studies which hoped to find genes with large effects to explain behavior. This “law” also “explains why the results of “candidate-gene” studies, which focus on a handful of genetic variants, usually fail to replicate in independent samples.” What this means to me is simple—it’s an ad-hoc account, meaning it was formulated to save the gene-searching by behavioral geneticists since the candidate gene era was a clear failure, as Jay Joseph noted in his discussion of the” 4th law.”

So here is the time line:

(1) Twin studies show above-0 heritabilities for behavioral traits.

(2) Since twin studies show high heritabilities for behavioral traits, then there must be genes that will be found upon analyzing the genome using more sophisticated methods.

(3) Once we started to peer into the genome after the completion of the human genome project, we then came to find candidate genes associated with behavior. Candidate gene studieslook at the genetic variation associated with disease within a limited number of pre-specified genes“, they refer to genes “believed to be” associated with a trait in question. Kwon and Goat (2000) wrote that “The candidate gene approach is useful for quickly determining the association of a genetic variant with a disorder and for identifying genes of modest effect.” But Sullivan (2020) noted that “Historical candidate gene studies didn’t work, and can’t work.” Charney (2022) noted that the candidate gene era was a “failure” and is now a “cautionary tale.”

Quite clearly, they were wrong then, and the failure of the candidate gene era led to the ad-hoc “4th law.” This has then followed us to the GWAS and PGS era, where it is claimed that we aren’t finding all of the heritability that twin studies say we should find with GWAS, since the traits under review are due to many genes of small effect. It’s literally just a shell game—when one claim is shown to be false, just make a reason why what you thought would be found wasn’t found, and then you can continue to search for genes “for” behavior. But genetic interactions create a “phantom heritability” (Zuk et al, 2011), while behavioral geneticists assume that the interactions are additive. They simply outright ignore interactions, although they pay it lip service.

So why, then, should we believe behavioral geneticists today in 2023 that we need larger and larger samples to find these mythical genes “for” behavior using GWAS? We shouldn’t. They will abandon GWAS and PGS in a few years when the new kid on the block shows up that they can they champion and claim that the mythical genes “for” behavior will finally be found.

(5) All phenotypic relationships are to some degree genetically mediated or confounded.

This claim is something that comes up a lot—the claim of genetic confounding (and mediation). A confound is a third variable that influences both the dependent and independent variable. The concept of genetic confounding was introduced during the era where it was debated whether or not smoking caused lung cancer (Pingault et al, 2021). (Do note that Ronald Fisher (1957), who was a part of this debate, claimed that smoking and lung cancer were both “influenced by a common cause, in this case individual genotype.

However, in order for the genetic confounding claim to work, they need to articulate a mechanism that explains the so-called genetic confounding. They need to articulate a genetic mechanism which causally explains X and Y, explains X independent of Y and explains Y independent of X. So for the cop-out genetic confounding claim to hold any water: G confounds X and Y iff there is a genetic mechanism which causally explains X and Y, causally explains X independent of Y and Y independent of X.

Conclusion

The “laws of behavioral genetics” uphold the false dichotomy of genes and environment, nature and nurture. Though, developmental systems theorists have rightly argued that it is a false dichotomy (Homans, 1979; Moore, 2002; Oyama, 2002; Moczek, 2012) and that it is just not biologically plausible (Lewkowicz, 2012). In fact, the h2 statistics assumes that G and E are independent, non-interacting factors, so if the claim is false then—for one of many reasons—we shouldn’t accept their conclusions. The fact that G and E interact means that, of course, we should reject h2 estimates, and along with it, the entire field of behavioral genetics.

Since the EEA is false, h2 equals c2. Furthermore, h2 equals 0. So Polderman’s (2015) meta analysis doesn’t show that for all traits in the analysis that h2 equals 49%. (See Jay Joseph’s critique.) Turkheimer (2000: 160) claimed that the nature-nurture debate is over, since everything is heritable. However, the debate is over because developmental systems approach has upended the false dichotomy of nature vs nurture, since all developmental resources interact and are therefore irreducible to development.

However, for the field to continue to exist, they need to promulgate the false dichotomy, since their heritability estimates depend on it. They also need to hold onto the claim that twin, family and adoption studies can show the “genetic influence” on traits to justify the continued search for genes “for” behavior. Zuk and Spencer (2020) called the nature-nurture “debate” “a zombie idea, one that, no matter how many times we think we have disposed of it, springs back to life.” This is just like Oyama (2000) who compared arguing against gene determinism like battling the undead (Griffiths, 2006).

Jay Joseph proposed a 5th “law” in 2015 where he stated:

Behavior genetic Laws 1-4 should be ignored because they are based on many false assumptionsconceptsand models, on negative gene finding attempts, and on decades of unsubstantiated gene discovery claims.

The “laws” should quite obviously be ignored. Since the whole field of behavioral genetics is based on them, why not abandon the search for “genes for behavior”? At the end of the day, it seems like there are no “laws” of behavioral genetics, since laws are strict and exceptionless. So why do they keep up with their claims that their “laws” tell us anything about human behavior? Clearly, it’s due to the ideology of those who hold that the all-important gene causes traits and behavior, so they will do whatever it takes to “find” them. But in 2023, we know that this claim is straight up false.

The Racial Identity Thesis: Why Race is a Social Construct of a Biological Reality

2250 words

Introduction

“Race” is a heavy topic. It influences many aspects of our daily lives, and in some cases, it even influences people to carry out heinous acts on groups of people deemed “inferior.” Such thoughts of “inferiority” of groups deemed (racialized as) races comes from horribly interpreting scientific studies, and in some cases, it is outright stated by the authors themselves, albeit using flowery language (eg Rushton, 2000). People that believe in the reality of race are called “race realists”, whether or not they hold a biological or social view of race.

On the socialrace side, social groups are racialized as races. “Social identities are contextual to historical and cultural elements.” “Racialization” is “the process through which groups come to be understood as major biological entities and human lineages, formed due to reproductive isolation, in which membership is transmitted through biological descent” (Hochman, 2019). Thinking about the root words for “racialization”, the two words are “racial formation.” Thus, when a group becomes racialized, it becomes a race based on societal expectations and thought.

“Hispanics” are an easy example to illustrate this case. In daily American discourse, people speak of “Hispanics”, “Spanish” people, and “Latinos/Latin Americans.” There are 33 Latin American countries to which this designation can be assigned to. In any case, what I term the “HLS distinction” (Hispanic, Latin American, Spanish distinction) is clearly a social designation and not a biological one. The group is an amalgamation of different peoples with differing amount ancestry to different countries of the world. “Hispanics” in some studies (eg Risch et al, 2002) don’t cluster in their own cluster (which would be taken to be a race, given that one has an argument for the claim), since they are an amalgam of different racial groups, it makes sense that they “clustered variously with the other groups.” Thus, “HLSs” are a social, not biological, kind.

Race concepts

Since “HBDers” call themselves “race realists” too, there needs to be a distinction between those who believe what I term “psychological race realism”—the claim that the psychological differences between the races are genetically transmitted and reduced to physical things—and regular old race realism—the claim that race exists as a social construct of a biological reality. Kaplan and Winther (2014) distinguish between bio-genomic/cluster race realism (the claim that race is real on the basis of genomic clusters from studies using programs like STRUCTURE); biological race realism affirms a kind of one-to-one mapping between social groups and clusters in DNA studies (though this not need be the case); and social race realism which is the racialization of social groups. They are anti-realists about biological race, but conventionalists about bio-genomic/cluster race. I think this is a good avenue to take, since the main program that psychological hereditarians push cannot ever be logically viable due to the irreducibility of the mental. (The “HBD” (Human Biological Diversity type racial realism is what I term psychological racial realism, while the type of racial realism I push is bio-genomic/cluster realism.)

In 2017, philosopher of race Michael Hardimon published his Rethinking Race: The Case for Deflationary Realism. In the book, Hardimon distinguishes between four race concepts: racialist races, in which it’s proponents attempt to rank-order racialized groups in socially-valued traits. Racialist races are socially constructed groups which then purport to pick out biological kinds. However, the concept of racialist race does not refer to any group in the world, genetic variation in Homo sapiens is nonconcordant, human variation is clinal, that is human groups aren’t sharply distinguished between one another on the basis of genomic data. Socialraces are groups taken to be racialist races, the social position taken by a group said to be a socialrace, or the system of positions that are social races (Hardimon, 2017: 131). So these two concepts need to be grouped together since they are hierarchical, though they need not be correlated with each other that strongly—one can be an anti-realist about biological races but be a realist about social races. This is why there is no contradiction in saying that biological race isn’t real but socialraces are. Socialraces also have a biological correlate, and these are what Hardimon terms “minimalist races”, which I will describe below.

Hardimon (2017: 69) has what he calls “an argument from the minimalist biological phenomenon of race” (Hardimon, 2017: 70):

Step 1. Recognize that there are differences in patterns of visible physical features of human beings that correspond to their differences in geographic ancestry.

Step 2. Observe that these patterns are exhibited by groups (that is, real existing groups).

Step 3. Note that the groups that exhibit these patterns of visible physical features correspond to differences in geographical ancestry satisfy the conditions of the minimalist concept of race.

Step 4. Infer that minimalist race exists.

Basically, if minimalist races exist then races exist because minimalist races are races. Contrast this argument (and Spencer’s newest argument arguing for the existence of OMB races) with hereditarian reasoning on race—basically just assume it’s existence.

The minimalist race concept does not state which populations are races, it just states that race exists. Hardimon’s populationist race concept (PRC) does, though. Although we don’t need genes to delineate race, using new technologies can and does help us to elucidate the existence of races. However, if minimalist races are populationist races, then the kind minimalist race equals populationist race. So if minimalist races are real, then so are populationist races.

Since 2014 in his paper A Radical Solution to the Race Problem, philosopher of race and science Quayshawn Spencer has been tinkering with an argument he now calls “the identity argument” (which I will provide below). The Office of Management and Budget has guidelines for the classification of people on the US census, being White, Black, Native American, Pacific Islander and Asian. The OMB never calls race a kind or a category, but they do refer to races as a set of categories or a proper name for population groups (Spencer, 2014). Spencer (2019a: 113) makes and defends 3 claims in regard to his OMB race theory:

(3.7) The set of races in OMB race talk is one meaning of ‘race’ in US race talk.

(3.8) The set of races in OMB race talk is the set of human continental populations.

(3.9) The set of human continental populations is biologically real.

In Spencer’s (2022) chapter in Remapping Race in a Global Context, A metaphysical mapping problem for race theorists and human population geneticists, Spencer has—in my opinion—articulated the best version of his OMB race argument.

Spencer’s identity argument

Of course, in the discussion of STRUCTURE and how many clusters it is told to construct from genomic data, the program is merely doing whag the human tells it to do. However, the five populations that come out in K= 5 “are genetically structured … which is to say, meaningfully demarcated solely on the basis of genetic markers” (Hardimon, 2017: 88). Hardimon (2017: 85) calls these clusters” continental-level minimalist races” whereas Spencer (2022: 278) calls these “the human continental populations” or “Blumenbachian partitions(Spencer, 2014: 1026).

K = 5 shows 5 human continental populations which are robust and have been replicated numerous times. The 5 human continental populations are Africans, Caucasians, East Asians, Native Americans, and Oceanians. K = 5 corresponds to the racial scheme used by the OMB, and by scientists and Americans in daily life. Human continental populations and OMB races correspond one-to-one in the following way: African, black; Caucasian, white; East Asian, Asian; Native American, American Indian; and Oceanian, Pacific Islander. So there is a metaphysical relation between the human continental populations in K = 5 and the OMB races. Spencer states that “identity” is what is exemplified between K = 5 and OMB races, and he then constructed this argument he calls “the identity thesis”:

(2.1) The identity thesis is true if, in OMB race talk, ‘American Indian,’ ‘Asian’, ‘Black’, ‘Pacific Islander,’ and ‘White’ are singular terms, and ‘American Indian’ means Native American, ‘Asian’ means East Asian, ‘Black’ means African, ‘Pacific Islander’ means Oceanian, and ‘White’ means Caucasian.

(2.2) In OMB race talk, the first conjunct in (2.1)’s antecedent is true.

(2.3) In OMB race talk, the second conjunct in (2.1)’s antecedent is true.

(2.4) So, the identity thesis is true.

(2.1) is true since the antecedent analytically entails the consequent; (2.2) is true based on the OMB’s intentions when they when coining the race terms, that is to provide a common language across US agencies; (2.3) is true since human continental populations are the best choices for the meanings of the OMB race terms—they posit referents which are part of the semantic content of the OMB race terms; it therefore follows that the conclusion (2.4) logically follows so the argument is deductively valid and I think it is sound.

Spencer (2019a) professes to be a radical pluralist about race—that is, there are many different race concepts for certain contexts based on American race talk. The referent “race” is merely a proper name for a set of human population groups. Race is therefore a social construct of a biological reality.

Conclusion

Race is meaningful in American social discourse—call it race talk. We should not be eliminativist about race, since there is medical relevance for certain groups deemed races. That is, we shouldn’t eliminate the concept RACE from our vocabulary, since it has a referent. RACE is a social construct of a biological reality. It must be said, though, that many people believe that since X is a social construct then X is therefore not real or doesn’t exist. But this line of thinking can be easily countered. Money is a social construct, so does that mean that money doesn’t exist so money isn’t real? No—the example perfectly shows the fallacious thinking of those who think that social constructivists are claiming that since X is socially constructed then X is not real. This couldn’t be further from the truth since social constructivists about race are realists about race. Pluralism about race is true—that is, there are many natures and realities for race based on the relevant context (Spencer, 2019: 27).

Lastly, these theories of race I have shown here do not license the claims that realists about racialist races or biological racial realists (as termed by Kaplan and Winther) have about these groups. While Hardimon distinguishes between the four concepts of race, Spencer’s concept and then argument tried to argue that race is a social construct of a biological reality.

Spencer even worries that those—like Charles Murray who don’t even have a coherent concept of race—may try to use his research for their own purposes:

Nevertheless, I do worry that politically right-winged people—for example, Charles Murray—might try to misuse my research for their own purposes. I’m also worried about the educational research that shows learning about human genetic differences in racial terms—for instance, lactase persistence alleles—increases racist attitudes among the learners.

Spencer took care of the first part as early as his 2014 paper and has reiterated it since—the DNA evidence that elucidates the reality of human races are on noncoding DNA, so Spencer (2014: 1036) states:

Nothing in Blumenbachian race theory entails that socially important differences exist among US races. This means that the theory does not entail that there are aesthetic, intellectual, or moral differences among US races. Nor does it entail that US races differ in drug metabolizing enzymes or genetic disorders. This is not political correctness either. Rather, the genetic evidence that supports the theory comes from noncoding DNA sequences. Thus, if individuals wish to make claims about one race being superior to another in some respect, they will have to look elsewhere for that evidence.

As for Spencer’s second worry, since racism is borne of ignorance, education can ameliorate racist attitudes (Hughes et al, 2007; Kuppens et al, 2014; Donovan, 2019, 2022).

Although I am a pluralist about race, the race concept I think best shows the reality of race is Spencer’s as he holds race to be a social construct of a biological reality in his OMB race theory. And we can easily see that RACE is a social construct of a biological reality with Spencer’s identity thesis by looking at the mapping between K = 5 and the OMB—the social part is the OMB designations of races, whereas the biological part are the clusters that appear in K = 5.

My thinking on race has changed a lot over the years. I used to be against the claim that race is a social construct. Though, if you know that being a social constructivist about race means that you don’t need to be an antirealist about the concept RACE as a whole, then you can state “Race is a social construct of a biological reality”, where “biological reality” means the concept RACE has some biological grounding, as seen in K = 5 STRUCTURE studies. RACE is an idea invented by people, which is the “social construct” claim. Thus, we impute racial categories onto people (the social construct part of the argument) and the racial categories we impute onto people have a biological grounding (as seen by K = 5).

P1: If RACE is a social construct of a biological reality, then RACE is not an inherent characteristic of the individual.

P2: If RACE is not an inherent characteristic of the individual, then one’s RACE is designated by the society they live in.

C: Thus, if RACE is a social construct of a biological reality, then one’s RACE is designated by the society they live in.

Evolutionary Psychology Does Not Explain Differences Between Rightists and Leftists

2000 words

Unless you’ve been living under a rock since the new year, you have heard of the “coup attempt” at the Capitol building on Wednesday, January 6th. Upset at the fact that the election was “stolen” from Trump, his supporters showed up at the building and rushed it, causing mass chaos. But, why did they do this? Why the violence when they did not get their way in a fair election? Well, Michael Ryan, author of The Genetics of Political Behavior: How Evolutionary Psychology Explains Ideology (2020) has the answer—what he terms “rightists” and “leftists” evolved at two different times in our evolutionary history which, then, explains the trait differences between the two political parties. This article will review part of the book—the evolutionary sections (chapters 1-3).

EP and ideology

Explaining why individuals who call themselves “rightists and leftists” behave and act differently than the other is Ryan’s goal. He argues, at length, that the two parties have two different personality profiles. This, he claims, is due to the fact that the ancestors of rightists and leftists evolved at two different times in human history. He calls this “Trump Island” and “Obama Island”—apt names, especially due to what occurred last week. Ryan claims that what makes Trump different from, say, Obama, is that his ancestors evolved at a different place in a different time compared to Obama’s ancestors. He further claims using the Stanford Prison Experiment that “we may not all be capable of becoming Nazis, after all. Just some, and conservatives especially so” (pg 12).

In the first chapter he begins with the usual adaptationism that Evolutionary Psychologists use. Reading between the lines in his implicit claims, he is arguing that “rightists and leftists” are natural kinds—that is, they are *two different kinds of people.* He explains some personality differences between rightists and leftists and then says that such trait differences are “rooted in biology and governed by genes” (pg 17). Ryan then makes a strong adaptationist claim—that traits are due to adaptation to the environment (pg 17). What makes you and I different from Trump, he claims, is that our ancestors and his ancestors evolved in different places at different times where different traits would be imperative to survival. So, over time, different traits got selected-for in these two populations leading to the trait differences we see today. So each environment led to the fixation of different adaptive traits which explains the differences we see today between the two parties, he claims.

Ryan then shifts from the evolution of personality differences to… The evolution of the beaks of Darwin’s finches and Tibetan adaptation to high-altitude living (pg 18), as if the evolution of physical traits is anything like the evolution of psychological traits. His folly is assuming that these physical traits can then be likened to personality/mental traits. The ancestors of rightists and leftists, like Darwin’s finches Ryan claims, evolved on different islands in different moments of evolutionary time. They evolved different brains and different adaptive behaviors on the basis of the evolution of those different brains. Trump’s ancestors were authoritarian, and this island occurred early in human history “which accounts for why Trump’s behavior seems so archaic at times” (pg 18).

The different traits that leftists show in comparison to rightists is due to the fact that their island came at a different point in evolutionary time—it was not recent in comparison to the so-called archaic dominance behavior portrayed by Trump and other rightists. Ryan says that Obama Island was more crowded than Trump Island where, instead of scowling, they smiled which “forges links with others and fosters reciprocity” (pg 19). So due to environmental adversity, they had a more densely populated “island”—in this novel situation, compared to the more “archaic” earlier time—the small bands needed to cooperate, rather than fight with each other, to survive. So this, according to Ryan, explains why studies show more smiling behavior in leftists compared to rightists.

Some of our ancestors evolved traits such as cooperativeness the aided the survival of all even though not everyone acquired the trait … Eventually a new genotype or subpopulation emerged. Leftist traits became a permanent feature of our genome—in some at least. (pg 19-20)

So the argument goes: Differences between rightists and leftists show us that the two did not evolve at the same points in time since they show different traits today. Different traits were adaptive at different points in time, some more archaic, some more modern. Since Trump Island came first in our evolutionary history, those whose ancestors evolved there show more archaic behavior. Since Obama Island came first, they show newer, more modern behaviors. Due to environmental uncertainty, those on Obama Island had to cooperate with each other. The trait differences between these two subpopulations were selected for in their environment that they evolved in, which is why they are different today. Now today, this led to the “arguing over the future direction of our species. This is the origin of human politics” (pg 20).

Models of evolution

Ryan then discusses four models of evolution: (1) the standard model, where “natural selection” is the main driver of evolutionary change; (2) epigenetic models like Jablonka’s and Lamb’s (2005) in Evolution in Four Dimensions; (3) where behavioral changes change genes; and (4) where organisms have phenotypic plasticity and is a way for the organism to respond to sudden environmental changes. “Leftists and rightists“, writes Ryan, “are distinguished by their own versions of phenotypic plasticity. They change behavior more readily than rightists in response to changing environmental signals” (pg 29-30).

In perhaps the most outlandish part of the book, Ryan articulates one of my now-favorite just-so stories. The passage is worth quoting in-full:

Our direct ancestor Homo erectus endured for two million years before going extinct 400,000 years ago when earth temperatures dropped far below the norm. Descendants of erectus survived till as recently as 14,000 years ago in Asia. The round head and shovel-shaped teeth of some Asians, including Vladimir Putin, are an erectile legacy. Archeologists believe erectus was a mix of Ted Bundy and Adolf Hitler. Surviving skulls point to a life of constant violence and routine killing. Erectile skulls are thick like a turtle’s, and the brow’s are ridged for protection from potentially fatal blows. Erectus’ life was precarious and violent. To survive, it had to evolve traits such as vigilant fearfulness, prejudice against outsiders, bonding with kin allies, callousness toward victims, and a penchant for inflexible habits of life that were known to guarantee safety. It had to be conservative. 34 Archeologists suggest that some of our most characteristic conservative emotions such as nationalism and xenophobia were forged at the time of Homo erectus. 35 (pg 33-34)

It is clear that Ryan is arguing that rightists have more erectus-like traits whereas leftists have more modern, Sapiens traits. “The contemporary coexistence of a population with more “modern” traits and a population with more “archaic” traits came into being” (pg 37). He is implicitly assuming that the two “populations” he discusses are natural kinds and with his “modern” “archaic” distinction (see Crisp and Cook 2005 who argue against a form of this distinction) he is also implying that there is a sort of “progress” to evolution.

Twin studies, it is claimed, show “one’s genetically informed psychological disposition” (Hatemi et al, 2014); they “suggest that leftists and rightists are born not made” while a so-called “consensus has emerged amongst scientists: political behavior is genetically controlled and heritable” (pg 43). But, Beckway and Morris (2008), Charney (2008), and Joseph (2009; 2013) argue that twin studies can do no such thing due to the violation of the equal environments assumption (Joseph, 2014; Joseph et al, 2015). Thus, Ryan’s claims of the “genetic origins” of political behavior rest on studies that cannot prove or disprove “genetic causation” (Shulitziner, 2017)—but since the EEA is false we must discount “genetic causation” for psychological traits, not least because it is impossible for genes to cause/influence psychological traits (see argument (iii)).

The arguments he provides are a form of inference to best explanation (IBE) (Smith, 2016). However, this is how just-so stories are created: the conclusion is already in mind, and then the story is crafted using “natural selection” to explain how a trait came to fixation and why it currently exists today. The whole book is full of such adaptive stories. Claiming that we have the current traits we do in the distributions they are in in the “populations” because they were, at a certain point in our evolutionary history, adaptive which then led to the individuals with those traits passing on more of their genes, eventually leading to trait fixation. (See Fodor and Piattelli-Palmarini, 2010).

Ryan makes such outlandish claims such as “Rightists are more likely than leftists to keep their desks neat. If in the distant past you knew exactly where the weapons were, you could find them quickly and react to danger more effectively. 26” (pg 45). He talks about how “time-consuming and effort-demanding accuracy of perception [were] more characteristic of leftist cognitionleftist cognition is more reflective” while “rightist cognition is intuitive rather than reflective” (pg 47). Rightists being more likely to endorse the status quo, he claims, is “an adaptive trait when scarce resources made energy management essential to getting by” (pg 48) Rightist language, he argues, uses more nouns since they are “more concrete, an anxious personalities prefer concrete to abstract language because it favors categorial rigidity and guarantees greater certainty” while leftists “use words that suggest anxiety, anger, threats, certainty, resistance to change, power, security, and conformity” (pg 49). There is “a connection between archaic physiology and rightist moral ideology” (pg 52). Certain traits that leftists have were “adaptive traits [that] were suited to later stage human evolution” (pg 53). Ryan just cites studies that show differences between rightists and leftists and then uses some great leaps and mental gymnastics to try to mold the findings as being due to evolution in the two different time periods he describes in chapter 1 (Trump and Obama Island).

Conclusion

I have not read one page in this book that does not have some kind of adaptive just-so story attempting to explain certain traits/behaviors between rightists and leftists in evolutionary terms. Ryan uses the same kind of “reasoning” that Evolutionary Psychologists use—have your conclusion in mind first and then craft an adaptive story to explain why the traits you see today are there. Ryan outright says that “[t]raits are the result of adaptation to the environment” (pg 17), which is a rare—strong adaptationist—claim to make.

His book ticks off all of the usual EP things: strong adaptationism, just-so storytelling, the claim that traits were selected-for due to their contribution in certain environments at different points in time. The strong adaptationist claims, for example, are where he says that erectus’ large brow “are rigid for protection from potentially fatal blows” (pg 34). Such strong adaptationist claims imply that Ryan believes that all traits are the result of adaptation and that they, as a result, are still here today because they all serve a function in our evolutionary past. His arguments are, for the most part, all evolutionary and follow the same kinds of patterns that the usual EP arguments do (see Smith, 2016 for an explication of just-so stories and what constitutes them). Due to the problems with evolutionary psychology, his adaptive claims should be ignored.

The arguments that Ryan provides are not scientific and, although they give off a veneer of being scientific by invoking “natural selection” and adaptationism, they are anything but. It is just a long-winded explanation for how and why rightists and leftists—liberals and conservatives—are different and why they cannot change, since these differences are “encoded” into our genome. The implicit claim of the book, then, that rightists and leftists are two different—natural—kinds, lies on the false bed of EP and, therefore, the arguments provided in the book fail to sway anyone that does not believe such fantastic storytelling masquerading as science. While he does discuss other evolutionary theories, such as epigenetic ones from Jablonka and Lamb (2005), the book is largely strongly adaptationist using “natural selection” to explain why we still have the traits we do in different “populations” today.

White Privilege: What It Is and Who Has It?

2550 words

Discussions about whiteness and privilege have become more and more common. Whites, it is argued, have a form of unearned societal privilege which therefore explains certain gaps between whites and non-whites. White privilege is the privilege that whites have in society—this type of privilege does not have to be in America, it can hold for groups that are viewed as ‘white’ in other countries. This, then, perpetrates social views of race, hence these people are realists about race but in a social/political context and do not have to recognize race as biological (although race can become biologicized through social/cultural practices). This article will discuss (1) What white privilege is; (2) Who has white privilege; (3) Arguments against white privilege; and (4) If race doesn’t exist, why does white privilege matter?

What is white privilege?

The concept of white privilege, like most concepts, evolves with the times and current social thought. The concept was originally created in order to account for whites’ (unearned) privileges and the conscious bias that went into creating and then maintaining these privileges, to unconscious favoritism/psychological advantages that whites give other whites (Bennett, 2012: 75). That is, white privilege is “an invisible package of unearned assets that I can count on cashing in each day, but about which I was “meant” to remain oblivious. White privilege is like an invisible weightless knapsack of special provisions, maps, passports, codebooks, visas, clothes, tools , and blank checks” (McIntosh, 1988).

More easily, we can say that white privilege is—the privilege conferred, either consciously or subconsciously, to one based on their skin color or, as Sullivan (2016, 2019) argues, their class status ALONG WITH their whiteness is what we should be talking about—white privilege with CLASS in between ‘white’ and ‘privilege’. In this sense, one’s class status AND their whiteness is explanatory, not only the concept of whiteness (i.e., their socialrace). The concept of whiteness—one’s skin color—as the privilege leaves out numerous intricacies in how whiteness gives and upholds systemic discrimination. When we add the concept of ‘class’ into ‘white privilege’ we get what Sullivan terms ‘white class privilege’.

While yes, one’s race is an important variable in whether or not they have certain privileges, such privileges are held for middle- to upper-middle class whites. Thus, numerous examples of ‘white privilege’ are better understood as examples of ‘white class privilege’, since lower-class whites don’t have the same kinds of privileges, outlooks, and social status as middle- and upper-middle class whites. Of course, though, lower-class whites can benefit from their whiteness—they definitely can. But the force of Sullivan’s concept of ‘white class privilege’ is this: white privilege is not monolithic towards whites, and some non-whites are better-off (economically and in regard to health) than whites. Thus, according to Sullivan, ‘white privilege’ should be amended to ‘white class privilege’.

Who has white privilege?

Lower-class whites could, in a way, be treated differently than middle- and upper-class whites—even though they are of the same race. Lower-class whites can be seen to have ‘white privilege’ on the basis of everyday thought, since most think of the privilege as down to just skin color, yet there is an untalked about class dimension at play here, which, then, even gives blacks an advantage while upholding the privilege of the upper-class whites.

Non-whites who have are of a higher social class than whites would also receive different treatment. Sullivan states that the revised concept of ‘white class privilege’ must be used intersectionally—that is, privilege must be considered interacting with class, gender, national, and other social experiences. Sure, lower-class whites may be treated differently than higher-class blacks in certain contexts, but this does not mean that the lower-class white has ‘more privilege’ than the upper-class black. This shows that we should not assume that lower-class whites have the same kinds of privilege conferred by society as middle- and upper-class whites. Upper-class blacks and ‘Hispanics‘ may attempt to distinguish themselves from lower-class blacks and ‘Hispanics’, as Sullivan (2019: 18-19) explains:

Class privilege shows up as a feature of most if not all racial groups in which members with “more”—more money, education, or whatever else is valued in society—are treated better than those with “less.” For that reason, we might think that white class privilege actually is an intragroup pattern of advantage and disadvantage among whites, rather than an intergroup pattern that gives white people a leg up over non-white people. After all, many Black middle-class and upper-middle-class Americans also go to great lengths to make sure that they are not mistaken for the Black poor in public spaces: when they are shopping, working, walking, or driving in town, and so on (Lacy, 2007). A similar pattern can be found with middle-to-upper-class Hispanic/Latinx people in the United States, who can “protect” themselves from being seen as illegal immigrants by ensuring that they are not identified as poor (Masuoka and Junn, 2013).

Sullivan then goes on to state that these situations are not equivalent, since wealth, fame, and education do not protect upper-class blacks from racial discrimination. The certain privileges that upper-class whites have, thusly, do not transfer to upper-class blacks. Further, middle- to upper-class whites distinguish themselves as ‘good whites’ who are not racist, while dumping all of the racism accusations on lower-class whites. “…the line between “good” and “bad” white people drawn by many (good) white people is heavily classed. Good white people tend to be middle-to-upper-class, and they often dump responsibility for racism onto lower-class white people” (Sullivan, 2019: 35). Even though the lower-class whites get used as a ‘shield’, so to speak, by upper-class whites, they still have some semblance of white privilege, in that they are not assumed to be non-citizens to the US—something that ‘Hispanics’ do have to deal with (no matter their race).

While wealthy white people generally have more affordances than poor white people do, in a society that prizes whiteness all white people have some racial affordances, at least some of the time.

Paradoxically, whites are not the only ones that benefit off of ‘white privilege’—even non-whites can benefit, though it ultimately helps upper-class whites. They can benefit by being brought up in a white home, around whites (like being adopted or having one white parent while spending most of their childhood with their white family). Thus, white privilege can cross racial lines all the while still benefitting whites.

Sullivan (2019: chapter 2) discusses some blacks who benefit from white privilege. One of the people she discusses has a white parent. This is what gives her her lighter skin, but that is not where her privilege comes from (think colorism in the black community where lighter skin is more prized than darker skin). Her privilege came from “her implicit knowledge of white norms, sensibilities, and ways of doing things that came from living with and being accepted by white family members” (Sullivan, 2019: 26). This is what Sullivan calls “family familiarity” and is one of the ways that blacks can benefit from white privilege. Another way in which blacks can benefit from white privilege is due to “ancestral ties to whiteness.”

Colorism is the discrimination within the black community by skin color. Certain blacks may talk about “light-” and “dark-skinned” blacks and they may—ironically or not—discriminate on the basis of skin color. Such colorism is even somewhat instilled in the black community—where darker-skinned black sons and lighter-skinned black daughters report higher-quality parenting. Landor et al (2014) report that their “findings provide evidence that parents may have internalized this gendered colorism and as a result, either consciously or unconsciously, display higher quality of parenting to their lighter skin daughters and darker skin sons.” Thus, even certain blacks—in virtue of being ‘part white’—would benefit from white (skin) privilege within their own (black) community, which would therefore give them certain advantages.

Arguments against white privilege

Two recent articles with arguments against white privilege (Why White Privilege Is Wrong — Quillette and The Fallacy of White Privilege — and How It Is Corroding Society) erroneously argue that since other minority groups quickly rose up upon arrival to America, therefore white privilege is a myth. These kinds of takes, though, are quite confused. It does not follow that since other groups have risen upon entry into America and that since whites have worse outcomes on some—and not other—health outcomes, that therefore the concept of white privilege is ‘fallacious’; we just need something more fine-grained.

For example, the claims that X minority group is over-represented compared to whites in America gets used as a point that ‘white privilege’ does not exist (e.g., Avora’s article). Avora discusses the experiences and data from many black immigrants, proclaiming:

These facts challenge the prevailing progressive notion that America’s institutions are built to universally favor whites and “oppress” minorities or blacks. On the whole, whatever “systemic racism” exists appears to be incredibly ineffectual, or even nonexistent, given the multitude of groups who consistently eclipse whites.

How does that follow? In fact, how does the whole discussion of, for example, Japanese now outperforming whites follow that white privilege therefore is a ‘fallacy’? I ask the question, since Asian immigrants to America are hyper-selected (Noam, 2014; Zhou and Lee, 2017), meaning that what explains higher Asian academic achievement is academic effort (Hsin and Xie, 2014) and the fact that Asians are hyper-selected—meaning that they have a higher chance of having a higher degree.

The educational credentials of these recent [Asian] arrivals are striking. More than six-in-ten (61%) adults ages 25 to 64 who have come from Asia in recent years have at least a bachelor’s degree. This is double the share among recent non-Asian arrivals, and almost surely makes the recent Asian arrivals the most highly educated cohort of immigrants in U.S. history.

Compared with the educational attainment of the population in their country of origin, recent Asian immigrants also stand out as a select group. For example, about 27% of adults ages 25 to 64 in South Korea and 25% in Japan have a bachelor’s degree or more.2 In contrast, nearly 70% of comparably aged recent immigrants from these two countries have at least a bachelor’s degree. (The Rise of Asian Americans)

Avora even discuses some African immigrants, namely Nigerians and Ghanaians. However, just like Asian immigrants to America, Nigerian and Ghanaian immigrants to America are more likely to hold advanced degrees, signifying that they are indeed hyper-selected in comparison to the population that they derive from (Duvivier, Burch, and Boulet, 2017). Thus, to go along with the stats that Avora cites on the children of Nigerian immigrants, their parents already had higher degrees, signifying that they are indeed a hyper-selected group. This means that such ethnic groups cannot be used to show that white privilege is explanatory.

While Avora does discuss “class” in his article, he shows that it’s not only ‘white privilege’, but the class element that comes along with whiteness in America. He therefore unknowingly shows that once you add the ‘class’ factor and create the concept of ‘white class privilege’, that this privilege can cross racial lines and benefit non-whites.

In the Harinam and Henderson Quillette article, they argue that since there are some things that we say are ‘good’ that non-whites have more of than whites, therefore the concept of ‘white privilege’ does not explain the existence of disparities between ethnic groups in the US since some some bad things happen to whites and some good things happen to non-whites—but this is an oversimplification. The fact of the matter is, whites that do receive privileges over other ethnic/racial groups do so not in virtue of their (white) skin privilege, but in virtue of their class privilege. This can be seen with the above citations on class being the explanatory variable regarding Asian academic success (showing how class values get reproduced in the new country which then explains the academic success of Asians in America).

The fact that both of these articles believe that by showing some minority groups in America have more ‘good’ things than whites or better outcomes for bad things (like suicides) misses the point. That whites kill themselves more than other American ethnic groups does not mean that whites do not have privilege in America compared to other groups.

If race doesn’t exist, then why does white privilege matter?

Lastly, those who argue against the concept of white privilege may say that those who are against the concept of white privilege would then at the same time say that race—and therefore whites—do not exist so, in effect, what are they talking about if ‘whites’ don’t exist because race does not exist? This is of course a ridiculous statement. One can indeed reject claims from biological racial realists and believe that race exists and is a socially constructed reality. Thus, one can reject the claim that there is a ‘biological’ European race, and they can accept the claim that there is an ever-changing ‘white’ race, in which groups get added or subtracted based on current social thought (e.g., the Irish, Italians, Jews), changing with how society views certain groups.

Though, it is perfectly possible for race to exist socially and not biologically. So the social creation of races affords the arbitrarily-created racial groups to be in certain areas on the hierarchy of races. Roberts (2011: 15) states that “Race is not a biological category that is politically charged. It is a political category that has been disguised as a biological one.” She argues that we are not biologically separated into races, we are politically separated into them, signifying race as a political construct. Most people believe that the claim “Race is a social construct” means that “Race does not exist.” However, that would be ridiculous. The social constructivist just believes that society divides people into races based on how we look (i.e., how we are born) and then society divides us into races on the basis of how we look. So society takes the phenotype and creates races out of differences which then correlate with certain continents.

So, there is no contradiction in the claim that “Race does not exist” and the claim that “Whites have certain unearned privileges over other groups.” Being an antirealist about biological race does not mean that one is an antirealist about socialraces. Thus, one can believe that whites have certain privileges over other groups, all the while being antirealists about biological races (saying that “Races don’t exist biologically”).

Conclusion

In this article I have explained what white privilege is and who has it. I have also discussed arguments against white privilege and claims that those who argue against race are hypocrites since they still talk about “whites” while claiming that race exists. After showing the conceptual confusions that people have about white privilege, along with the fact that groups that do better than whites in America (the groups that supposedly show that white privilege is “a fallacy”), I then forward Sulllivan’s (2016, 2019) argument on white class privilege. This shows that their whiteness is not the sole reason why they prosper—their whiteness along with their middle-to-upper-middle-class status explains why they prosper. It also, furthermore, shows that while lower-class whites do have some sort of white privilege, they do not have all of the affordances of white privilege due to their class status. Blacks can, too, benefit from white privilege, whether it’s due to their proximity to whiteness or their ancestral heritage.

White privilege does exist, but to fully understand it, we must add in the nexus of class with it.

Herrnstein’s Syllogism

2650 words

1. If differences in mental abilities are inherited, and
2. if success requires those abilities, and
3. if earnings and prestige depend on success,
4. then social standing will be based to some extent on inherited differences among people. (Herrnstein, 1971)

Richard Herrnstein’s article I.Q. in The Atlantic (Herrnstein, 1971) caused much controversy (Herrnstein and Murray, 1994: 10). Herrnstein’s syllogism argued that as environments become more similar and if differences in mental abilities are inherited and that success in life requires such abilities and if earning and prestige depends on success which is required by inheritable mental abilities then social standing will be based, “to some extent on inherited differences among people.” Herrnstein does not say this outright in the syllogism, but he is quite obviously talking about genetic inheritance. One can, however, look at the syllogism with an environmental lens, as I will show. Lastly, Herrnstein’s syllogism crumbles since social class is predictive of success in life when both IQ and social class are equated. So since family background and schooling explains the IQ-income relationship (a measure of success) then Herrnstein’s argument falls.

Note that Herrnstein came to measurement due to being a student of William Sheldon’s somatotyping. “Somatotyping lured the impressionable and young Herrnstein into a world promising precision and human predictability based on the measurement of body parts” (Hilliard, 2012: 22).

  1. If differences in mental abilities are inherited

Premise 1 is simple: “If differences in mental ability are inherited …” Herrnstein is obviously talking about genetic transmission, but we can look at this through a cultural/environmental lens. For example, Berg and Belmont (1990) showed that Jewish children of different socio-cultural backgrounds had different patterns of mental abilities, which were clustered in certain socio-cultural groups (all Jewish), showing that mental abilities are, in large part, culturally derived. Another objection could be that since there are no laws linking psychological/mental states with physical states (the mental is irreducible to the physical—meaning that mental states cannot be transmitted through (physical) genes) then such genetic transmission of psychological/mental traits is impossible. In any case, one can look at cultural transmission of mental abilities and disregard genetic transmission of psychological traits and the argument fails.

We can accept all of the premises of Herrnstein’s syllogism and argue an environmental case, in fact (bracketed words are my additions):

1. If differences in mental abilities are [environmentally] inherited, and
2. if success requires those [environmentally inherited] abilities, and
3. if earnings and prestige depend on [environmentally inherited] success,
4. then social standing will be based to some extent on [enviromnentally] inherited differences among people.

The syllogism hardly changes, but my additions change what Herrnstein was arguing for—environmental, not genetic differences cause success and along with it social standing among groups of people.

The Bell Curve (Herrnstein and Murray, 1994) can, in fact, be seen as an at-length attempt to prove the validity of the syllogism in an empiric matter. Herrnstein and Murray (1994: 105, 108-110) have a full discussion of the syllogism. “As stated, the syllogism is not fearsome” (Herrnstein and Murray, 1994: 105). They go on to state that if intelligence (IQ scores, AFQT scores) is only a bit influenced by genes and if success is only a bit influenced by intelligence then only a small amount of success is inherited (genetically). Note that their measure of “IQ” is the AFQT—which is a measure of acculturated learning, measuring school achievement (Roberts et al, 2000; Cascio and Lewis, 2005).

How much is IQ a matter of genes?“, Herrnstein and Murray ask. They then discuss the heritability of IQ, relying, of course, on twin studies. They claim that the heritability of IQ is .6 based on the results of many twin studies. But the fatal flaw with twin studies is that the EEA is false and, therefore, genetic conclusions should be dismissed outright (Burt and Simons, 2014, 2015; Joseph, 2015; Joseph et al, 2015; Fosse, Joseph, and Richardson, 2015; Moore and Shenk, 2016). Herrnstein (1971) also discusses twin studies in the context of heritability, attempting to buttress his argument. But if the main vehicle used to show that “intelligence” (whatever that is) is heritable is twin studies, why, then, should we accept the conclusions of twin research if the assumptions that make the foundation of the field are false?

Block (1995) quotes Murray’s misunderstanding about heritability in an interview Murray had while making tours for The Bell Curve:

When I – when we – say 60 percent heritability, it’s not 60 percent of the variation. It is 60 percent of the IQ in any given person.” Later, he repeated that for the average person, “60 percent of the intelligence comes from heredity” and added that this was true of the “human species,” missing the point that heritability makes no sense for an individual and that heritability statistics are population-relative.

So Murray used the flawed concept of heritability in the wrong way—hilarious.

So the main point of Herrnstein’s argument is that environments become more uniform for everyone, then the power of heredity will shine through since the environment is uniform—the same—for everyone. But even if we could make the environment “the same”. What does this even mean? How is my environment the same, even if the surroundings are the same, say, if I would react or see something differently than you do on the same thing? The subjectivity of the mental disproves the claim that environments can be “more uniform.” Herrnstein claimed that if no variance in environment exists, then the only thing that can influence success is heredity. This is not wrong, but how would it be possible to equalize environments? Are we supposed to start from square one? Give up the wealth and status of the rich and powerful and “equalize environments” and, according to Herrnstein and the ‘meritocracy’, those who had earnings and prestige, which depended on success which depended on inherited mental abilities would still float to the top.

But what happens when both social class and IQ are equated? What predicts life success? Stephen Ceci reanalyzed the data from Terman’s Termites (the term coined for those in the study) and found something quite different from what Terman had assumed. There were three groups in Terman’s study—group A, B, and C. Groups A and C comprised the top and bottom 20 percent of the full sample in terms of life success. So at the start of the study, all of the children “were about equal in IQ, elementary school grades, and home evaluations” (Ceci, 1996: 82). Depending on the test used, the IQs of the children ranged between 142 to 155, which then decreased by ten points during the second wave due to regression and measurement error. So although group A and C had equivalent IQs, they had starkly different life outcomes. (Group B comprised 60 percent of the sample and enjoyed mediocre life success.)

Ninety-nine percent of the men in the group that had the best professional and personal accomplishments, i.e., group A were individuals who came from professional or business-managerial families that were well educated and wealthy. In contrast, only 17% if the children from group C came from professional and business families, and even these tended to be poorer and less well educated than their group A peers. The men in the two groups present a contrast on all social indicators that were assesssed: group A individuals preferred to play tennis, while group C men preferred to watch football and baseball; as children, the group A men were more likely to collect stamps, shells, and coinds than were the group C men. Not only were the fathers of the group A men better educated than those of group C, but so were their grandfathers. In short, even though the men in group C had equivalent IQs to group A, they did not have equivalent social status. Thus, when IQ is equated and social class is not, it is the latter that seems to be deterministic of professional success. Therefore, Terman’s findings, far from demonstrating that high IQ is associated with real-world success, show that the relationship is more complex and that the social status of these so-called geniuses’ families had a “long reach,” influencing their presonal and professional achievments throughout their adult lives. Thus, the title of Terman’s volumes Genetic studies of Genius, appears to have begged the question of the causation of genius. (Ceci, 1996: 82-83)

Ceci used the Project Talent dataset to analyze the impact of IQ on occupational success. This study, unlike Terman’s, looked at a nationally representative sample of 400,000 high-school students “with both intellectual aptitude and parental social class spanning the entire range of the population” (Ceci, 1996: 85). The students were interviewed in 1960, then about 4,000 were again interviewed in 1974. “For all practical purposes, this subgroup of 4,000 adults represents a stratified national sample of persons in their early 30s” (Ceci, 1996: 86). So Ceci and his co-author, Henderson, ran several regression analyses that involved years of schooling, family and social background and a composite score of intellectual ability based on reasoning, math, and vocabulary. They excluded those who were not working at the time due to being imprisoned, being housewives or still being in school. This then left them with a sample of 2,081 for the analysis.

They looked at IQ as a predictor of variance in adult income in one analysis, which then showed an impact for IQ. “However, when we entered parental social status and years of schooling completed as additional covariates (where parental social status was a standardized score, mean of 100, SD = 10, based on a large number of items having to do with parental income, housing costs, etc.—ranging from low of 58 to high of 135), the effects of IQ as a predictor were totally eliminated” (Ceci, 1996: 86). Social class and education were very strongly related to predictors of adult income. So “this illustrates that the relationship between IQ and adult income is illusory because the more completely specified statistical model demonstrates its lack of predictive power and the real predictive power of social and educational variables” (Ceci, 1996: 86).

The considered high, average, and low IQ groups, about equal size, while examining the regressions of earnings on social class and education within the groups.

Regressions were essentially homogeneous and, contrary to the claims by those working from a meritocratic perspective, the slope for the low IQ group was steepest (see Figure 4.1). There was no limitation imposed by low IQ on the beneficial effects of good social background on earnings and, if anything, there was a trend toward individuals with low IQ actually earning more than those with average IQ (p = .09). So it turns out that although both schooling and parental social class are powerful determinants of future success (which was also true in Terman’s data), IQ adds little to their influence in explaining adult earnings. (Ceci, 1996: 86)

The same was also true for the Project Talent participants who continued school. For each increment of school completed, there was also an effect on their earnings.

Individuals who were in the top quartile of “years of schooling completed” were about 10 times as likely to be receiving incomes in the top quartile of the sample as were those who were in the bottom quartile of “years of schooling completed.” But this relationship does not appear to be due to IQ mediating school attainment or income attainment, because the identical result is found even when IQ is statistically controlled. Interestingly, the groups with the lowest and highest IQs both earned slightly more than average-IQ students when the means were adjusted for social class and education (unadjusted meansat the modal value of social class and education = $9,094, $9,242, and $9,997 for low, average, and hhigh IQ groups, whereas the unadjusted means at this same modal value = $9,972, $9,9292, and $9,9278 for the low, average, and high IQs.) (Perhaps the low IQ students were tracked into plumbing, cement finishing and other well-paying jobs and the high-IQ students were tracked intothe professions, while average IQ students became lower paid teachers. social workers, ministers, etc.) Thus, it appears that the IQ-income relationship is really the result of schooling and family background, and not IQ. (Incidentally, this range in IQs from 70 to 130 and in SES from 58 to 135 covers over 95 percent of the entire population.) (Ceci, 1996: 87-88)

Ceci’s analysis is just like Bowles and Nelson’s (1974) analysis in which they found that earnings at adulthood were more influenced by social status and schooling, not IQ. Bowles and Nelson (1974: 48) write:

Evidently, the genetic inheritance of IQ is not the mechanism which reproduces the structure of social status and economic privilege from generation to generation. Though our estimates provide no alternative explanation, they do suggest that an explanation of intergeneration immobility may well be found in aspects of family life related to socio-economic status and in the effects of socio-economic background operating both directly on economic success, and indirectly via the medium of inequalities in educational attainments.

(Note how this also refutes claims from PumpkinPerson that IQ explains income—clearly, as was shown, family background and schooling explain the IQ-income relationship, not IQ. So the “incredible correlation between IQ and income” is not due to IQ, it is due to environmental factors such as schooling and family background.)

Herrnstein’s syllogism—along with The Bell Curve (an attempt to prove the syllogism)—is therefore refuted. Since social class/family background and schooling explains the IQ-income relationship and not IQ, then Herrnstein’s syllogism crumbles. It was a main premise of The Bell Curve that society is becoming increasingly genetically stratified, with a “cognitive elite”. But Conley and Domingue (2015: 520) found “little evidence for the proposition that we are becoming increasingly genetically stratified.”

IQ testing legitimizes social hierarchies (Chomsky, 1972; Roberts, 2015) and, in Herrnstein’s case, attempted to show that social hierarchies are an inevitability due to the genetic transmission of mental abilities that influence success and income. Such research cannot be socially neutral (Roberts, 2015) and so, this is yet another reason to ban IQ tests, as I have argued. IQ tests are a measure of social class (Ceci, 1996; Richardson, 2002, 2017), and such tests were created to justify existing social hierarchies (Mensh and Mensh, 1991).

Thus, the very purpose of IQ tests was to confirm the current social order as naturally proper. Intelligence tests were not misused to support hereditary theories of social hierarchies; they were perfected in order to support them. The IQ supplied an essential difference among human beings that deliberately reflected racial and class stratifications in order to justify them as natural.9 Research on the genetics of intelligence was far from socially neutral when the very purpose of theorizing the heritability of intelligence was to confirm an unequal social order. (Roberts, 2015: S51)

Herrnstein’s syllogism seems valid, but in actuality, it is not. Herrnstein was implying that genes were the casue of mental abilities and then, eventually, success and prestige. But one can look at Herrnstein’s syllogism from an environmentalist point of view (do note that the hereditarian/environmentalist debate is futile and continues the claim that IQ tests test ‘intelligence’, whatever that is). When matched for IQ—in regard to Terman’s Termites—family background and schooling explained the IQ-income relationship. Further analyses showed that this, again, was the case. Ceci (1996) showed again, replicating Terman’s and Bowles’ and Nelson’s (1974) analyses that social class and schooling, not IQ, explains income’s relationship with IQ.

The conclusion of Herrnstein’s argument can, as I’ve already shown, be an environmental one—through cultural, not genetic, transmission. Such arguments that IQ is ‘genetic’ and, thusly, certain individuals/groups will tend to stay in their social class, as Pinker (2002: 106) states: “Smarter people will tend to float into the higher strata, and their children will tend to stay there.” This, as has been shown, is due to social class, not ‘smarts’ (scores on an IQ test). In any case, this is yet another reason why IQ tests and the research behind them should be banned: IQ tests attempt to justify the current social order as ‘inevitable’ due to genes that influence mental abilities. This claim, though, is false and, therefore—along with the fact that America is not becoming more genetically stratified (Conley and Domigue, 2015)—Herrnstein’s syllogism crumbles. The argument attempts to justify the claim that class has a ‘genetic’ component (as Murray, 2020, attempts to show) but subsequent analyses and arguments have shown that Herrnstein’s argument does not hold.

Test Validity, Test Bias, Test Construction, and Item Selection

3400 words

Validity for IQ tests is fleeting. IQ tests are said to be “validated” on the basis of performance with other IQ tests and that of job performance (see Richardson and Norgate, 2015). Further, IQ tests are claimed to not be biased against social class or racial group. Finally, through the process of “item selection”, test constructors make the types of distributions they want (normal) and get the results the want through the subjective procedure of removing items that don’t agree with their pre-conceived notions on who is or is not “intelligent.” Lastly, “intelligence” is descriptive measure, not an explanatory concept, and treating it like an explanatory measure can—and does—lead to circularity (of which is rife in the subject of IQ testing; see Richardson, 2017b and Taleb’s article IQ is largely a psuedoscientific swindle). This article will show that, on the basis of test construction, item analysis (selection and deselection of items) and the fact that there is no theory of what is being measured in so-called intelligence tests that they, in fact, do not test what they purport to.

Richardson (1991: 17) states that “To measure is to give … a more reliable sense of quantity than our senses alone can provide”, and that “sensed intelligence is not an objective quantity in the sense that the same hotness of a body will be felt by the same humans everywhere (given a few simple conditions); what, in experience, we choose to call ‘more’ intelligence, and what ‘less’ a social judgement that varies from people to people, employing different criteria or signals.Richardson (1991: 17-18) goes on to say that:

Even if we arrive at a reliable instrument to parallel the experience of our senses, we can claim no more for it than that, without any underlying theory which relates differences in the measure to differences in some other, unobserved, phenomena responsible for those differences. Without such a theory we can never be sure that differences in the measure correspond with our sensed intelligence aren’t due to something else, perhaps something completely different. The phenomenon we at first imagine may not even exist. Instead, such verification most inventors and users of measures of intelligence … have simply constructed the source of differences in sensed intelligence as an underlying entity or force, rather in the way that children and naïve adults perceive hotness as a substance, or attribute the motion of objects to a fictitious impetus. What we have in cases like temperature, of course, are collateral criteria and measures that validate the theory, and thus the original measures. Without these, the assumed entity remains a fiction. This proved to be the case with impetus, and with many other naïve conceptions of nature, such as phlogiston (thought to account for differences in health and disease). How much greater such fictions are likely to be unobserved, dynamic and socially judged concepts like intelligence.

Richardson (1991: 32-35) then goes on to critique many of the old IQ tests, in that they had no way of being construct valid, and that the manuals did not even discuss the validity of the test—it was just assumed.

If we do not know what exactly is being measured when test constructors make and administer these tests, then how can we logically state that “IQ tests test intelligence”? Even Arthur Jensen admitted that psychometricians can create any type of distribution they please (1980: 71); he tacitly admits that tests are devised through the selection and deselection of items on IQ tests that correspond to the test constructors preconceived notions on what “intelligence” is. This, again, is even admitted by Jensen (1980: 147-148) who writes “The items must simply emerge arbitrarily from the heads of test constructors.

We know, to build on Richardson’s temperature example, that we know exactly is what being measured when we look at the amount of mercury in a thermometer. That is, the concept of “temperature” and the instrument to measure it (the thermometer) were verified independently, without circular reliance on the thermometer itself (see Hasok Chang’s 2007 book Inventing Temperature). IQ tests, on the other hand, are, supposedly, “validated” through measures of job performance and correlations with other, previous tests assumed to be (construct) valid—but they were, of course, just assumed to be valid, it was never shown.

For another example (as I’ve shown with IQ many times) of a psychological construct that is not valid is ASD (autism spectrum disorder). Waterhouse, London, and Gilliberg (2016) write that “14 groups of findings reviewed in this paper that together argue that ASD lacks neurobiological and construct validity. No unitary ASD brain impairment or replicated unitary model of ASD brain impairment exists.” That a construct is valid—that is, it tests what it purports to, is of utmost importance to test measurement. Without it, we don’t know if we’re measuring something else completely different from what we hope—or purport—to.

There is another problem: the fact that, for one of the most-used IQ tests that there is no underlying theory of item selection, as seen in John Raven’s personal notes (see Carpenter, Just, and Shell, 1990). Items on the Raven were selected based on Raven’s intuition, and not any formal theory—the same can be said about, of course, modern-day IQ tests. Carpenter, Just, and Shell (1990) write that John Raven “used his intuition and clinical experience to rank order the difficulty of the six problem types . . . without regard to any underlying processing theory.”

These preconceived notions on what “intelligence” is, though, fail without (1) a theory of what intelligence is (which, as admitted by Ian Deary (2001), there is no theory of human intelligence like the way physics has theories); and (2) what ultimately is termed “construct validity”—that a test measures what it purports to. There are a few kinds of validity: and what IQ-ists claim the most is that IQ tests have predictive validity—that is, they can predict an individual’s outcome in life, and job performance (it is claimed). However, “intelligence” is “a descriptive measure, not an explanatory concept … [so] measures of intelligence level have little or no predictive value” (Howe, 1988).

Howe (1997: ix) also tells us that “Intelligence is … an outcome … not a cause. … Even the most confidently stated assertions about intelligence are often wrong, and the inferences that people have drawn from those assertions are unjustified.

The correlation between IQ and school performance, according to Richardson (1991: 34)may be a necessary aspect of the validity of tests, but is not a sufficient one. Such evidence, as already mentioned, requires a clear connection between a theory (a model of intelligence), and the values on the measure.” But, as Richardson (2017: 85) notes:

… it should come as no surprise that performance on them [IQ tests] is associated with school performance. As Robert L. Thorndike and Elizabeth P. Hagen explained in their leading textbook, Educational and Psychological Measurement, “From the very way in which the tests were assembled [such correlation] could hardly be otherwise.”

Gottfredson (2009) claims that the construct validity argument against IQ is “fallacious”, noting it as one of her “fallacies” on intelligence testing (one of her “fallacies” was the “interactionism fallacy”, which I have previously discussed). However, unfortunately for Gottfredson (2009), “the phenomena that testers aim to capture” are built into the test and, as noted here numerous times, preconceived by the constructors of the test. So, Gottfredson’s (2009) claim fails.

Such kinds of construction, too, come into the claim of a “normal distribution.” Just like with preconceptions of who is or is not “intelligent” on the basis of preconceived notions, the normal distribution, too, is an artifact of test construction, along the selection and deselection of items to conform with the test constructors’ presuppositions; the “bell curve” of IQ is created by the presuppositions that the test constructors have about people and society (Simon, 1997).

Charles Spearman, in the early 1900s, claims to have found a “general factor” that explains correlations between different tests. This positive manifold he termed “g” for “general intelligence.” Spearman stated “The (g) factor was taken, pending further information, to consist in something of the nature of an ‘energy’ or ‘power’…” (quoted in Richardson, 1991: 38). The refutation of “g” is a simple, logical, one: While a correlation between performances “may be a necessary requirement for a general factor … it is not a sufficient one.” This is because “it is quite possible for quite independent factors to produce a hierarchy of correlations without the existence of any underlying ‘general’ factor (Fancer, 1985a; Richardson and Bynner, 1984)” (Richardson, 1991: 38). The fact of the matter is, Spearman’s “g” has been refuted for decades (and was shown to be reified by Gould (1981), and further defenses of his concepts on “general intelligence”, like by Jensen (1998) have been refuted, most forcefully by Peter Schonemann. Though, “g” is something built into the test by way of test construction (Richardson, 2002).

Castles (2013: 93) notes that “Spearman did not simply discover g lurking in his data. Instead, he chose one peculiar interpretation of the relationships to demonstrate something in which he already believed—unitary, biologically based intelligence.”

So what explains differences in “g”? The same test construction noted above along with differences in social class, due to stress, self-confidence, test preparedness and other factors correlated with social class, termed the “sociocognitive-affective nexus” (Richardson, 2002).

Constance Hilliard, in her book Straightening the Bell Curve (Hilliard, 2012), notes that there were differences in IQ between rural and urban white South Africans. She notes that differences between those who spoke Afrikaans and those who spoke another language were completely removed through test construction (Hilliard, 2012: 116). Hilliard (2012) notes that if the tests that the constructors formulate don’t agree with their preconceived notions, they are then thrown out:

If the individuals who were supposed to come out on top didn’t score highly or, conversely, if the individuals who were assumed would be at the bottom of the scores didn’t end up there, then the test designers scrapped the test.

Sex differences in “intelligence” (IQ) have been the subject of some debate in the early-to-mid-1900s. Test constructors debated amongst themselves what to do about such differences between the sexes. Hilliard (2012) quotes Harrington (1984; in Perspectives on Bias in Mental Testing) who writes about normalizing test scores between men and women:

It was decided [by IQ test writers] a priori that the distribution of intelligence-test scores would be normal with a mean (X=100) and a standard deviation (SD=15), also that both sexes would have the same mean and distribution. To ensure the absence of sex differences, it was arranged to discard items on which the sexes differed. Then, if not enough items remained, when discarded items were reintroduced, they were balanced, i.e., for every item favoring males, another one favoring females was also introduced.

While Richardson (1998: 114) notes that test constructors had two choices when looking at sex differences in the items they administered to the sexes:

One who would construct a test for intellectual capacity has two possible methods of handling the problem of sex differences.
1 He may assume that all the sex differences yielded by his test items are about equally indicative of sex differences in native ability.
2 He may proceed on the hypothesis that large sex differences on items of the Binet type are likely to be factitious in the sense that they reflect sex differences in experience or training. To the extent that this assumption is valid, he will be justified in eliminating from his battery test items which yield large sex differences.
The authors of the New Revision have chosen the second of these alternatives and sought to avoid using test items showing large differences in percents passing. (McNemar 1942:56)

Change “sex differences” to “race” or “social class” differences and we can, too, change the distribution of the curve, along with notions of who is or is not “intelligent.” Previously low scorers can, by way of test construction, become high scorers, vice-versa for high scorers being made into low scorers. There is no logical—or empirical—justification for the inclusion of specific items on whatever IQ test is in question. That is, to put it another way, the inclusion of items on a test is subjective, which comes down to the test designers’ preconceived notions, and not an objective measure of what types of items should be on the test—as Raven stated, there is no type of underlying theory for the inclusion of items in the test, it is based on “intuition” (which is the same thing that modern-day test constructors do). These two quotes from IQ-ists in the early 20th century are paramount in the attack on the validity of IQ tests—and the causes for differences in scores between groups.

He and van de Vijver (2012: 7) write that “An item is biased when it has a different psychological meaning across cultures. More precisely, an item of a scale (e.g., measuring anxiety) is said to be biased if persons with the same trait, but coming from different cultures, are not equally likely to endorse the item (Van de Vijver & Leung, 1997).” Indeed, Reynolds and Suzuki (2012: 83) write that “Item bias due to“:

… “poor item translation, ambiguities in the original item, low familiarity/appropriateness of the item content in certain cultures, or influence of culture specifics such as nuisance factors or connotations associated with the item wording” (p. 127) (van de Vijver and Tanzer, 2004)

Drame and Ferguson (2017) note that their “Results indicate that use of the Ravens may substantially underestimate the intelligence of children in Mali” while the cause may be due to the fact that:

European and North American children may spend more time with play tasks such as jigsaw puzzles or connect the dots that have similarities with the Ravens and, thus, train on similar tasks more than do African children. If African children spend less time on similar tasks, they would have fewer opportunities to train for the Ravens (however unintentionally) reflecting in poorer scores. In this sense, verbal ability need not be the only pitfall in selecting culturally sensitive IQ testing approaches. Thus, differences in Ravens scores may be a cultural artifact rather than an indication of true intelligence differences. [Similar arguments can be found in Richardson, 2002: 291-293]

The same was also found by Dutton et al (2017) who write that “It is argued that the undeveloped nature of South Sudan means that a test based around shapes and analytic thinking is unsuitable. It is likely to heavily under-estimate their average intelligence.” So if the Raven has these problems cross-culturally (country), then it SHOULD have such biases within, say, America.

It is also true that the types of items on IQ tests are not as complex as everyday life (see Richardson and Norgate, 2014). Types of questions on IQ tests are, in effect, ones of middle-class knowledge and skills and, knowing how IQ tests are structured will make this claim clear (along with knowing the types of items that eventually make it onto the particular IQ test itself). Richardson (2002) has a few questions on modern-day IQ tests whereas Castles (2013), too, has a few questions from the Stanford-Binet. This, of course, is due to the social class of the test constructors. Some examples of some questions can be seen here:

‘What is the boiling point of water?’ ‘Who wrote Hamlet?’ ‘In what continent is Egypt?’ (Richardson, 2002: 289)

and

‘When anyone has offended you and asks you to excuse him—what ought you do?’ ‘What is the difference between esteem and affection?’ [this is from the Binet Scales, but “It is interesting to note that similar items are still found on most modern intelligence tests” (Castles, 2013).]]

Castles (2013: 150) further notes made-up examples of what is on the WAIS (since she cannot legally give questions away since she is a licensed psychologist), and she writes:

One section of the WAIS-III, for example, consists of arithmetic problems that the respondent must solve in his or her head. Others require test-takers to define a series of vocabulary words (many of which would be familiar only to skilled-readers), to answer school-related factual questions (e.g., “Who was the first president of the United States?” or “Who wrote the Canterbury Tales?”), and to recognize and endorse common cultural norms and values (e.g., “What should you do it a sale clerk accidentally gives you too much change?” or “Why does our Constitution call for division of powers?”). True, respondents are also given a few opportunities to solve novel problems (e.g., copying a series of abstract designs with colored blocks). But even these supposedly culture-fair items require an understanding of social conventions, familiarity with objects specific to American culture, and/or experience working with geometric shapes or symbols.

All of these factors coalesce into forming the claim—and the argument—that IQ tests are one of middle-class knowledge and skills. The thing is, contrary to the claims of IQ-ists, there is no such thing as a culture-free IQ test. Richardson (2002: 293) notes that “Since all human cognition takes place through the medium of cultural/psychological tools, the very idea of a culture-free test is, as Cole (1999) notes, ‘a contradiction in terms . . . by its very nature, IQ testing is culture bound’ (p. 646). Individuals are simply more or less prepared for dealing with the cognitive and linguistic structures built in to the particular items.

Cole (1981) notes that “that the notion of a culture free IQ test is an absurdity” because “all higher psychological processes are shaped by our experiences and these experiences are culturally organized” (this is a point that Richardson has driven home for decades) while also—rightly—stating that “IQ tests sample school activities, and therefore, indirectly, valued social activities, in our culture.

One of the last stands for the IQ-ist is to claim that IQ tests are useful for identifying at-risk individuals for learning disabilities (as Binet originally created the first IQ tests for). However, it is noted that IQ tests are not necessary—nor sufficient—for the identification of those with learning disabilities. Siegal (1989) states that “On logical and empirical grounds, IQ test scores are not necessary for the definition of learning disabilities.

When Goddard brought the first IQ tests to America and translated them into English from French is when the IQ testing conglomerate really took off (see Zenderland, 1998 for a review). These tests were used to justify current social ranks. As Richardson (1991: 44) notes, “The measurement of intelligence in the twentieth century arose partly out of attempts to ‘prove’ or justify a particular world view, and partly for purposes of screening and social selection. It is hardly surprising that its subsequent fate has been one of uncertainty and controversy, nor that it has raised so many social and political issues (see, for example, Joynson 1989 for discussion of such issues).” So, what actual attempts at validation did the constructors of such tests need in the 20th century when they knew full-well what they wanted to show and, unsurprisingly, they observed it (since it was already going to happen since they construct the test to be that way)?

The conceptual arguments just given here point to a few things:

(1) IQ tests are not construct valid because there is no theory of intelligence, nor is there an underlying theory which relates differences in IQ (the unseen function) to, for example, a physiological variable. (See Uttal, 2012; 2014 for arguments against fMRI studies that purport to show differences in physiological variables cognition.)

(2) The fact that items on the tests are biased against certain classes/cultures; this obviously matters since, as noted above, there is no theory for the inclusion of items, it comes down to the subjective choice of the test designers, as noted by Jensen.

(3) ‘g’ is a reified mathematical abstraction; Spearman “discovered” nothing, he just chose the interpretation that, of course, went with his preconceived notion.

(4) The fact that sex differences in IQ scores were seen as a problem and, through item analysis, made to go away. This tells us that we can do the same for class/race differences in intelligence. Score differences are a function of test construction.

(5) The fact that the Raven has been shown to be biased in two African countries lends credence to the claims here.

So this then brings us to the ultimate claim of this article: IQ tests don’t test intelligence; they test middle-class knowledge and skills. Therefore, the scores on IQ tests are not that of intelligence, but of an index of one’s cultural knowledge of the middle class and its knowledge structure. This, IQ scores are, in actuality, “middle-class knowledge and skills” scores. So, contra Jensen (1980), there is bias in mental testing due to the items chosen for inclusion on the test (we have admission that score variances and distributions can change from IQ-ists themselves)

Chopsticks Genes and Population Stratification

1200 words

Why do some groups of people use chopsticks and others do not? Years back, created a thought experiment. So he found a few hundred students from a university and gathered DNA samples from their cheeks which were then mapped for candidate genes associated with chopstick use. Come to find out, one of the associated genetic markers was associated with chopstick use—accounting for 50 percent of the variation in the trait (Hamer and Sirota, 2000). The effect even replicated many times and was highly significant: but it was biologically meaningless.

One may look at East Asians and say “Why do they use chopsticks” or “Why are they so good at using them while Americans aren’t?” and come to such ridiculous studies such as the one described above. They may even find an association between the trait/behavior and a genetic marker. They may even find that it replicates and is a significant hit. But, it can all be for naught, since population stratification reared its head. Population stratification “refers to differences in allele frequencies between cases and controls due to systematic differences in ancestry rather than association of genes with disease” (Freedman et al, 2004). It “is a potential cause of false associations in genetic association studies” (Oetjens et al, 2016).

Such population stratification in the chopsticks gene study described above should have been anticipated since they studied two different populations. Kaplan (2000: 67-68) described this well:

A similar argument, bu the way, holds true for molecular studies. Basically, it is easy to mistake mere statistical associations for a causal connection if one is not careful to properly partition one’s samples. Hamer and Copeland develop and amusing example of some hypothetical, badly misguided researchers searching for the “successful use of selected hand instruments” (SUSHI) gene (hypothesized to be associated with chopstick usage) between residents in Tokyo and Indianapolis. Hamer and Copeland note that while you would be almost certain to find a gene “associated with chopstick usage” if you did this, the design of such a hypothetical study would be badly flawed. What would be likely to happen here is that a genetic marker associated with the heterogeneity of the group involved (Japanese versus Caucasian) would be found, and the heterogeneity of the group involved would independently account for the differences in the trait; in this case, there is a cultural tendency for more people who grow up in Japan than people who grow up in Indianapolis to learn how to use chopsticks. That is, growing up in Japan is the causally important factor in using chopsticks; having a certain genetic marker is only associated with chopstick use in a statistical way, and only because those people who grow up in Japan are also more likely to have the marker than those who grew up in Indianapolis. The genetic marker is in no way causally related to chopstick use! That the marker ends up associated with chopstick use is therefore just an accident of design (Hamer and Copeland, 1998, 43; Bailey 1997 develops a similar example).

In this way, most—if not all—of the results of genome-wide association studies (GWASs) can be accounted for by population stratification. Hamer and Sirota (2000) is a warning to psychiatric geneticists to not be quick to ascribe function and causation to hits on certain genes from association studies (of which GWASs are).

Many studies, for example, Sniekers et al (2017), Savage et al (2018) purport to “account for” less than 10 percent of the variance in a trait, like “intelligence” (derived from non-construct valid IQ tests). Other GWA studies purport to show genes that affect testosterone production and that those who have a certain variant are more likely to have low testosterone (Ohlsson et al, 2011). Population stratification can have an effect here in these studies, too. GWASs; they give rise to spurious correlations that arise due to population structure—which is what GWASs are actually measuring, they are measuring social class, and not a “trait” (Richardson, 2017b; Richardson and Jones, 2019). Note that correcting for socioeconomic status (SES) fails, as the two are distinct (Richardson, 2002). (Note that GWASs lead to PGSs, which are, of course, flawed too.)

Such papers presume that correlations are causes and that interactions between genes and environment either don’t exist or are irrelevant (see Gottfredson, 2009 and my reply). Both of these claims are false. Correlations can, of course, lead to figuring out causes, but, like with the chopstick example above, attributing causation to things that are even “replicable” and “strongly significant” will still lead to false positives due to that same population stratification. Of course, GWAS and similar studies are attempting to account for the heriatbility estimates gleaned from twin, family, and adoption studies. Though, the assumptions used in these kinds of studies are shown to be false and, therefore, heritability estimates are highly exaggerated (and flawed) which lead to “looking for genes” that aren’t there (Charney, 2012; Joseph et al, 2016; Richardson, 2017a).

Richardson’s (2017b) argument is simple: (1) there is genetic stratification in human populations which will correlate with social class; (2) since there is genetic stratification in human populations which will correlate with social class, the genetic stratification will be associated with the “cognitive” variation; (3) if (1) and (2) then what GWA studies are finding are not “genetic differences” between groups in terms of “intelligence” (as shown by “IQ tests”), but population stratification between social classes. Population stratification still persists even in “homogeneous” populations (see references in Richardson and Jones, 2019), and so, the “corrections for” population stratification are anything but.

So what accounts for the small pittance of “variance explained” in GWASs and other similar association studies (Sniekers et al, 2017 “explained” less than 5 percent of variance in IQ)? Population stratification—specifically it is capturing genetic differences that occurred through migration. GWA studies use huge samples in order to find the genetic signals of the genes of small effect that underline the complex trait that is being studied. Take what Noble (2018) says:

As with the results of GWAS (genome-wide association studies) generally, the associations at the genome sequence level are remarkably weak and, with the exception of certain rare genetic diseases, may even be meaningless (1321). The reason is that if you gather a sufficiently large data set, it is a mathematical necessity that you will find correlations, even if the data set was generated randomly so that the correlations must be spurious. The bigger the data set, the more spurious correlations will be found (3).

Calude and Longo (2016; emphasis theirs) “prove that very large databases have to contain arbitrary correlations. These correlations appear only due to the size, not the nature, of data. They can be found in “randomly” generated, large enough databases, which — as we will prove — implies that most correlations are spurious.”

So why should we take association studies seriously when they fall prey to the problem of population stratification (measuring differences between social classes and other populations) along with the fact that big datasets lead to spurious correlations? I fail to think of a good reason why we should take these studies seriously. The chopsticks gene example perfectly illustrates the current problems we have with GWASs for complex traits: we are just seeing what is due to social—and other—stratification between populations and not any “genetic” differences in the trait that is being looked at.

The “Interactionism Fallacy”

2350 words

A fallacy is an error in reasoning that makes an argument invalid. The “interactionism fallacy” is the fallacy—coined by Gottfredson (2009)—that since genes and environment interact, that heritability estimates are not useful—especially for humans (they are for nonhuman animals where environments can be fully controlled; see Schonemann, 1997; Moore and Shenk, 2016). There are many reasons why this ‘fallacy’ is anything but a fallacy; it is a simple truism: genes and environment (along with other developmental products) interact to ‘construct’ the organism (what Oyama, 2000 terms ‘constructive interactionism—“whereby each combination of genes and environmental influences simultaneously interacts to produce a unique result“). The causal parity thesis (CPT) is the thesis that genes/DNA play an important role in development, but so do other variables, so there is no reason to privilege genes/DNA above other developmental variables (see Noble, 2012 for a similar approach). Genes are not special developmental resources and so, nor are they more important than other developmental resources. So the thesis is that genes and other developmental resources are developmentally ‘on par’.

Genes need the environment. Without the environment, genes would not be expressed. Behavior geneticists claim to be able to partition genes from environment—nature from nurture—on the basis of heritability estimates, mostly gleaned from twin and adoption studies. However, the method is flawed: since genes interact with the environment and other genes, how would it be possible to neatly partition the effects of genes from the effects of the environment? Behavior geneticists claim that we can partition these two variables. Behavior geneticists—and others—cite the “Interactionism fallacy”, the fallacy that since genes interact with the environment that heritability estimates are useless. This “fallacy”, though, confuses the issue.

Behavior geneticists claim to show how genes and the environment affect the ontogeny of traits in humans with twin and adoption studies (though these methods are highly flawed). The purpose of this “fallacy” is to disregard what developmental systems theorists claim about the interaction of nature and nurture—genes and environment.

Gottfredson (2009) coins the “interactionism fallacy”, which is “an irrelevant truth [which is] that an organism’s development requires genes and environment to act in concert” and the “two forces are … constantly interacting” whereas “Development is their mutual product.” Gottfredson also states that “heritability … refers to the percentage of variation in … the phenotype, which has been traced to genetic variation within a particular population.” (She also makes the false claim that “One’s genome is fixed at birth“; though this is false, see epigenetics/methylation studies.) Heritability estimates, according to Phillip Kitcher are “‘irrelevant’ and the fact that behavior geneticists persist
in using them is ‘an unfortunate tic from which they cannot free themselves’ (Kitcher,
2001: 413)” (quoted in Griffiths, 2002).

Gottfredson is engaging in developmental denialism. Developmental denialismoccurs when heritability is treated as a causal mechanism governing the developmental reoccurrence of traits across generations in individuals.” Gottfredson, with her “interactionism fallacy” is denying organismal development by attempting to partition genes from environment. As Rose (2006) notes, “Heritability estimates are attempts to impose a simplistic and reified dichotomy (nature/nurture) on non-dichotomous processes.” The nature vs nurture argument is over and neither has won—contra Plomin’s take—since they interact.

Gottfredson seems confused, since this point was debated by Plomin and Oyama back in the 80s (Plomin’s review of Oyama’s book The Ontogeny of Information; see Oyama, 1987, 1988; Plomin, 1988a, b). In any case, it is true that development requires genes to interact. But Gottfredson is talking about the concept of heritability—the attempt to partition genes and environment through twin, adoption and family studies (which have a whole slew of problems). For example, Moore and Shenk (2016: 6) write:

Heritability statistics do remain useful in some limited circumstances, including selective breeding programs in which developmental environments can be strictly controlled. But in environments that are not controlled, these statistics do not tell us much.

Susan Oyama writes in The Ontogeny of Information (2000, pg 67):

Heritability coefficients, in any case, because they refer not only to variation in genotype but to everything that varied (was passed on) with it, only beg the question of what is passed on in evolution. All too often heritability estimates obtained in one setting are used to infer something about an evolutionary process that occurred under conditions, and with respect to a gene pool, about which little is known. Nor do such estimates tell us anything about development.

Characters are produced by the interaction of nongenetic and genetic factors. The biological flaw, as Moore and Shenk note, throw a wrench into the claims of Gottfredson and other behavior geneticists. Phenotypes are ALWAYS due to genetic and nongenetic factors interacting. So the two flaws of heritability—the environmental and biological flaw (Moore and Shenk, 2016)—come together to “interact” to refute such simplistic claims that genes and environment—nature and nurture—can be separated.

For instance, as Moore (2016) writes, though “twin study methods are among the most powerful tools available to quantitative behavioral geneticists (i.e., the researchers who took up Galton’s goal of disentangling nature and nurture), they are not satisfactory tools for studying phenotype development because they do not actually explore biological processes.” (See also Richardson, 2012.) This is because twin studies ignore biological/developmental processes that lead to phenotypes.

Gamma and Rosenstock (2017) write that the concept of heritability that behavioral geneticists use is “is a generally useless quantity” while “the behavioral genetic dichotomy of genes vs environment is fundamentally misguided.” This brings us back to the CPT; there is causal parity to all processes/interactants that form the organism and its traits, thus the concept of heritability that behavioral geneticists employ is a useless measure. Oyama, Griffiths, and Gray (2001: 3) write:

These often overlooked similarities form part of the evidence for DST’s claim of causal parity between genes and other factors of development. The “parity thesis” (Griffiths and Knight 1998) does not imply that there is no difference between the particulars of the causal roles of genes and factors such as endosymbionts or imprinting events. It does assert that such differences do not justify building theories of development and evolution around a distinction between what genes do and what every other causal factor does.

Behavior geneticists’ endeavor, though, is futile. Aaron Panofsky (2016: 167) writes that “Heritability estimates do not help identify particular genes or ascertain their functions in development or physiology, and thus, by this way of thinking, they yield no causal information.” (Also see Panofsky, 2014; Misbehaving Science: Controversy and the Development of Behavior Genetics.) So, the behavioral genetic method of partitioning genes and environment does not—and can not—show causation for trait ontogeny.

Now, while people like Gottfredson and others may deny it, they are genetic determinists. Genetic determinism, as defined by Griffiths (2002) is “the idea that many significant human characteristics are rendered inevitable by the presence of certain genes.” Using this definition, many behavior geneticists and their sympathizers have argued that certain traits are “inevitable” due to the presence of certain genes. Genetic determinism is literally the idea that genes “determine” aspects of characters and traits, though it has been known for decades that it is false.

Now we can take a look at Brian Boutwell’s article Not Everything Is An Interaction. Boutwell writes:

Albert Einstein was a brilliant man. Whether his famous equation of E=mc2 means much to you or not, I think we can all concur on the intellectual prowess—and stunning hair—of Einstein. But where did his brilliance come from? Environment? Perhaps his parents fed him lots of fish (it’s supposed to be brain food, after all). Genetics? Surely Albert hit some sort of genetic lottery—oh that we should all be so lucky. Or does the answer reside in some combination of the two? How very enlightened: both genes and environment interact and intertwine to yield everything from the genius of Einstein to the comedic talent of Lewis Black. Surely, you cannot tease their impact apart; DNA and experience are hopelessly interlocked. Except, they’re not. Believing that they are is wrong; it’s a misleading mental shortcut that has largely sown confusion in the public about human development, and thus it needs to be retired.

[…]

Most traits are the product of genetic and environmental influence, but the fact that both genes and environment matter does not mean that they interact with one another. Don’t be lured by the appeal of “interactions.” Important as they might be from time to time, and from trait to trait, not everything is an interaction. In fact, many things likely are not.

I don’t even know where to begin here. Boutwell, like Gottfredson, is confused. The only thing that needs to be retired because it “has largely sown confusion in the public about human development” is, ironically, the concept of heritability (Moore and Shenk, 2016)! I have no idea why Boutwell claimed that it’s false that “DNA and experience [environment] are hopelessly interlocked.” This is because, as Schneider (2007) notes, “the very concept of a gene requires an environment.” Since the concept of the gene requires the environment, how can we disentangle them into neat percentages like behavior geneticists claim to do? That’s right: we can’t. Do be lured by the appeal of interactions; all biological and nonbiological stuff constantly interacts with one another.

Boutwell’s claims are nonsense. It would be worth it to quote Richard Lewontin’s forward in the 2000 2nd edition of Susan Oyama’s The Ontogeny of Information (emphasis Lewontin’s):

Nor can we partition variation quantitatively, ascribing some fraction of variation to genetic differences and the remainder to environmental variation. Every organism is the unique consequence of the reading of its DNA in some temporal sequence of environments and subject to random cellular events that arise because of the very small number of molecules in each cell. While we may calculate statistically an average difference between carriers of one genotype and another, such average differences are abstract constructs and must not be reified with separable concrete effects of genes in isolation from the environment in which the genes are read. In the first edition of The Ontogeny of Information Oyama characterized her construal of the causal relation between genes and environment as interactionist. That is, each unique combination of genes and environment produces a unique and a priori unpredictable outcome of development. The usual interactionist view is that there are separable genetic and environmental causes, but the effects of these causes acting in combination are unique to the particular combination. But this claim of ontogenetically independent status of the causes as causes, aside from their interaction in the effects produced, contradicts Oyama’s central analysis of the ontogeny of information. There are no “gene actions” outside environments, and no “environmental actions” can occur in the absence of genes. The very status of environment as a contributing cause to the nature of an organism depends on the existence of a developing organism. Without organisms there may be a physical world, but there are no environments. In like the manner no organisms exist in the abstract without environments, although there may be naked DNA molecules lying in the dust. Organisms are the nexus of external circumstances and DNA molecules that make these physical circumstances into causes of development in the first place. They become causes only at their nexus, and they cannot exist as causes except in their simultaneous action. That is the essence of Oyama’s claim that information comes into existence only in the process of ontogeny. (Oyama, 2000: 16)

There is an “interactionist consensus” (see Oyama, Griffiths, and Grey, 2001; What is Developmental Systems Theory? pg 1-13): the organism and the suite of traits it has is due to the interaction of genetic/environmental/epigenetic etc. resources at every stage of development. Therefore, for organismal development to be successful, it always requires the interaction of genes, environment, epigenetic processes, and interactions between everything that is used to ‘construct’ the organism and the traits it has. Thus “it makes no sense to ask if a particular trait is genetic or environmental in origin. Understanding how a trait develops is not a matter of finding out whether a particular gene or a particular environment causes the trait; rather, it is a matter of understanding how the various resources available in the production of the trait interact over time” (Kaplan, 2006).

Lastly, I will shortly comment on Sesardic’s (2005: chapter 2) critiques on developmental systems theorists and their critique of heritability and the concept of interactionism. Sesardic argues in the chapter that interaction between genes and environment, nature and nurture, does not undermine heritability estimates (the nature and nurture partition). Philosopher of science Helen Longino argues in her book Studying Human Behavior (2013):

By framing the debate in terms of nature versus nurture and as though one of these must be correct, Sesardic is committed to both downplaying the possible contributions of environmentally oriented research and to relying on a highly dubious (at any rate, nonmethodological) empirical claim.

In sum, the “interactionist fallacy” (coined by Gottfredson) is not a ‘fallacy’ (error in reasoning) at all. For, as Oyama writes in Evolution’s Eye: A Systems View of the Biology-Culture DivideA not uncommon reaction to DST is, ‘‘That’s completely crazy, and besides, I already knew it” (pg 195). This is exactly what Gottfredson (2009) states, that she “already knew” that there is an interaction between nature and nurture; but she goes on to deny arguments from Oyama, Griffiths, Stotz, Moore, and others on the uselessness of heritability estimates along with the claim that nature and nurture cannot be neatly partitioned into percentages as they are constantly interacting. Causal parity between genes and other developmental resources, too, upends the claim that heritability estimates for any trait make sense (not least for how heritability estimates are gleaned for humans—mostly twin, family, and adoption studies). Developmental denialism—what Gottfredson and others often engage in—runs rampant in the “behavioral genetic” sphere; and Oyama, Griffiths, Stotz, and others show how we should not deny development and we should discard with these estimates for human traits.

Heritability estimates imply that there is a “nature vs nurture” when it is “nature and nurture” which are constantly interacting—and, due to this, we should discard with these estimates due to the interaction of numerous developmental resources; it does not make sense to partition an interacting, self-organizing developmental system. Claims from behavior geneticists—that genes and environment can be separated—are clearly false.

Men Are Stronger Than Women

1200 words

The claim that “Men are stronger than women” does not need to be said—it is obvious through observation that men are stronger than women. To my (non-)surprise, I saw someone on Twitter state:

“I keep hearing that the sex basis of patriarchy is inevitable because men are (on average) stronger. Notwithstanding that part of this literally results from women in all stages of life being denied access to and discourage from physical activity, there’s other stuff to note.”

To which I replied:

“I don’t follow – are you claiming that if women were encouraged to be physically active that women (the population) can be anywhere *near* men’s (the population) strength level?”

I then got told to “Fuck off,” because I’m a “racist” (due to the handle I use and my views on the reality of race). In any case, while it is true that part of this difference does, in part, stem from cultural differences (think of women wanting the “toned” look and not wanting to get “big and bulky”—as if it happens overnight) and not wanting to lift heavy weights because they think they will become cartoonish.

Here’s the thing though: Men have about 61 percent more muscle mass than women (which is attributed to higher levels of testosterone); most of the muscle mass difference is allocated to the upper body—men have about 75 percent more arm muscle mass than women which accounts for 90 percent greater upper body strength in men. Men also have about 50 percent more muscle mass than women, while this higher percentage of muscle mass is then related to men’s 65 percent greater lower body strength (see references in Lassek and Gaulin, 2009: 322).

Men have around 24 pounds of skeletal muscle mass compared to women, though in this study, women were about 40 percent weaker in the upper body and 33 percent weaker in the lower body (Janssen et al, 2000). Miller et al (1993) found that women had a 45 percent smaller cross-section area in the brachii, 45 in the elbow flexion, 30 percent in the vastus lateralis, and 25 percent smaller CSA in the knee extensors, as I wrote in Muscular Strength by Gender and Race, where I concluded:

The cause for less upper-body strength in women is due the distribution of women’s lean tissue being smaller.

Men have larger fibers, which in my opinion is a large part of the reason for men’s strength advantage over women. Now, even if women were “discouraged” from physical activity, this would be a problem for their bone density. Our bones are porous, and so, by doing a lot of activity, we can strengthen our bones (see e.g., Fausto-Sterling, 2005). Bishop, Cureton, and Collins (1987) show that the sex difference in strength in close-to-equally-trained men and women “is almost entirely accounted for by the difference in muscle size.” Which lends credence to my claim I made above.

Lindle et al (1997) conclude that:

… the results of this study indicate that Con strength levels begin to decline in the fourth rather than in the fifth decade, as was previously reported. Contrary to previous reports, there is no preservation of Ecc compared with Con strength in men or women with advancing age. Nevertheless, the decline in Ecc strength with age appears to start later in women than in men and later than Con strength did in both sexes. In a small subgroup of subjects, there appears to be a greater ability to store and utilize elastic energy in older women. This finding needs to be confirmed by using a larger sample size. Muscle quality declines with age in both men and women when Con peak torque is used, but declines only in men when Ecc peak torque is used. [“Con” and “Ecc” strength refer to concentric and eccentric actions]

Women are shorter than men and have less fat-free muscle mass than men. Women also have a weaker grip (even when matched for height and weight, men had higher levels of lean mass compared to women (92 and 79 percent respectively; Nieves et al, 2009). So men had greater bone mineral density (BMD) and bone mineral content (BMC) compared to women. Now do some quick thinking—do you think that one with weaker bones could be stronger than someone with stronger bones? If person A had higher levels of BMC and BMD compared to person B, who do you think would be stronger and have the ability to do whatever strength test the best—the one with the weaker or stronger muscles? Quite obviously, the stronger one’s bones are the more weight they can bare on them. So if one has weak bones (low BMC/BMD) and they put a heavy load on their back, while they’re doing the lift their bones could snap.

Alswat (2017) reviewed the literature on bone density between men and women and found that men had higher BMD in the hip and higher BMC in the lower spine. Women also had bone fractures earlier than men. Some of this is no doubt cultural, as explained above. However, even if we had a boy and a girl locked in a room for their whole lives and they did the same exact things, ate the same food, and lifted the same weights, I would bet my freedom that there still would be a large difference between the two, skewing where we know it would skew. Women are more likely to suffer from osteoporosis than are men (Sözen, Özışık, and Başaran 2016).

So if women have weaker bones compared to men, then how could they possibly be stronger? Even if men and women had the same kind of physical activity down to the tee, could you imagine women being stronger than men? I couldn’t—but that’s because I have more than a basic understanding of anatomy and physiology and what that means for differences in strength—or running—between men and women.

I don’t doubt that there are cultural reasons that account for the large differences in strength between men and women—I do doubt, though, that the gap can be meaningfully closed. Yes, biology interacts with culture. So the developmental variables that coalesce to make men “Men” and those that coalesce to make women “Women” converge in creating the stark differences in phenotype between the sexes which then explains how the sex differences between the sexes manifest itself.

Differences in bone strength between men and women, along with distribution of lean tissue, differences in lean mass, and differences in muscle size explain the disparity in muscular strength between men and women. You can even imagine a man and woman of similar height and weight and they would, of course, look different. This is due to differences in hormones—the two main players being testosterone and estrogen (see Lang, 2011).

So yes, part of the difference in strength between men and women are rooted in culture and how we view women who strength train (way more women should strength train, as a matter of fact), though I find it hard to believe that even if the “cultural stigma” of the women who lifts heavy weights at the gym disappeared overnight, that women would be stronger than men. Differences in strength exist between men and women and this difference exists due to the complex relationship between biology and culture—nature and nurture (which cannot be disentangled).