NotPoliticallyCorrect

Home » Psychology

Category Archives: Psychology

An Argument for the Existence of Mind and Intentional Consciousness

2100 words

Consciousness and mind are uniquely human attributes. They allow us to reason and act intentionally. But what establishes the claim that consciousness and mind exist if they are immaterial? I have a few arguments for the claim, and I will also add a semi-updated version of the argument I made against animal mentality. So I will combine the argument that humans possess a mind (that is, that the mind is real and has referents) with the argument that nonhuman animals lack propositional attitudes and so they lack language and intentional states, and so they lack minds like humans. So the conclusion will be guaranteed—minds exist and only humans are in possession of them. First I will provide the argument that the mind exists since the existence of consciousness implies the existence of a nonphysical entity. After defending the premises, I will then shift to the argument that the problem of mental causation isn’t a problem for dualism. Then I will argue that only humans have minds and, due to the nature of intentionality and normativity of psychological traits, nonhuman animals lack minds like humans since they lack the prerequisites that entail having a mind.

An argument for the existence of mind

P1: Consciousness is a real, undeniable phenomenon that cannot be fully explained by physical or material processes.
P2: If consciousness cannot be fully explained by physical or material processes, then it must be a nonphysical phenomenon.
C1: Thus consciousness is a nonphysical phenomenon.
P3: The existence of a nonphysical phenomenon requires the existence of a nonphysical entity that can support or generate such phenomena.
C2: So the existence of consciousness implies the existence of a nonphysical entity.
C3: This nonphysical entity is the mind, so the mind exists.

Call this the argument from consciousness. Consciousness is a subjective, first-personal experience that everyone has and which cannot be reduced to physical or material processes. It is an experiential fact that each and every one of us is aware of their experiences and their thoughts, feelings, and perceptions; these cannot be explained by material or physical brain processes, due to the explanatory gap argument.

Premise 1: Of course we can study NCCs (neural correlates of consciousness), and this is what neuroscience does, but that’s not the same as studying the mind. The actualization of mind is not the same as that of digestion; gastroenterologists can study the physiological process of digestion, but neuroscientists can’t study the mind, since the mind isn’t merely brain physiology or brain/CNS (central nervous system activity) activity.

Premise 2: We can study material and physical processes, like brain states/physiology/CNS, using scientific inquiry. But if consciousness cannot be fully explained by physical or material processes, then it must be explained by the existence of a nonphysical entity and this also suggests that consciousness is an irreducible, nonphysical phenomenon.

Conclusion 1 then logically follows from P1 and P2.

Premise 3: Like P2, if consciousness is a nonphysical phenomenon, then it must be explained by a nonphysical entity since physical accounts cannot fully account for consciousness.

Conclusion 2 then logically follows from P3 and P4, and then conclusion 3 then follows from C2. So the argument is established on the grounds that consciousness exists and cannot be explained by material or physical processes, and if something exists which cannot be explained by physical or material processes then this implies the existence of a nonphysical entity that can support or generate such consciousness, and this nonphysical entity is referred to as MIND.

The limitations of material and physical explanations entail that the mind isn’t a physical process or a function of physical processes. The Knowledge Argument concludes that everything can’t be explained by physical or material processes, which would then strengthen the overall argument for the existence of an immaterial substance that explains consciousness. So there is a knowledge gap between physical explanations and subjective, individual explanations, and this is what the Knowledge Argument gets at. (Morch’s explanatory knowledge argument against physicalism establishes that some facts are nonphysical, which establishes the existence MIND.)

If the mind is immaterial, then how does it interact with the physical brain?

This question has been said to be a knockdown argument against dualism. If M and P are two different substances, how can they be said to interact? How, then, can the mind cause things to happen in the physical world? This is known as the problem of mental causation. How can mental events have any causal efficacy on physical events?we then need to establish between event causation and intentional causation (Lowe, 2001, 2009).

Intentional causation is mental causation (fact causation), and bodily causation is physical causation. Mental causation doesn’t reduce to physical causation, and this is because mental causation is intentional whereas physical causation isn’t. We have voluntary control over our actions, and so, we can intend to do things. So in Lowe’s (2006, 2010, 2012) non-Cartesian substance dualism (NCSD), persons or selves are distinct from their physical bodies and parts of their physical bodies. NCSD can better explain mental causation than the alternative materialist/physicalist theories, and so it can’t explain the intentionality of mental causation. The self is not the body, mental states aren’t physical states/processes (nor are they reducible to physical states/processes). The intentional content of mental states explains their uniqueness in contrast to physical states and event causation (Lowe, 1999).

P1: If M events cause P events, then mental causation isn’t a problem for dualism.
P2: Mental events do cause physical events.
C: So mental causation isn’t a problem for dualism.

So this is based on Lowe’s distinction between event and agential causation. Agential causation means that an agent can cause events that mere physical event causation cannot and agential causation cannot be reduce to event causation. Agents aren’t events nor are they processes, al they are entities with powers that can enact causal chains. So mental events which are caused by agents can cause physical events sans violating any laws of nature, and this doesn’t require that mental events are identical to or reducible to physical events. Thus if mental events can cause physical events—which we have experiential and empirical evidence that they do—then the problem of mental causation does not pose a problem for dualism.

Premise 2 now needs defense. Mental events like thoughts, beliefs and desires can cause bodily movements. If you desire to go to the store and buy something, then your desires are directing your actions; the mental intention to move the body is then carried out. Also, experiencing pain (a mental event) can cause a reaction, which would then cause the avoidance of the source of the pain. Furthermore, we also deliberate on what to do, while considering different outcomes and options and then act based on our mental states at the time. Lastly, if mental events did not cause physical events, then we wouldn’t be able to hold people responsible for their intentional actions. So P2 is true and the conclusion then follows that mental causation isn’t a problem for dualism.

Then we have Krodel’s (2013) counterfactual argument for mental causation which can be formed like this:

If the mind were different, then the physical world would have been different, al M events cause P events in the world. Krodel (2013: 3) puts it like this:

My headache caused me to take an aspirin. This claim sounds as natural as any. A dualist too can make it. More importantly, a dualist can provide a rigorous argument for it. Nothing depends on the specifics of headaches and our reactions to them, so let m be some actually occurring mental event and b it’s actually occurring later behavioral effect (‘putative behavioral effect’, if you like, to quell any suspicion of begging the question). The argument has a complicated part with the conclusion that if m had not occurred, then b would not have occurred, and a simple part with the conclusion that m caused b. Let us start with the complicated part:

(1) If none of m’s physical bases had occurred, then b would not have occurred.
(∼∪P → ∼B)

(2) If m had not occurred, then none of m’s physical bases would have occurred.
(∼M → ∼∪P)

(3) If none of m’s physical bases had occurred, then m would not have occurred.
(∼∪P → ∼M)

———————
(4) If m had not occurred, then b would not have occurred.
(∼M → ∼B)

It can also be put like this:

P1: If a mental event didn’t occur, then a physical event wouldn’t have occurred.
P2: If a mental event didn’t occur, then a different physical event would have occurred.
C: So the mental event causally contributed to the physical event.

So mental events make the difference to the counterfactual dependence of the physical events on the mental causes. So the mental event is causally relevant to the instantiation of the physical event, even if the mental event isn’t causally necessary. So again, the problem of mental causation isn’t a problem for dualism.

The argument for the uniqueness of human intentional consciousness

This is going to be a long argument, so bear with me.

P1: Humans are capable of intentional action, which involves the use of reason and purpose to achieve goals.
P2: Intentional action requires a mental capacity to represent and reason about the world.
P3: This mental capacity is what we refer to as MIND.
C1: Thus, humans possess a mind. (modus ponens, P1, P2, P3)
P4: Humans have intentional consciousness which involves being aware of thoughts, beliefs, and desires along with the ability to form and pursue goals.
P5: Intentional consciousness requires the ability to represent and reason about mental states.
P6: The possession of a mind is necessary for intentional consciousness.
C2: So humans have intentional consciousness (modus ponens, P6, C1).
P7: To be able to think, an organism must have a full range of propositional attitudes like beliefs, desires, intentions, and knowledge.
P8: Having a full range of propositional attitudes rests on having language.
P9: Nonhuman animals lack language.
P10: Nonhuman animals lack a full range of propositional attitudes.
P11: Since a full range of propositional attitudes is necessary for thinking, then nonhuman animals can’t think.
C3: So nonhuman animals lack MIND. (modus tollens, P3, P10, P11)
P12: Nonhuman animals have phenomenal consciousness, that is, there is something it is like to be a certain animal.
P13: Phenomenal consciousness does not require the ability to represent and reason about mental states.
P14: Intentional consciousness is a higher level of consciousness that requires both phenomenal consciousness and the ability to reason about and represent mental states.
C4: Therefore, no nonhuman animal possesses intentional consciousness. (modus tollens, C2, P12, P14)

P1-P6 establish that humans possess a mind and intentional consciousness and that nonhuman animals lack intentional consciousness. P7-P11 build on this and introduce the ideas that in order to think, an organism needs to have a full range of propositional attitudes, and since a full range of propositional attitudes rests on having language, and nonhuman animals lack language, then nonhuman animals can’t think which then leads to the conclusion that nonhuman animals lack MIND because thinking is a necessary component of MIND. P12-P14 state the distinction between phenomenal and intentional consciousness, and show that nonhuman animals have phenomenal consciousness but not intentional consciousness. Phenomenal consciousness does not require the ability to represent and reason about mental states, which is necessary for intentional consciousness.

So, minds exist and humans have them, humans have intentional consciousness (since humans can reason to achieve goals), nonhuman animals lack mind (since they lack language and therefore propositional attitudes) and nonhuman animals lack intentional consciousness (since they have phenomenal consciousness and no nonhuman animal can have intentional consciousness).

Conclusion

I have argued that humans have consciousness and that consciousness isn’t reducible to physical or material processes. Consciousness is a nonphysical phenomenon and since it is nonphysical, then only an immaterial, nonphysical thing can support or generate nonphysical consciousness and this immaterial thing is the mind. Mental causation isn’t a problem for dualism, since mental events can and do cause physical events. Krodel’s counterfactual argument for mental causation was provided to help establish the claim. Lastly, I argued that humans are capable of intentional action and so they possess minds, while arguing that humans have intentional consciousness nonhuman animals lack mind and so nonhuman animals lack intentional consciousness. This is due to the fact that nonhuman animals lack language and so they lack propositional attitudes and therefore intentional states.

So the ultimate conclusion here is that humans are special, a part of our constitution is nonphysical and irreducible (MIND), and so nonhuman animals don’t share MIND since they lack language and propositional attitudes, so they lack intentional consciousness.

Advertisement

Arguments Against Measurement Invariance

2100 words

Introduction

Measurement invariance (MI) is a concept used in psychology, sociology, and education research which purports to describe the property that is being measured across different groups and/or contexts (eg Boorsboom, 2006), while measurement noninvariance (MNI) means that “a construct has a different structure or meaning to different groups or on different measurement occasions in the same group, and so the construct cannot be meaningfully tested or construed across groups or across time” (Putnick and Borstein, 2017). MI assumes that psychological constructs are measurable and that the same construct can be measured across different groups and contexts.

But MI doesn’t need to be a concrete physical reality to show that the construct that’s supposedly being measured is the same between cultures/contexts, so the story goes. Furthermore, establishing MI is important in showing that a construct is a valid one—that is, if a so-called measure is construct valid. But upon further examination of MI, the concept seems to be highly flawed. It assumes that psychological constructs are fixed, objective and can be measured with the same precision as physical objects. However, this assumption runs counter to the arguments I’ve been making for years about the immateriality of mind and psychological processes along with the limits of scientific inquiry.

In this article, I will provide an argument against the concept of MI and defend it’s premises, arguing that MI is an unrealistic, invalid concept due to the assumptions both implicit and explicit in the concept. Further, if a test is measurement invariant, then scores are comparable between groups and any differences between the groups can be said to be down to true differences in the hypothesized construct that is supposedly being measured. But if this construct is a psychological one, then there is no measurement occurring. But it is important to note that even if a test is so-called measurement invariant, that does not mean that a test is free from bias.

Arguing against measurement invariance

So MI is concerned with operationalizing and measuring certain indicators. However, since MI assumes that the underlying psychological construct that is purported to be measured exists, is objective, and a stable entity that can be measured accurately can be seen to be a form of reification, just like the concept of “general intelligence.” Although I reject MI due to its conceptual and theoretical assumptions as I will argue below, it is worth noting that Wicherts (2016; cf Shuttleworth-Edwards, 2016) showed that not all IQ tests are measurement invariant—that is, some are measurement noninvariant. Wicherts states that “psychological test scores need to be valid and reliable“—though I am not aware of any IQ test that is indeed construct valid (Richardson and Norgate, 2015).

Construct validity (CV) is also related to the overarching arguments I will mount. CV refers to whether or not a test or measure accurately assessed what it is purported it does. So when it comes to. CV for “intelligence” and other psychological traits, how can CV be established if there is no physicality to the construct—if they are immaterial? Further, the neurosciences have attempted to show CV for IQ, but this fails too.

I’m not really worried about whether or not IQ tests results are or are not measurement invariant. I am, however, concerned with the conceptual and theoretical underpinnings and assumptions of MI. It is these assumptions and theoretical underpinnings, I will argue, that invalidate the concept of MI. Long-time readers will know that I argue against the materiality (and so argue for) of psychological traits. Since MI is concerned with operationalizing and measuring certain indicators that supposedly relate to the constructs in question, the arguments I am about to give and have given in the past completely invalidate MI. These arguments are conceptual, and so empirical evidence is irrelevant to them. MI assumes that psychological constructs are measurable—that is, there is a physical basis for psychological constructs. Though, metaphysical/philosophical arguments refute this claim.

Now I will give the arguments against MI.

P1: MI assumes that the same construct is being measured in the same way across different groups and contexts.
P2: The meaning of a construct may differ between different cultural and social groups.
P3: Differences in cultural and social meaning can affect the reliability and validity of the measurement instruments which are used to assess the construct.
C: Thus, it may be inappropriate to assume that measures of psychological constructs are invariant across different cultural and social groups.

Premise 1: P1 is definitional, and is the accepted definition of MI in the literature.

Premise 2: Cultural and social differences can influence the meaning of a construct. A construct like self-esteem may be interpreted differently in individualistic and collectivist cultures, where an individualistic culture would look at self-esteem as an individual attribute while a collectivist culture would look at it as a relational attribute (eg between Western and East Asian cultures) (Heine and Hamamura, 2007). Further, aggressiveness and assertiveness could also be valued differently in individualistic and collectivist cultures (Church, 2000).

Premise 3: So-called measurement instruments that are used to assess psychological constructs could be influenced by cultural and social differences. A psychological test developed and “validated” in one culture may not be “valid” due to differences in culture, language, or norms (van de Vijver and Leung, 1997). Furthermore, differences in item content or even response styles may also lead to measurement bias which could then affect the reliability and validity of the so-called measurement instruments which then could affect the reliability and validity of the so-called measures.

So the conclusion here logically follows from the premises: so it could be inappropriate to assume that measures of psychological constructs are comparable between groups.


P1: If MI is a valid concept, then the same construct should be measured in the same way across groups/contexts.
P2: If cultural and social differences can affect the meaning and measurement of a construct, then the same construct cannot be measured in the same way across groups/contexts.
P3: Cultural or social differences can affect the meaning of a construct.
C: So MI isn’t a valid concept. (Hypothetical syllogism)

If psychological constructs can have different meanings across cultural and social groups, and if these differences can affect the “measurement” of the so-called construct, then it is impossible to achieve MI as defined in P1. While it is important to note that MI would then need to be redefined or applies with caution and awareness of cultural and social factors relevant to the so-called measurement, it is also important to note that different cultures convey different meaning to concepts. This argument is a sound one.

Premise 1: Again, this is definitional and a widely-accepted definition of MI.

Premise 2: Different cultures have different psychological and cultural tools, which then affect how those who find themselves in that culture think and behave (eg Markus and Kitayama, 1991; Nisbett, 2003; Heine, 2008). The implication here is obviously that psychological constructs have different meanings based on cultural, national, and social context. So the references provided show a valid concern which challenges a main assumption of MI.

Premise 3: Along with the references to Markus and Kitayama, Nisbett, and Heine, it has also been observed that cultural differences in beliefs and values could affect how people respond to personality tests, and that language can affect the interpretation of so-called psychological scales. So P3 is well-evidenced.

So the conclusion logically follows and what I wrote above in the preceding paragraph below the argument is the same.


P1: Psychological constructs are fundamentally subjective and culturally situated.
P2: MI assumes the possibility of objective and universal measurement of psychological constructs.
C: Therefore, MI is completely invalid.

Premise 1: Psychological constructs are abstract, subjective constructs which reflect the subjective experiences, values, and beliefs of individuals and communities. They are, furthermore, also shaped by cultural and social contexts which could affect the meanings and associations of particular constructs (Markus and Kitayama, 1991; Nisbett, 2003; Heine, 2008). So this implies that psychological constructs cannot be measured in a consistent and objective way across cultural and social contexts, since they are inherently subjective and culturally situated.

Premise 2: This, again, is definitional. So based on the well-accepted definition of MI, psychological constructs are universal and can be measured in a consistent and objective way across cultures. But P1 directly contradicts this—psychological constructs are subjective and culturally situated.

Conclusion: So based on P1 and P2, MI is invalid. If psychological constructs are fundamentally subjective and culturally situated, and if MI assumes the possibility and objective measurement of these constructs, then MI is fundamentally (conceptually and theoretically) flawed and so it cannot be applied in a meaningful way. So MI is an invalid concept.

The conceptual arguments I have given against MI challenge the assumptions and underlying theoretical framework of MI. So, the next and final argument I will provide will prove that the nature and immateriality of psychological traits means that MI is an invalid concept.


P1: If construct validity requires that the same construct is being measured in the same way across contexts and groups, and if psychological traits are immaterial and so immeasurable, then MI is an invalid concept.
P2: Psychological traits are immaterial and so immeasurable.
P3: Construct validity requires that the same construct is being measured in the same way across contexts and groups.
C: So MI is an unrealistic and invalid concept.

Premise 1: CV requires that the same construct is being measured across all groups and contexts (Kane, 2006). If we can it ensure that we are measuring the same construct across all groups and contexts, then we cannot establish construct validity. So if psychological traits are immaterial and immeasurable, and we cannot ensure that we are measuring the same construct across different groups and contexts, then the concept of MI is unrealistic and therefore invalid.

Premise 2: This premise is of course widely-debated in philosophy of mind (Chalmers, 1996), and I have provided numerous arguments that establish the immateriality and immeasurability of psychological traits (hereherehere and here). So since psychological traits are immaterial and not directly observable or measurable, then they pose a huge and insurmountable obstacle for the psychometric claims that mentality can be measured.

Premise 3: This premise establishes that for the same construct to be measured across groups and contexts, then it needs to be established that the same construct is indeed being measured. However, see the discussion of the conclusion below.

Conclusion: Since psychological traits are immaterial and so immeasurable, and construct validity requires that the same construct is being measured across all groups and contexts, then MI isn’t a valid concept.

Conclusion

I have discussed what MI is and what is required for MI in the literature. I then gave four arguments which conclude that MI isn’t a valid concept. This is because different cultures have different psychological and cultural tools, and so different concepts would have different meaning in different cultures and nations, which does indeed have empirical support.

Most importantly, the immateriality of psychological traits means that they are not directly observable or measurable, and therefore, this is yet another reason why the concept of MI is invalid. For X to be measured, X needs a specified measured object, object of measurement and measurement unit (Berka, 1983; Nash, 1990). I have termed this “the Berka/Nash measurement objection.” For if those three conditions do not hold, then one cannot logically state that they are measuring anything. Furthermore—regarding standardized tests—they exist to assess social function (Garrison, 2009: 5). Psychometricians also assume that what they are measuring is quantitative, without showing that it is. That “something” is being measured that we don’t know due to the the fact that tests and questionnaires obtain numerical values is patently ridiculous. Psychometricians design that test first and then attend it to ascertain what it measures (Nash, 1990).

Now, while I think I have refuted the concept of MI, of course this will continue to be used in psychometric research. However, just because a thing is continously used and just because a whole field continues even when there are lethal conceptual objections against it doesn’t mean that they are in the clear. The fact of the matter is, psychometrics isn’t measurement (Uher, 2021), and no experiment can establish that it is—psychometricians need to grapple with the conceptual arguments, first. MI just isn’t a valid concept due to the arguments and reasons I have given.

Empirical Evidence is Irrelevant to Conceptual Arguments

2050 words

Introduction

Empirical arguments rely on scientific data—data derived from the five senses. Conceptual arguments don’t rely on empirical evidence—they rely on thinking about concepts. So then we can say that it’s about a priori vs. a posteriori knowledge. A posteriori knowledge is empirical/scientific knowledge while a priori knowledge is conceptual/logical knowledge. An argument about concepts would be analyzing the concepts which make up a proposition while an empirical argument would be an argument derived from our senses and what we observe. In this article, I will articulate the distinction between empirical and conceptual argument and then provide arguments why empirical evidence is irrelevant to conceptual objections. This, then, has implications for things like the reducibility of the mental to the physical.

Empirical argument

An empirical argument is one in which scientific data is paramount to it and observations and using our five senses are key to gaining knowledge. A related inquiry is what philosopher of mind Markus Gabriel calls rampant empiricism in his book I am not a Brain. Rampant empiricism is the philosophical claim that all knowledge can be derived from our five senses. However the claim that all knowledge derives from sense experience is a philosophical—conceptual—claim, and can’t be corrected with sense experience. In my article Against Scientism, I articulated an argument against the claim that scientism is true:

Premise 1Scientism is justified.
Premise 2If scientism is justified, then science is the only way we can acquire knowledge.
Premise 3We can acquire knowledge through logic and reasoning, along with science.
Conclusion: Therefore scientism is unjustified since we can acquire knowledge through logic and reasoning.

While empirical evidence and argument do, of course, hold value—as evidenced with scientific inquiry—it is quite clear that we can nevertheless gain knowledge through thinking about concepts, through a priori reasoning. That is, we can acquire knowledge through logic and reasoning, so there is more than one way to gain knowledge. So evidence is empirical if it is derived through one of the five senses, that is if it is accessible to sense experience.

Empirical evidence is concerned with the physical world. That is, what we can see and measure. It is based on observation, experience, and measurement of physical quantities. So, for example, if X isn’t quantifiable, then X can’t be measured, therefore X wouldn’t be subject to empirical verification so it would be subject to conceptual argumentation.

Conceptual argument

A conceptual argument is an argument that doesn’t rely on empirical evidence—it is a priori (Bojanic, Laquinto, and Torrengo, 2018). It merely relies on logic and reasoning to gain knowledge. However, “logic without empirical supports can only be used to prove conceptual truths” (Icefield, 2020). Such arguments are based on abstract concepts, ideas, and principles, and claims are established based on a logical analysis of concepts.

Conceptual arguments are based on a priori knowledge such as mathematical proofs, conceptual definitions and logical principles. Knowledge like this can be established through reasoning and reasoning alone without appeal to empirical evidence. Since a priori knowledge is knowledge gained without appeal to empirical evidence, or observation, empirical evidence is thusly irrelevant to conceptual arguments. Concepts are general meanings of linguistic predicates, and so philosophy itself is an a priori, conceptual discipline, which relies on deduction.

Conclusions in an a priori, conceptual argument are established with logic and reasoning without appealing to empirical data. Such examples are, of course, the relationship between the mind and body, and the nature of causation. There could be no scientific experiments which would establish which theory in philosophy of mind would be true, nor could there be a scientific experiment which would establish the nature of causation.

For many, philosophy is essentially the a priori analysis of concepts, which can and should be done without leaving the proverbial armchair. We’ve already seen that in the paradigm case, an analysis embodies a definition; it specifies a set of conditions that are individually necessary and jointly sufficient for the application of the concept. For proponents of traditional conceptual analysis, the analysis of a concept is successful to the extent that the proposed definition matches people’s intuitions about particular cases, including hypothetical cases that figure in crucial thought experiments.

A related attraction is that conceptual analysis explains how philosophy could be an a priori discipline, as many suppose it is. If philosophy is primarily about concepts and concepts can be investigated from the armchair, then the a priori character of philosophy is secured (Jackson 1998). (SEP, Concepts)

The argument against empirical evidence being relevant to conceptual arguments

I have articulated a few arguments for the claim that empirical evidence is irrelevant to conceptual arguments.

P1: If empirical evidence is relevant to conceptual arguments, then empirical evidence can be used to support or refute conceptual claims.

P2: Empirical evidence cannot be used to support or refute conceptual claims.

C: Thus, empirical evidence is irrelevant to conceptual arguments.

Conceptual claims are based on a priori truths and so we can know things without experience, whereas empirical evidence is based on experience and observations and so they cannot refute conceptual claims. There is, though, no kind of empirical evidence that can refute conceptual claims from philosophical analysis.


P1: If empirical evidence is relevant to conceptual arguments, then all conceptual arguments would be subject to empirical verification.

P2: There are conceptual arguments that are not subject to empirical verification.

C: Therefore, empirical evidence is irrelevant to conceptual arguments.

A few conceptual arguments that aren’t subject to empirical verification include: Fodor and Piattelli-Palmarini’s (2009, 2010) argument against natural selection, the argument against the reducibility of the mental, the Berka/Nash measurement objection, a theory/definition of “intelligence“, the argument against localization of cognitive functions in the brain (Uttal, 2001, 2012, 2014), the argument against animal mentality (Davidson, 1982), the argument against machine intentionality/mindedness, the argument against the possibility of science being able to study subjective states, and Gettier’s argument against knowledge as justified true belief (Chalmers and Jackson, 2001). These all have one thing in common: There can never be any empirical evidence that would validate or invalidate these arguments, and so falsification would be irrelevant. One would disprove the arguments not by empirical means—if they could—but by dealing with the logic/concepts of the arguments. These arguments deal with abstract, theoretical concepts which cannot be observed or tested empirically; they rely on logical reasoning like deduction and using inferences and not observation and measurement ; many involve normative judgments; and they are subject to interpretation. While empirical evidence is concerned with the directly observable, measurable physical world.

Furthermore, arguments like those in Gourionouva and Mansvelder (2019) fall prey to the facts: that there is no definition or theory of “intelligence”; that twin studies don’t show genetic influence (Joseph, 2014) since the “laws” of behavioral genetics don’t hold; and last but not least that neuroimaging studies don’t and can’t do what they set out to do (Uttal, 2001, 2012, 2014), since neuroreductionism is false along with the fact that neuroscience assumes mind is physical and reducible to the CNS.


P1: Empirical evidence relies on the observation and measurement of physical phenomena.

P2: Conceptual arguments are based on abstract concepts and logical deduction.

P3: Abstract concepts cannot be observed or measured through empirical means.

C: Thus, empirical evidence is irrelevant to conceptual arguments.

P1 asserts that empirical evidence is based on observation and measurement of physical phenomena. P2 establishes that conceptual arguments are based on logical deduction and abstract concepts which aren’t observable or measurable empirically. P3 follows from the definition of abstract concepts and logical deduction and how they aren’t directly observable or measurable phenomena. So the conclusion necessarily and logically follows from the premises—it’s impossible to directly measure or observe abstract concepts or logical deductions which are the basis of conceptual arguments, so empirical evidence is irrelevant to conceptual arguments.


P1: If a proposition is conceptual, then it is not based on empirical evidence.

P2: If a proposition is not based on empirical evidence, then it is not subject to empirical testing.

P3: So if a proposition is conceptual, then it is not subject to empirical testing.

P4: If a proposition is not subject to empirical testing, then empirical evidence is irrelevant to assessing the truth or validity of the proposition.

C: Thus, if a proposition is conceptual, then empirical evidence is irrelevant to assessing the truth or validity of the proposition.

This argument builds on the others, and using hypothetical syllogism, successfully concludes that if a proposition is conceptual then it isn’t subject to empirical verification. Thus, taken together, empirical evidence is irrelevant to conceptual arguments, and so falsification, too, is irrelevant. Since falsification is irrelevant, then testability of the conceptual argument is, too, irrelevant. Conceptual arguments are irrefutable using the methods of scientific inquiry and can only be refuted using philosophical inquiry.


P1: If empirical evidence is relevant to an argument, then the argument must be testable using empirical methods.

P2: Conceptual arguments aren’t testable using empirical methods.

C: So empirical evidence is irrelevant to conceptual arguments.

This argument is simple: If empirical evidence is relevant to an argument, but conceptual arguments aren’t testable through empirical methods, then empirical evidence isn’t relevant to conceptual arguments. This is due to the distinction between empirical and conceptual arguments/evidence.


Now I can use destructive dillema to argue that empirical evidence is irrelevant to the mind-body problem.

P1: If an argument is conceptual, then it is based on abstract concepts and logical deduction.

P2: If an argument is based on abstract concepts and logical deduction, then it cannot be observed or measured through empirical means.

P3: If an argument cannot be observed or measured through empirical means, then empirical evidence is irrelevant to that argument.

P4: The mind-body problem is a conceptual argument.

C: Therefore, empirical evidence is irrelevant to the mind-body problem.

If the mind-body problem is a conceptual argument (P4), then it is based on abstract concepts and logical deduction (P1), so it cannot be observed or measured through empirical means (P2), thus empirical evidence is irrelevant to the mind-body problem (P3), ultimately meaning that the mind-body problem—along with the mental and first-personal subjective states—cannot be studied by science. This of course—as I have been arguing for years—has implications for the so-called hereditarian hypothesis and any kind of genetic or neuroimaging studies they attempt and assert that shows that the mental is reducible to genes or brain physiology. Indeed, a priori philosophical conceptual analysis shows that consciousness cannot be reduced to any material features—meaning it is outside of the bounds of scientific explanation, just like first-personal subjective states (Chalmers and Jackson, 2001).

Conclusion

I have articulated the distinction between empirical and conceptual arguments, and provided valid and sound arguments which distinguish between both types of argument. The nonidentity between how the types of argument work and gather support for premises shows why empirical evidence is irrelevant to conceptual arguments.

Quite clearly, there is evidence that isn’t empirical and so it isn’t subject to empirical verification/falsification/testing. This is because logical concepts/argumentation/reasoning are subject to falsification from the scientific method. For one to be able to successfully reject conceptual arguments, they need to grapple with the logic and reasoning of the arguments. No empirical evidence would be able to show that a proposition is false, since empirical evidence wasn’t used for the proposition.

The most powerful arguments against so-called hereditarianism are conceptual, and no matter what kinds of studies the hereditarian conjures up, none of them will refute the arguments against the possibility of psychophysical reductionism. So if consciousness (mind) cannot be reductively explained by the scientific method, then hereditarianism fails and therefore, there cannot be a science of the mind.

Since empirical arguments rely on observation or experimentation, and conceptual arguments rely on logic and reasoning, they have different bases of evidence. Empirical arguments are concerned with observable phenomena, while conceptual arguments are concerned with abstract concepts. Empirical arguments rely on the scientific method and data collection, while conceptual arguments rely on philosophical frameworks, logic, and reasoning.

The validity of a conceptual argument relies on the logic of its inferences along with the consistency and coherence of its logical framework and the soundness of the logic underlying the premises of the argument. The distinction and non-identity between the two types of evidence allows us to rightly state that, for these reasons why empirical evidence is irrelevant to conceptual arguments.

Three Recent Environmentalist Books that Perpetuate the “IQ-As-A-Measure-of-Intelligence” Myth

2500 words

The hereditarian-environmentalist debate has been ongoing for over 100 years. In this time frame, many theories have been forwarded to explain the disparity between individuals and groups. In one camp you have the hereditarians who claim that any non-zero heritability for IQ scores means that hereditarianism is true (eg Warne, 2020); while in the other camp you have the environmentalists who claim that differences in IQ are explained by environmental factors. This debate has been raging since the 1870s when Francis Galton coined the “nature-nurture” dichotomy still rages today. Unfortunately, the environmentalists lend credence to IQ-ist claims that, however imperfect, IQ tests are “measures” of intelligence.

Three recent books on the matter are A Terrible Thing to Waste: Environmental Racism and its Assault on the American Mind (Washington, 2019), Making Kids Cleverer: A Manifesto for Closing the Advantage Gap (Didau, 2019), and Young Minds Wasted: Reducing Poverty by Enhancing Intelligence in Known Ways (Schick, 2019). All three of these authors are clearly environmentalists and they accept the IQ-ist canard that IQ—however crudely—is a “measure” of “intelligence.”

There are, however, no sound arguments that IQ tests “measure” intelligence and there is no response to the Berka/Nash measurement objection for the claim that IQ tests are a “measure” since no hereditarian can articulate the specified measured object, the object of measurement and the measurement unit for IQ; there is, also, no accepted definition or theory of “intelligence”. So how can we say that some”thing” is being “measured” with a certain instrument if we have no satisfactorily defined what we claim to be measuring with a well-accepted theory of what we are measuring (Richardson and Norgate, 2015; Richardson, 2017), with a specified measured object, object of measurement, and measurement unit (Berka, 1983a, 1983b; Nash, 1990; Garrison, 2003, 2009) for the construct we want to measure?

But the point of this article is that environmentalists push the hereditarian canard that IQ is equal to, however crudely, intelligence. And though the authors do have great intentions and are pointing to things that we can do to attempt to ameliorate differences between individuals in different environments, they still lend credence to the hereditarian program.

A Terrible Thing to Waste

Washington (2019) discusses the detrimental effects (and possible effects of others) of lead, mercury and other metals that are more likely to be found in low-income black and “Hispanic” communities along with iodine deficiencies. These environmental exposures retard normal brain development. But, one is not justified in claiming that they are measures of “intelligence”—at best, as Washington (2019) argues, we can claim that they are indexes of environmental polluters on the brains of developing children.

Intelligence is a product of environment and experience that is forged, not inherited; it is malleable, not fixed. (Washington, 2019: 20)

While it is true, as Washington claims, that we can mitigate these problems from the toxic metals and lack of other pertinent nutrients for brain development by addressing the problems in these communities, it does not follow that IQ is a “biological” thing. Yes, IQ is malleable (contra hereditarian claims), and Headstart does work to improve life outcomes, even though such gains “fade out” after the child leaves the enriched environment. Lead poisoning, for example, has led to a decrease in 23 million IQ points per year (Washington, 2019: 15). But I am not worried about lost IQ points (even though by saving the IQ points from being lost, we would then be directly improving the environments that lead to such a decrease). I am worried about the detrimental effects of these toxic chemicals on the developing minds of children; lost IQ points are an outcome of this effect. At best, IQ tests can track cognitive damage due to pollutants in these communities (Washington, 2019) but they do NOT “measure” intelligence. (Also note that lead consumption is associated with higher rates of crime so this is yet another reason to reduce the consumption of lead in these communities.)

Speaking of “measuring intelligence”, Washington (2019: 29) noted that Jensen (1969: 5) stated that while “intelligence” is hard to define, it can be measured… But how does that make any sense? How can you measure what you can’t define? (See arguments (i), (ii), and (iii) here.)

Big Lead, though, “actively encouraged landlords to rent to families with vulnerable young children by offering financial incentives” (Washington, 2019: 55). This was in reference to the researchers who studied the deleterious effects of lead consumption on developing humans. “The participation of a medical researcher, who is ethically and legally responsible for protecting human subjects, changes the scenario from a tragedy to an abusive situation. Moreover, this exposure was undertaken to enrich landlords and benefit researchers at the detriment of children” (Washington, 2019: 55). We realized that lead had deleterious effects on development as early as the 1800s (Rabin, 2008), but Big Lead pushed back:

[Lead Industries Association’s] vigorous “educational” campaign sought to rehabilitate lead’s image, muddying the waters by extolling the supposed virtues of lead over other building materials. It published flooding guides and dispatched expert lecturers to tutor architects, water authorities, plumbers, and federal officials in the science of how to repair and “safely” install lead pipes. All the while the [Lead Industries Association] staff published books and papers nd gave lectures to architects and water authorities that downplayed lead’s dangers. 11 (Washington, 2019: 60)

In any case, Washington’s book is a good read into the effects of toxic metals on brain development, and while we must do what we can to ameliorate the effects of these metals in low-income communities, IQ increases are a side effect of ameliorating the toxic metals in these communities.

Making Kids Cleverer

Didau (2019: 86) outright claims that “intelligence is measured by IQ tests”—he is outright pushing the hereditarian view that IQ tests “measure intelligence.” (A strange claim since on pg 95-96 he says that IQ tests are “a measure of relative intelligence.”)

In the book, Didau accepts many hereditarian premises—like the claim IQ tests measure intelligence, that heritability can partition genetic and environmental variation. Further, Didau says in the Acknowledgements (pg 11) that Ritchie’s (2015) Intelligence: All That Matters forms the backbone for much of the information in Chapters 3 and 5.” So we can see here how the hereditarian IQ-ist stance colors his view on the relationship between “IQ” and “intelligence.” He also makes the bald claims that “intelligence is a good candidate for being the best researched and best understood characteristic of the human brain” and that it’s “also probably the most stable construct in all psychology” (pg 81).

Didau takes the view that intelligence is both a way to acquire knowledge as well as what type of knowledge we know (pg 83)—basically, it’s what we know and what we do with what we know along with ways to acquire said knowledge. What one knows is obviously a product of the environment they find themselves growing up in, and what we do with the knowledge we have is similarly down to environmental factors. Didau states that “Possibly the strongest correlations [with IQ] are those with educational outcomes” (pg 92). But Didau, it seems, fails to realize that this strong correlation is built into the test since IQ tests and scholastic achievement tests are different versions of the same test (Schwartz, 1975, Richardson, 2017).

In one of the “myths of intelligence” (Myth 3: Intelligence cannot be increased, pg 102) he discusses, Didau uses a similar analogy as myself. In an article on “the fade-out effect“, I argued that if one goes to the gym, works out and gets bigger and then stops going, we can then say that going to the gym is useless since once they leave the enriched environment they lose their gains. The direct parallels to Headstart, then, is clear with my gym/muscle-building analogy.

In another myth (Myth 4: IQ tests are unfair), Didau claims that if you get a low IQ score then you are probably unintelligent, while if you get a high one, it means you know the answers to the questions—which is obviously true. Of course, to know the answers to the questions (and to be able to reason the answers for some of the questions), one must be exposed to the knowledge that is contained in that test, or they won’t score high.

We can reject the use of IQ scores by racists, he says, who would use it to justify the superiority of their own groups and the inferiority of “the other”, all while not rejecting that IQ tests are valid (where have they been validated?). “Something real and meaningful” is being measured by these tests, and we have chosen to call this “intelligence” (pg 107). But we can say this about anything. Imagine having a test Y for X. But we don’t really know what X is, nor that Y really measures it. But because it accords with our a priori biases and since we have constructed Y to get the results we think we should see, even though we have no idea what X is, we assume that we are measuring what we set out to all without the basic requirements of measurement.

While Didau does seem to agree with some of the criticisms I’ve levied on IQ tests over the years (cross-cultural testing is pointless, IQ scores can be changed), he is, obviously, pushing a hereditarian IQ-ist agenda, cloaked as an environmentalist. He contradicts himself by saying that intelligence is measured by IQ tests without then saying what he says later about them—and I don’t think one should assume that he meant they are an “imperfect measure” of intelligence. (Imagine an imperfect measure of length—would we still be using it to build houses if it was only somewhat accurate?) Didau also agrees with the g theorists, in that there is a “general cognitive ability”, as well. He also agrees with Ritchie and Tucker-Drob (2018) and Ceci (1996) that schooling can and does increase IQ scores (as summer vacations show that IQ scores do decrease without schooling) (see Didau, 2018: Chapter 5). So while he does agree that IQ isn’t static and that education can and does increase it, he is still pushing a hereditarian IQ-ist model of “intelligence”—even though, as he admits, the concept of “intelligence” has yet to be satisfactorily defined.

Young Minds Wasted

In the last book, Young Minds Wasted (Schick, 2019), while he does dispense with many hereditarian myths (such as the myth of the normal distribution, see here), he still—through an environmentalist lens—justifies the claims that IQ tests test intelligence. While he masterfully dispenses with the “IQ is normally distributed” claim (see discussion in pg 180-186), the tagline of the book is “reducing poverty by increasing intelligence, in known ways.”

The poor’s intelligence is wasted, he says, by an intelligence-depressing environment. We can see the parallels here with Washington’s (2019) A Terrible Thing to Waste. Schick claims that “the single most important and widespread cause of poverty is the environmental constraints on intelligence” (pg 12, Schick’s emphasis). Now, like Washington, Schick says that a whole slew of chemicals and toxins decrease IQ (a truism) and by identity, intelligence. Of course, living in a deprived environment where one is exposed to different kinds of toxins and chemicals can retard brain development and lead to deleterious life outcomes down the line. But this fact does not mean that intelligence is being measured by these tests; it only shows that there are environments that can impede brain development which then is mirrored in a decrease in IQ scores.

Schick says that as intelligence increases, societal problems decrease. But, as I have argued at length, this is due to the way the tests themselves are constructed, involving the a priori biases of the test’s constructors. If we can construct a test with any kind of distribution we want to, and if the items emerge arbitrarily from the heads of the test’s constructors who then try them out on a standardized sample (Jensen, 1980: 71) looking for the results they want and assume a priori, then we can make it so that what we accept as truisms regarding the relationship between IQ and life events can be turned on their head, with no logical reason to accept one set of items over another, other than that one set has a bias in which it upholds a test constructor’s previously-held biases.

Schick does agree that “intelligent behavior” can change throughout life, based on one’s life experiences. But “Human intelligence is based on several genetically determined capabilities such as cognitive functions” (pg 39). He also claims that genetic factors determine while environmental factors influence cognitve functions, memory, and universal grammar.

Along with his acceptance that genetic factors can influence IQ scores and other aspects of the mind, he also champions heritability estimates as being able to partition genetic and environmental variation in traits (even though it can do no such thing; Moore and Shenk, 2016). He—uncritically—accepts the 80/20 genetic environmental heritability from Bouchard and the 60/40 genetic environmental heritability from Jensen and Murray and Herrnstein. These “estimates”—drawn mostly from family, twin, and adoption studies (Joseph, 2015)—though, are invalid due to the false assumptions the researchers hold, neverminding the conceptual difficulties with the concept of heritability (Moore and Shenk, 2016).

Conclusion

While Washington and Schick both make important points—that those who live in poor environments are at-risk of being exposed to certain things that disrupt their development—they both, along with Didau, accept the hereditarian claim that IQ tests are tests of intelligence. While each author has their own specific caveats (some of which I agree with, and other I do not), they keep the hereditarian claim alive by lending credence to their arguments, but not looking at it through a genetic lens.

While the authors have good intentions in mind and while the research they discuss is extremely important and interesting (like the effects of toxins and metals on the development of the brain and the development of the child), they—like their intellectual environmentalist ancestors—unwittingly lend credence to hereditarian claims that IQ tests measure intelligence but they go about the causes of individual and group differences in completely different ways. These authors, with their assertions, then, accept the claim that certain groups are less “intelligent” than others. But it’s not genes that are the cause—it’s the differences in environment that cause it. And while that claim is true—that the deleterious effects Washington and Schick discuss can and do retard normal development—it, in no way shape or form, means that “intelligence” is being measured.

Normal (brain) development is indeed a terrible thing to waste; we can teach kids more by exposing them to more things, and young minds are wasted by poverty. But by accepting these premises, one does not need to accept the hereditarian dogma that IQ tests are measures of some undefined thing with no theory. That poverty and the environments that those in poverty live in impedes normal brain development which is then reflected in IQ scores, it does not follow that these tests are “measuring” intelligence—they, at best, show environmental challenges that change the brain of the individual taking the test.

One needs to be careful with the language they use, lest they lend credence to hereditarian pseudoscience.

Conceptual Arguments Against Heredetarianism

3300 words

Introduction

Hereditarianism is the theory that differences in psychology between individuals and groups have a ‘genetic’ or ‘innate’ (to capture the thought before the ‘gene’ was conceptualized) cause to them—which therefore would explain the hows and whys of, for example, the current social hierarchy. The term ‘racism’ has many referents—and using one of the many definitions of ‘racism’, one could say that the hereditarian theory is racist since it attempts to justify and naturalize the current social hierarchy.

In what I hope is my last word on the IQ/hereditarian debate, I will provide three conceptual arguments against hereditarianism: (1) psychologists don’t ‘measure’ any’thing’ with their psychological tests since there is no object of measurement, no specified measured object, and measurement unit for any specific trait; only physical things can be measured and psychological ‘traits’ are not physical so they cannot be measured (Berka, 1983; Nash, 1990; Garrison, 2009); (2) there is no theory or definition of “intelligence” (Lanz, 2000; Richardson, 2002; Richardson and Norgate, 2015; Richardson, 2017) so there can be no ‘measure’ of it, the example of temperature and thermometers will be briefly mentioned; (3) the logical impossibility of psychophysical reduction entails that mental abilities/psychological traits cannot be genetically inherited/transmitted; and (4) psychological theories are influenced by the current problems/going-on in society as well as society influencing psychological theories. These four objections are lethal to hereditarianism, the final showing that psychology is not an ‘objective science.’

(i) The Berka/Nash measurement objection

The Berka/Nash measurement objection is simply: if there is no specified measured object, object of measurement or measuring unit for the ‘trait’, then no ‘thing’ is truly being ‘measured’ as only physical things can be measured. Nash gives the example of a stick—the stick is the measured object, the length of the stick is the object of measurement (the property being measured) and inches, centimeters etc are the measuring units. Being that the stick is in physical space, its property can be measured—its length. Since psychological traits are not physical (this will also come into play for (ii) as well) nor do they have a physical basis, there can be no ‘measuring’ of psychological traits. Indeed, since scaling is accepted by fiat to be a ‘measure’ of something. This, though, leads to confusion, especially to psychologists.

The most obvious problem with the theory of IQ measurement is that although a scale of items held to test ‘intelligence’ can be constructed, there are no fixed points of reference. If the ice point of water at one atmosphere fixes 276.16 K, what fixes 140 points of IQ? Fellows of the Royal Society? Ordinal scales are perfectly adequate for certain measurements. Moh’s scale of scratch hardness consists of ten fixed points from talc to diamond, and is good enough for certain practical purposes. IQ scales (like attainment test scales) are ordinal scales, but this is not really to the point, for whatever the nature of the scale it could not provide evidence for the property IQ or, therefore, that IQ has been measured. (Nash, 1990: 131)

In first constructing its scales and only then preceding to induce what they ‘measure’ from correlational studies, psychometry has got into the habit of trying to do what cannot be done and doing it the wrong way round anyway. (Nash, 1990: 133)

The fact of the matter is, IQ tests don’t even meet the minimal theory of measurement since there is no—non-circular—definition of what this ‘general cognitive ability’ even is.

(ii) No theory or definition of intelligence

This also goes back to Nash’s critique of IQ (since there can be no non-circular definition of what “IQ tests” purport to measure): There is no theory or definition of intelligence therefore there CAN BE no ‘measure’ of it. Imagine saying that you have measured temperature without a theory behind it. Indeed, I have explained in another article that although IQ-ists like Jensen and Eysenck emphatically state that the ‘measuring’ of ‘intelligence’ with “IQ tests” is “just like” the measuring of temperature with thermometers, this claim fails as there is no physical basis to psychological traits/mental abilities so they, therefore, cannot be measured. If “intelligence” is not like height or weight, then “intelligence”‘ cannot be measured. “Intelligence” is not like height or weight. Therefore, “intelligence” cannot be measured.

We had a theory and definition of temperature and then the measuring tool was constructed to measure our new construct. The construct of temperature was then verified independently of the instrument used to originally measure it, with the thermoscope which then was verified with human sensation. Thus, temperature was verified in a non-circular way. On the other hand, “intelligence tests” are “validated” circularly, if the tests correlate highly with other older tests (like Terman’s Stanford-Binet), it is held that the new test ‘measures’ the construct of ‘intelligence’—even if none of the previous tests have themselves been validated!

Therefore, this too is a problem for IQ-ists—their scale was first constructed (to agree with the social hierarchy, no less; Mensh and Mensh, 1991) and then they set about trying to see what their scales ‘measure’ with correlational studies. But we know that since two things are correlated that doesn’t mean that one causes the other—there could be some unknown third variable causing the relationship or the relationship could be spurious. In any case, this conceptual problem, too, is a problem for the IQ-ist. IQ is nothing like temperature since temperature is an actual physical measure that was verified independently of the instrument constructed to measure the construct in the first place.

Claims of individuals as ‘intelligent’ (whatever that means) or not are descriptive, not explanatory—it is the reflection of one’s current “ability” (used loosely) in relation to their current age norms (Anastasi; Howe, 1997).

(iii) The logical impossibility of psychophysical reduction

I will start this section off with two (a priori) arguments:

Anything that cannot be described in material terms using words that only refer to material properties is immaterial.
The mind cannot be described in material terms using words that only refer to material properties.
Therefore the mind is immaterial; materialism is false.

and

If physicalism is true then all facts can be stated using a physical vocabulary.
But facts about the mind cannot be stated using a physical vocabulary.
So physicalism is false.

(Note that the arguments provided are valid, and I hold them to be sound thus an objector would need to reject then refute a premise.)

Therefore, if all facts cannot be stated using a physical vocabulary and if the mind cannot be described in material terms using words that only refer to material properties, then there can, logically, be no such thing as ‘mental measurement’—no matter what IQ-ists try to tell you.

Different physical systems can give rise to different mental phenomena—what is known as the argument from multiple realizability. Thus, since psychological traits/mental states are multiply realizable, then it is impossible for psychology to reduce to mental kinds to reduce to physical kinds—the mental kind can be realized by multiple physical states. Psychological states are either multiply realizable or they are type identical to the physio-chemical states of the brain. That is a kind of mind-brain identity thesis—that the mind is identical to the states of the brain. Although they are correlated, this does not mean that the mind is the brain or that the mind can be reduced to physio-chemical states, as Putnam’s argument from multiple realizability concludes. If type-physicalism is true, then it must be true that every and all mental properties can be realized in the same exact way. But, empirically, it is highly plausible that mental properties can be realized in multiple ways. Therefore, type-identity theory is false.

Psychophysical laws are laws connnecting mental abilities/psychological traits with physical states. But, as Davidson famously argued in his defense of Anomalous Monism, there can be no such laws linking mental and physical events. There are no mental laws therefore there can be no scientific theory of mental states. Science studies the physical. The mental is not physical. Thus, science cannot study the mental. Indeed, since there are no bridge laws that link the mental and physical and the mental is irreducible to and underdetermined by the physical, it then follows that science cannot study the mental. Therefore, a science of the mind is impossible.

Further note that the claim “IQ is heritable” reduces to “thinking is heritable”, since the main aspect of test-taking is thinking. Thinking is a mental activity which results in a thought. If thinking is a mental activity which results in a thought, then what is a thought? A thought is a mental state of considering a particular idea or answer to a question or committing oneself to an idea or an answer. These mental states are, or are related to, beliefs. When one considers a particular answer to a question, they are paving the way to holding a certain belief. So when they have committed themselves to an answer, they have committed themselves to a new belief. Since beliefs are propositional attitudes, believing p means adopting the belief attitude that p. So, since cognition is thinking, then thinking is a mental process that results in the formation of a propositional belief. Thus, since thinking is related to beliefs and desires (without beliefs and desires we would not be able to think), then thinking (cognition) is irreducible to physical/functional states, meaning that the main aspect of test-taking (thinking) is irreducible to the physical thus physical states don’t explain thinking which means the main aspect of (IQ) test-taking is irreducible to the physical.

(iv) Reflexivity in psychology

In this last section, I will discuss the reflexivity—circularity—problem for psychology. This is important for psychological theorizing since, to its practitioners, psychology is seen to be an ‘objective science.’ If you think about psychology (and science) and how it is practiced, it (they) investigates third-personal, not first-personal, states. Thus, there can be no science of the mind (what psychology purports to be) and psychology can, therefore, not be an ‘objective science’ as the hard sciences are. The ‘knowledge’ that we attain from psychology comes from, obviously, the study of people. As Wade (2010: 5) notes, the knowledge that people and society are the object of study “creates a reflexivity, or circular process of cause and effect, whereby the ‘objects’ of study can and do change their behavior and ideas according to the conclusions that their observers draw about their behavior and ideas.”

It is quite clear that such academic concepts do not arise independently—in the history of psychology, it has been used in an attempt to justify the current social hierarchy of the time (as seen in 1900s America, Germany, and Britain). Psychological theories are influenced by current social goings-on. Thus, it is influenced by the bias of the psychologists in question. “The views, attitudes, and values of psychologists influence the claims they make” (Jones, Elcock, and Tyson, 2011: 29).

… scientific ideas did not develop in a vacuum but rather reflected underlying political or economic trends. 15

The current social context influences the psychological discourse and the psychological discourse influences the current social context. The a priori beliefs that one holds will influence what they choose to study. An obvious example being, hereditarian psychologists who believe there are innate differences in ‘IQ’ (they use ‘IQ’ and ‘intelligence’ interchangeably as if there is an identity relation) will undertake certain studies in order to ‘prove’ that the relationship they believe to be true holds and that there is indeed a biological cause to mental abilities within and between groups and individuals. Do note, however, that we have the data (blacks score lower on IQ tests) and one must then make an interpretation. So we have three possible scenarios: (1) differences in biology cause differences in IQ; (2) differences in experience cause differences in IQ; or (3) the tests are constructed to get the results the IQ-ists want in order to justify the current social hierarchy. Mensh and Mensh (1991) have succinctly argued for (3) while hereditarians argue for (1) and environmentalists argue for (2). While it is indeed true that one’s life experiences can influence their IQ scores, we have seen that it is logically impossible for genes to influence/cause mental abilities/psychological traits.

The only tenable answer is (3). Such relationships, as noted by Mensh and Mensh (1991), Gould (1996), and Garrison (2009), between test scores and the social hierarchy are interpreted by the hereditarian psychologist thusly: (1) our tests measure an innate mental ability; (2) if our tests measure an innate mental ability, then differences in the social hierarchy are due to biology, not environment; (3) thus, environmental differences cannot account for what is innate between individuals so our tests measure innate biological potential for intelligence.

The [IQ] tests do what their construction dictates; they correlate a group’s mental worth with its place in the social hierarchy. (Mensh and Mensh, 1991)

Richards (1997) in his book on racism in the history of psychology, identified 331 psychology articles published between 1909 (the first conceptualization of the ‘gene’, no less) and 1940 which argued for biology as a difference-maker for psychological traits while noting that 176 articles for the ‘environment’ side were published in that same time period.

EfV3_d2XgAEkn3c

Note that the racist views of the psychologists in question more than likely influenced their research interests—they set out to ‘prove’ their a priori biases. Indeed, they even modeled their tests after such biases. Tests that were constructed that agreed with their a priori pre-suppositions on who was or was not intelligence was kept whereas those that did not agree with those notions were thrown out (as noted by Hilliard, 2012). This is just as Jones, Elcock, and Tyson (2011: 67) note with the ‘positive manifold’ (‘general intelligence’):

Subtests within a battery of intelligence tests are included n the basis of them showing a substantial correlation with the test as a whole, and tests which do not show such correlations are excluded.

From this, it directly follows that psychometry (and psychology) are not sciences and do not ‘measure’ anything (returning to (i) above). What psychometrics (and psychology) do is attempt to use their biased tests in order to sort individuals into where they ‘belong’ on the social hierarchy. Standardized testing (IQ tests were one of the first standardized tests, along with the SAT)—and by proxy psychometrics—is NOT a form of measurement. The hierarchy that the tests ‘find’ is presupposed to exist and then constructed into existence using the test to ‘prove’ their biases ‘right.’

Indeed, Hilliard (2012) noted that in South Africa in the 1950s that there was a 15-point difference in IQ between two white cultural groups. Rather than fan flames of political tension between the groups, the test was changed in order to eliminate the difference between the two groups. The same, she notes, was the case regarding IQ differences between men and women—Terman eliminated such differences by picking and choosing certain items that favored one group and balanced them out so that they scored near-equally. These are two great examples from the 20th century that demonstrate the reflexivity in psychology—how one’s a priori biases influence what they study and the types of conclusions they draw from data.

Psychology, at least when it comes to racial differences in ‘IQ’, is being used to confirm pre-existing prejudices and not find any ‘new objective facts.’ “… psychology [puts] a scientific gloss on the accepted social wisdom of the day” (Christian, 2008: 5). This can be seen with a reading into the history of “IQ tests” themselves. The point is, that psychology and society influence each other in a reflexive—circular—manner. Thus, psychology is not and cannot be an ‘objective science’ and when it comes to ‘IQ’ the biases that led to the bringing of the tests to America and concurrently social policy are still—albeit implicitly—believed today.

Psychology originally developed in the US in the 19th century in order to attempt to fix societal problems—there needed to be a science of the mind and psychology purported to be just that. They, thusly, needed a science of ‘human nature’, and it was for this reason that psychology developed in the US. The first US psychologists were trained in Germany and then returned to the US and developed an American psychology. Though, do note that in Germany psychology was seen as the science of the mind while in America it would then turn out to be the science of behavior (Jones, Elcock, and Tyson, 2011). This also does speak to the eugenic views held by certain IQ-ists in the 20th and into the 21st century.

In Nazi Germany, Jewish psychologists were purged since their views did not line-up with the Nazi regime.

Psychology appealed to the Nazi Party for two reasons: because psychological theory could be used to support Nazi ideology, and because psychology could be applied in service to the state apparatus. Those psychologists who remained adapted their theories to suit Nazi ideology, and developed theories that demonstrated the necessary inferiority of non-Aryan groups (Jones and Elcock, 2001). These helped to justify actions by the state in discriminating against, and ultimately attempting to eradicate, thse other groups. (Jones, Elcock, and Tyson (2011: 38-39)

These examples show that psychology is influenced by society but also that society influences psychological theorizing. Clearly, what psychologists choose to study, since society influences psychology, is a reflection of a society’s social concerns. In the case of IQ, crime, etc, the psychologist attempts to naturalize and biologicize such differences in order to explain them as ‘innate’ or ‘genetic’. The rise of IQ tests in America, too, also coincided with the worry that ‘national intelligence’ was declining and so, the IQ test would need to be used to ‘screen’ prospective immigrants. (See Richardson, 2011 for an in-depth consideration on the tests and conditions that the testees were exposed to on Ellis Island; also see Gould, 1996.)

Conclusion

(i) The Berka/Nash measurement objection is one of the most lethal arguments for IQ-ists. If they cannot state the specified measured object, the object of measurement, and the measuring unit for IQ then they cannot say that any’thing’ is being ‘measured’ by ‘IQ tests.’ This then brings us to (ii). Since there is no theory or definition of what is being ‘measured’, and if the tests were constructed first before the theory, then there will necessarily be a built-in bias to what is being ‘measured’ (namely, so-called ‘innnate mental potential’). (iii) Since it is logically impossible for psychology to reduce to physical structure, and since all facts cannot be stated using a physical vocabulary nor can the mind be described using material terms that only refer to material properties, then this is another blow to the claim that psychology is an ‘objective science’ and that some’thing’ is being ‘measured’ by their tests (constructed to agree with their a priori biases). And (iv) The bias that is inherent in psychology (for both the right and the left) influences the practitioners’ theorizing and how they interpret data. Society has influenced psychology (and psychology has influenced the society) and we only need to look at America and Nazi Germany in the 20th century to see that this holds.

The relationship between psychology and society is inseparable—it is a truism that what psychologists choose to study and how and why they formulate their conclusions will be influenced by the biases they already hold about society and how and why it is the way it is. For these reasons, psychology/psychometry are not ‘sciences’ and hereditarianism is not a logically sound position. Hereditarianism, then, stays what it was when it was formulated—a racist theory that attempts to bilogicize and justify the current social hierarchy. Thus, one should not accept that psychologists ‘measure’ any’thing’ with their tests; one should not accept the claim that mental abilities can be genetically transmitted/inherited; one should not accept the claim that psychology is an objective ‘science’ due to the reflexive relationship between psychology and society.

The arguments given show why hereditarianism should be abandoned—it is not a scientific theory, it just attempts to naturalize biological inequalities between individuals and groups (Mensh and Mensh, 1991; Gould, 1996; Garrison, 2009). Psychometrics (what hereditarians use to attempt to justify their claims) is, then, nothing more than a political ring.

Superiority, Psychometrics, and Measurement

2800 words

In a conversation with an IQ-ist, one may eventually find themselves discussing the concept of “superiority” or “inferiority” as it Regards IQ. The IQ-ist may say that only critics of the concept of IQ place any sort of value-judgments on the number one gets when they take an IQ test. But if the IQ-ist says this, then they are showing their ignorance regarding the history of the concept of IQ. The concept was, in fact, formulated to show who was more “intelligent”—“superior”—and who was less “intelligent”—“inferior.” But here is the thing, though: The terms “superior” and “inferior” are, however, anatomic which shows the folly of the attempted appropriation of the term.

Superiority and inferiority

If one wants to find early IQ-ists talking about superiority and inferiority regarding IQ, they would only need to check out Lewis Terman’s very first Stanford-Binet tests. His scales—now in their fifth edition—state that IQs between 120 and 129 are “superior” while 130-144 is “gifted or very advanced” and 145-160 is “very gifted” or “highly advanced.” How strange… But, the IQ-ist can say that they were just products of their time and that no serious researcher believes such foolish things, that one is “superior” to another on the basis of an IQ score. What about proximal IQs? Lateral IQs? Posterior IQs? Distal IQs? It’s ridiculous to use anatomic terminology (for physical things) and attempt to use them to describe mental “things.”

But, perhaps the most famous hereditarian Arthur Jensen, as I have noted, wrongly stated that heritability estimates can be used to estimate one’s “genetic standing” (Jensen, 1970) and that if we continue our current welfare policies then we are in danger of creating a “genetic underclass” (Jensen, 1969). This, as does the creation of the concept of IQ in the early 1900s, speaks to the hereditarian agenda and the reason for the IQ enterprise as a whole. (See Taylor, 1980 for a wonderful discussion of Jensen’s confusion on the concept of heritability.)

This is no surprise when you understand that IQ tests were created to rank people on a mental hierarchy that reflected the current social hierarchy of the time which would then be used as justification for their spot on the social hierarchy (Mensh and Mensh, 1991). So it is no surprise that anatomic terminology was hijacked in an attempt at forwarding eugenic ideas. But the eugenicists concept of superiority didn’t always pan out the way they wanted it to go, which is evidenced a few decades before the conceptualization of standardized testing.

Galton attempted to show that those with the fastest reaction times were more intelligent, but when he found out that the common man had just as quick of a reaction time, he abandoned this test. Then Cattell came along and showed that no relationship existed between sensory perception and IQ scores. Finally, Binet showed that measures of the skull did not correspond with teacher’s assessment of who is or is not “intelligent.” Then, some decades later, Binet and Simon finally construct a test that discriminates between who they feel is or is not intelligent—which discriminated by social class. This test was finally the “measure” that would differentiate between social classes since it was based on a priori notions of an individual’s place in the social hierarchy (Garrison, 2009: 75). Binet and Simon’s “ideal city” would use test scores as a basis to shuttle people into occupations they “should be” in on the basis of their IQ scores which would show how they would work based on their “aptitudes” (Mensh and Mensh, 1991: 24; Garrison, 2009: 79). Bazemore-James, Shinaorayoon, and Martin (2017) write that:

The difference in racial subgroup mean scores mimics the intended outcomes of the original standardized IQ tests, with exception to Asian Americans. Such tests were invented in the 1910s to demonstrate the superiority of rich, U.S.-born, White men of northern European Descent over non-Whites and recent immigrants (Gersh, 1987). By developing an exclusion-inclusion criteria that favored the aforementioned groups, test developers created a norm “intelligent” (Gersh, 1987, p. 166) populationiot “to differentiate subjects of known superiority from subjects of known inferiority” (Terman, 1922, p. 656).

So, as one can see, this “superiority” was baked-in to IQ tests from the very start and the value-judgments, then, are not in the minds of IQ critics but is inherent in the scores themselves as stated by the pioneers of IQ testing in America and the originators of the concept that would become IQ. Garrison (2009: 79) writes:

With this understanding it is possible to make sense of Binet’s thinking on intelligence tests as group differentiation. That is, the goal was to group children as intelligent and unintelligent, and to grade (value) the various levels of the unintelligent (also see Wolf 1973, 152–154). From the point of view of this goal, it mattered little whether such differences were primarily biological or environmental in origin. The genius of the theory rests in how it postulates one group as “naturally” superior to the other without the assumptions of biology, for reason had already been established as a natural basis for distinction, irrespective of the origin of differences in reasoning ability.

While Binet and Simon were agnostic on the nature-nurture debate, the test items that they most liked were those items that differentiated between social classes the most (which means they were consciously chosen for those goals). But reading about their “ideal city”, we can see that those who have higher test scores are “superior” to those who do not. They were operating under the assumption that they would be organizing society along class lines with the tests being measures of group mental ability. For Binet and Simon, it does not matter whether or not the “intelligence he sought to define” was inherited or acquired, they just assumed that it was a property of groups. So, in effect, “Binet and Simon developed a standard whereby the value of people’s thinking could be judged in a standard way, in a way that corresponded with the exigencies of social reproduction at that time” (Garrison, 2009: 94). The only thing such tests do is reproduce the differences they claim to measure—making it circular (Au, 2009).

But the whole reason why Binet and Simon developed their test was to rank people from “best” to “worst”, “good” to “bad.” But, this does not mean that there is some “thing” inherent in individuals or groups that is being “measured” (Nash, 1990). Thus, since their inception, IQ tests (and by proxy all standardized testing) has pronouncements of such ranking built-in, even if it is not explicitly stated today. Such “measures” are not scientific and psychometrics is then shown for what it really is: “best understood as the development of tools for vertical classification and the production of social value” (Garrison, 2009: 5).

The goal, then, of psychometry is clear. Garrison (2009: 12) writes:

Ranking human worth on the basis of how well one competes in academic contests, with the effect that high ranks are associated with privilege, status, and power, suggests that psychometry is premised, not on knowledge of intellectual or emotional development, but on Anglo-American political ideals of rule by the best (most virtuous) and the brightest (most talented), a “natural aristocracy” in Jeffersonian parlance.

But, such notions of superiority and inferiority, as I have stated back in 2018, are nonsense when taken out of anatomic context:

It should be noted that the terms “superior” and “inferior” are nonsensical, when used outside of their anatomic contexts.

An IQ-ist may exclaim “Are you saying that you can’t say that person A has superior sprinting ability or breath-holding ability!? Are you denying that people are different?!” No, what I’m saying is that it is absurd to take anatomic terminology (physical measures) and attempt to liken it to IQ—this is because nothing physical is being measured, not least because the mental isn’t physical nor reducible to it.

They were presuming to measure one’s “intelligence” and then stating that one has ‘superior’ “intelligence” to another—and that IQ tests were measuring this “superiority”. However, pscyhometrics is not a form of measurement—rankings are not measures.

Knowledge becomes reducible to a score in regard to standardized testing, so students, and in effect their learning and knowledge, are then reduced to their scores on these tests. And so, “such inequalities [with the SAT, which holds for all standardized testing] are structured into the very foundations of standardized test construction itself” (Au, 2009: 64). So what is built into a test can also be built out of it (Richardson, 1990, 2000; Hilliard, 2012).

Measurement

In first constructing its scales and only then preceding to induce what they ‘measure’ from correlational studies, psychometry has got into the habit of trying to do what cannot be done and doing it the wrong way round anyway. (Nash, 1990: 133)

…psychometry fails to meet its claim of measurement and … its object is not the measurement of nonphysical human attributes, but the marking of some human beings as having more worth or value than other human beings … Psychometry’s claim to measurement serves to veil and justify the fundamentally political act of marking social value, and the role this practice plays in legitimating vast social inequalities. (Garrison, 2009: 30-31)

One of the best examples of a valid measure is temperature—and it has a long history (Chang, 2007). It is valid because there is a well-accepted theory of temperature, what is hot and what is cold. It is a physical property of measure which quantitatively expressed heat and cold. So thermometers were invented to quantify temperature, whereas “IQ” tests were invented to quantify “intelligence.” Those, like Jensen, attempt to make the analogy between temperature and IQ, thermometers and IQ tests. Thermometers, with a high degree of reliability, measure temperature and so do, Jensen, claims, IQ tests. But this, then, presumes that there is something “increasing”—presumably—physiologically in the brain, when there is no such evidence. So for this and many more reasons, the attempted comparison of temperature to intelligence, thermometers to intelligence tests, fails.

So, IQ-ists claim, temperature is measured by thermometers, by definition, therefore intelligence is what IQ tests measure, by definition. But there is a problem with claims such as this. Temperature was verified independently of the measuring device originally used to measure it. Fixed points were first established, and then numerical thermometers could be constructed in which we then find a procedure to assign numbers to degrees of heat between and beyond the fixed points. The thermoscope was what was used for the establishment of fixed points, The thermoscope has no fixed points, so we do not have to circularly rely on the concept of fixed points for reference. And if it goes up and down, we can then rightly infer that the temperature of blood is not stable. But what validates the thermoscope? Human sensation. We can see that when we put our hand into water that is scalding hot, if we put the thermoscope in the same water and note that it rises rapidly. So the thermoscopes agreement with our basic sensations of ‘hot’ and ‘cold’ serve as reliability for the fact that thermoscopes reliably justify (in a non-circular way) that temperature is truly being measured. We are trusting the physical sensation we get from whichever surface we are touching, and from this, we can infer that thermoscopes do indeed validate thermometers making the concept of temperature validated in a non-circular manner and a true measure of hot and cold. (See Chang, 2007 for a full discussion on the measurement of temperature.)

Thermometers could be tested by the criterion of comparability, whereas IQ tests, on the other hand, are “validated” circularly with tests of educational achievement, other IQ tests which were not themselves validated. and job performance (Howe, 1997; Richardson and Norgate, 2015; Richardson, 2017) which makes the “validation” circular since IQ tests and achievement tests are different versions of the same test (Schwartz, 1975).

For example, take intro chemistry. When one takes the intro course, they see how things are measured. Chemists may be measuring in mols, grams, the physical state of a substance, etc. We may measure water displacement, reactions between different chemicals or whatnot. And although chemistry does not reduce to physics, these are all actual physical measures.

But the same cannot be said for IQ (Nash, 1990). We can rightly say that one scores higher than another on an IQ tests but that does not signify that some “thing” is being measured and this is because, to use the temperature example again, there is no independent validation of the “construct.” IQ is a (latent) construct but temperature is a quantitative measure of hot and cold. It really exists, though the same cannot be said about IQ or “intelligence.” The concept of “intelligence” does not refer to something like weight and temperature, for example (Midgley, 2018).

Physical properties are observables. We observe the mercury in a thermometer change based on the temperature inside a building or outside. One may say that we observe “intelligence” daily, but that is NOT a “measure”, it’s just a descriptive claim. Blood pressure is another physical measure. It refers to the pressure in large arteries of the system. This is due to the heart pumping blood. An IQ-ist may say that intelligence is the emergent product of thinking and that this is due to the brain and that correlations between life outcomes, IQ tests and educational achievements then validate the measure. But, as noted above, this is circular. The two examples given—blood pressure and temperature—are real things that are physically measurable, unlike IQ (a latent construct).

It also should be noted that Eysenck claimed that if the measurement of temperature is scientific, then so is the measurement of intelligence. But thermometers are not identical to standardized scales. However, this claim fails, as Nash (1990: 131) notes:

In order to measure temperature three requirements are necessary: (i) a scale, (ii) some thermometric property of an object and, (iii) fixed points of reference. Zero temperature is defined theoretically and successive interval points are fixed by the physical properties of material objects. As Byerly (p. 379) notes, that ‘the length of a column of mercury is a thermometric property presupposes a lawful relationship between the order of length and the temperature order under certain conditions.’ It is precisely this lawful relationship which does not exist between the normative IQ scale and any property of intelligence.

This is where IQ-ists go the most wrong: the emphatically state that their tests are measuring SOMETHING! which is important for life success since they correlate with them. Though, there is no precise specification of the measured object, no object of measurement and no measurement unit, so this “means that the necessary conditions for metrication do not exist [for IQ]” (Nash, 1990: 145).

Since IQ tests have a scoring system, the general impressions is that IQ tests measure intelligence just like thermometers measure temperature—but this is a nonsense claim. IQ is an artifact of the test’s norming population. These points do not reflect any inherent property of individuals, they reflect one’s relation to the society they are in (since all standardized tests are proxies for social class).

Conclusion

One only needs to read into the history of IQ testing—and standardized testing as a whole—to see how and why these tests were first devised. From their beginnings wkth Binet and then over to Terman, Yerkes, and Goddard, the goal has been clear—enact eugenic policies on those deemed “unintelligent” by IQ tests which just so happen to correspond with lower classes in virtue of how the tests were constructed, which goes back originally to Binet and Simon. The history of the concept makes it clear that it’s not based on any kind of measurement theory like blood pressure and temperature. It is based on a priori notions of the structure and distribution of “intelligence” which then reproduces the social structure and “justifies” notions of superiority and inferiority on the basis of “intelligence tests” (Mensh and Mensh, 1991; Au, 2009; Garrison, 2009).

The attempts to hijack anatomic terminology, as I have shown, are nonsense since one doesn’t talk in other anatomic terminology about other kinds of things; the first IQ-ists’ intentions were explicit in what they were attempting to “show” which still holds for all standardized testing today.

Binet, Terman, Yerkes, Goddard and others all had their own priors which then led them to construct tests in such a way that would lead to their desired conclusions. No “property” is being “measured” by these tests, nor can they be used to show one’s “genetic standing” (Jensen, 1970) which implies that one is “genetically superior” (this can be justified by reading Jensen’s interview with American Renaissance and his comments on the “genetic enslavement” of a group of we continued our welfare policy).

Physiological measures, such as blood pressure, and measures of hot and cold, such as temperature, are valid measures and in no way, shape or form—contra Jensen—like the concept of IQ/”intelligence”, which Jensen conflates (Edwards, 1973). Intelligence (which is extra-physical) cannot be measured (see Berka, 1983 and see Nash, 1990: chapter 8 for a discussion of the measurement objection of Berka).

For these reasons, we should not claim that IQ tests ‘measure’ “intelligence”, nor do they measure one’s “genetic standing” or how “superior” one is to another and we should claim that psychometrics is nothing more than a political ring.

Race, Test Bias, and ‘IQ Measurement’

1800 words

The history of standardized testing—including IQ testing—has a contentious history. What causes score distributions between groups of people? I stated at least four reasons why there is a test gap:

(1) Differences in genes cause differences in IQ scores;

(2) Differences in environment cause differences in IQ scores;

(3) A combination of genes and environment cause differences in IQ scores; and

(4) Differences in IQ scores are built into the test based on the test constructors’ prior biases.

I hold to (4) since, as I have noted, the hereditarian-environmentalist debate is frivolous. There is, as I have been saying for years now, no agreed-upon definition of ‘intelligence’, since there are such disparate answers from the ‘experts’ (Lanz, 2000; Richardson, 2002).

For the lack of such a definition only reflects the fact that there is no worked-out theory of intelligence. Having a successful definition of intelligence without a corresponding theory would be like having a building without foundations. This lack of theory is also responsible for the lack of some principled regimentation of the very many uses the word ‘intelligence’ and its cognates are put to. Tao many questions concerning intelligence are still open, too many answers controversial. Consider a few examples of rather basic questions: Does ‘intelligence’ name some entity which underlies and explains certain classes of performances1, or is the word ‘intelligence’ only sort of a shorthand-description for ‘being good at a couple of tasks or tests’ (typically those used in IQ tests)? In other words: Is ‘intelligence’ primarily a descriptive or also an explanatorily useful term? Is there really something like intelligence or are there only different individual abilities (compare Deese 1993)? Or should we turn our backs on the noun ‘intelligence’ and focus on the adverb ‘intelligently’, used to characterize certain classes of behaviors? (Lanz, 2000: 20)

Nash (1990: 133-4) writes:

Always since there are just a series of tasks of one sort or another on which performance can be ranked and correlated with other performances. Some performances are defined as ‘cognitive performances’ and other performances as ‘attainment performances’ on essentially arbitrary, common sense grounds. Then, since ‘cognitive performances’ require ‘ability’ they are said to measure that ‘ability’. And, obviously, the more ‘cognitive ability’ an individual posesses the more that individual can acheive. These procedures can provide no evidence that IQ is or can be measured, and it is rather besides the point to look for any, since that IQ is a metric property is a fundamental assumption of IQ theory. It is imposible that any ‘evidence’ could be produced by such procedures. A standardised test score (whether on tests designated as IQ or attainment tests) obtained by an individual indicates the relative standing of that individual. A score lies within the top ten perent or bottom half, or whatever, of those  gained by the standardisation group. None of this demonstrates measurement of any property. People may be rank ordered by their telephone numbers but that would not indicate measurement of anything. IQ theory must demonstrate not that it has ranked people according to some performance (that requires no demonstration) but that they are ranked according to some real property revealed by that performance. If the test is an IQ test the property is IQ — by definition — and there can in consequence be no evidence dependent on measurement procedures for hypothesising its existence. The question is one of theory and meaning rather than one of technique. It is impossible to provide a satisfactory, that is non-circular, definition of the supposed ‘general cognitive ability’ IQ tests attempt to measure and without that definition IQ theory fails to meet the minimal conditions of measurement.

These is similar to Mary Midgley’s critique of ‘intelligence’ in her last book before her death What Is Philosophy For? (Midgley, 2018). The ‘definitions’ of ‘intelligence’ and, along with it, its ‘measurement’ have never been satisfactory. Haier (2016: 24) refers to Gottfredson’s ‘definition’ of ‘intelligence, stating that ‘intelligence’ is a ‘general mental ability.’ But if that is the case, that it is a ‘general mental ability’ (g) then ‘intelligence’ does not exist because ‘g’ does not exist as a property in the brain. Lanz’s (2000) critique is also like Howe’s (1988; 1997) that ‘intelligence’ is a descriptive, not explanatory, term.

Now that the concept of ‘intelligence’ has been covered, let’s turn to race and test bias.


Test items are biased when they have different psychological meanings across cultures (He and van de Vijver 2012: 7). If they have different meanings across cultures, then the tests will not reflect the same ‘ability’ between cultures. Being exposed to the knowledge—and correct usage of it—on a test is imperative for performance. For if one is not exposed to the content on the test, how are they expected to do well if they do not know the content? Indeed, there is much evidence that minority groups are not acculturated to the items on the test (Manly et al, 1997; Ryan et al, 2005; Boone et al, 2007). This is what IQ tests measure: acculturation to the the tests’ constructors, school cirriculum and school teachers—aspects of white, middle-class culture (Richardson, 1998). Ryan et al (2005) found that reading and and educational level, not race or ethnicity, was related to worse performance on psychological tests.

Serpell et al (2006) took 149 white and black fourth-graders and randomly assigned them to ethnically homogeneous groups of three, working on a motion task on a computer. Both blacks and whites learned equally well, but the transfer outcomes were better for blacks than for whites.

Helms (1992) claims that standardized tests are “Eurocentric”, which is “a perceptual set in which European and/ or European American values, customs, traditions and characteristics are used as exclusive standards against which people and events in the world are evaluated and perceived.” In her conclusion, she stated that “Acculturation
and assimilation to White Euro-American culture should enhance one’s performance on currently existing cognitive ability tests” (Helms, 1992: 1098). There just so happens to be evidence for this (along with the the studies referenced above).

Fagan and Holland (2002) showed that when exposure to different kinds of information was required, whites did better than blacks but when it was based on generally available knowledge, there was no difference between the groups. Fagan and Holland (2007) asked whites and blacks to solve problems found on usual IQ-type tests (e.g., standardized tests). Half of the items were solvable on the basis of available information, but the other items were solveable only on the basis of having acquired previous knowledge, which indicated test bais (Fagan and Holland, 2007). They, again, showed that when knowledge is equalized, so are IQ scores. Thus, cultural differences in information acquisition explain IQ scores. “There is no distinction between crassly biased IQ test items and those that appear to be non-biased” (Mensh and Mensh, 1991). This is because each item is chosen because it agrees with the distribution that the test constructors presuppose (Simon, 1997).

How do the neuropsychological studies referenced above along with Fagan and Holland’s studies show that test bias—and, along with it test construction—is built into the test which causes the distribution of the scores observed? Simple: Since the test constructors come from a higher social class, and the items chosen for inclusion on the test are more likely to be found in certain cultural groups than others, it follows that the reason for lower scores was that they were not exposed to the culturally-specific knowledge used on the test (Richardson, 2002; Hilliard, 2012).


The [IQ] tests do what their construction dictates; they correlate a group’s mental worth with its place in the social hierarchy. (Mensh and Mensh, 1991)

This is very easily seen with how such tests are constructed. The biases go back to the beginning of standardized testing—the first one being the SAT. The tests’ constructors had an idea of who was or was not ‘intelligent’ and so constructed the tests to show what they already ‘knew.’

…as one delves further … into test construction, one finds a maze of arbitrary steps taken to ensure that the items selected — the surrogates of intelligence — will rank children of different classes in conformity with a mental hierarchy that is presupposed to exist. (Mensh and Mensh, 1991)

Garrison (2009: 5) states that standardized tests “exist to assess social function” and that “Standardized testing—or the theory and practice known as “psychometrics” … is not a form of measurment.” The same way tests were constructed in the 1900s is the same way they are constructed today—with arbitrary items and a presuppossed mental hiearchy which then become baked into the tests by virtue of how they are constructed.

IQ-ists like to say that certain genes are associated with high intelligence (using their GWASes), but what could the argument possibly be that would show that variation in SNPs would cause variation in ‘intelligence’? What would a theory of that look like? How is the hereditarian hypothesis not a just-so story? Such tests were created to justify the hierarchies in society, the tests were constructed to give the results that they get. So, I don’t see how genetic ‘explanations’ are not just-so stories.

1 Blacks and whites are different cultural groups.

2 If (1), then they will have different experiences by virtue of being different cultural groups.is

3 So blacks and whites, being different cultural groups, will score differently on tests of ability, since they are exposed to different knowledge structures due to their different cultures and so, all tests of ability are culture-bound. Knowledge, Culture, Logic, and IQ

Rushton and Jensen (2005) claim that the evidence they review over the past 30 years of IQ testing points to a ‘genetic component’ to the black-white IQ gap, relying on the flawed Minnesota study of twinsreared apart” (Joseph, 2018)—among other methods—to generate heritability estimates and state that “The new evidence reviewed here points to some genetic component in Black–White differences in mean IQ.” The concept of heritability, however, is a flawed metric (Bailey, 1997; Schonemann, 1997; Guo, 2000; Moore, 2002; Rose, 2006; Schneider, 2007; Charney, 2012, 2013; Burt and Simons, 2014; Panofsky, 2014; Joseph et al, 2015; Moore and Shenk, 2016; Panofsky, 2016; Richardson, 2017). That G and E interact means that we cannot tease out “percentages” of nature and nurture’s “contribution” to a “trait.” So, one cannot point to heritability estimates as if they point to a “genetic cause” of the score gap between blacks and whites. Further note that the gap has closed in recent years (Dickens and Flynn, 2006; Smith, 2018).

And now, here is another argument based on the differing experiences that cultural groups experience which then explains IQ score differences (eg Mensh and Mensh, 1991; Manly et al, 1997; Kwate, 2001; Fagan and Holland, 2002, 2007; Cole, 2004; Ryan et al, 2005; Boone et al, 2007; Au, 2008; Hilliard, 2012; Au, 2013).

(1) If children of different class levels have experiences of different kinds with different material; and
(2) if IQ tests draw a disproportionate amount of test items from the higher classes; then
(c) higher class children should have higher scores than lower-class children.

Does Playing Violent Video Games Lead to Violent Behavior?

1400 words

President Trump was quoted the other day saying We have to look at the Internet because a lot of bad things are happening to young kids and young minds and their minds are being formed,” Trump said, according to a pool report, “and we have to do something about maybe what they’re seeing and how they’re seeing it. And also video games. I’m hearing more and more people say the level of violence on video games is really shaping young people’s thoughts.” But outside of broad assertions like this—that playing violent video games cause violent behavior—does it stack up to what the scientific literature says about it? In short, no, it does not. (A lot of publication bias exists in this debate, too.) Why do people think that violent video games cause violent behavior? Mostly due to the APA and their broad claims with little evidence.

Just doing a cursory Google search of ‘violence in video games pubmed‘ brings up 9 journal articles, so let’s take a look at a few of those.

The first article is titled The Effect of Online Violent Video Games on Levels of Aggression by Hollingdale and Greitemeyer (2014). They took 101 participants and randomized them to one of four experimental conditions: neutral, offline; neutral online; (Little Big Planet 2) violent offline; and violent online video games (Call of Duty: Modern Warfare). After they played said games, they answered a questionnaire and then measured aggression using the hot sauce paradigm (Lieberman et al, 1999) to measure aggressive behavior. Hollingdale and Greitemeyer (2014) conclude that “this study has identified that increases in aggression are not more pronounced when playing a violent video game online in comparison to playing a neutral video game online.”

Staude-Muller (2011) finds that “it was not the consumption of violent video games but rather an uncontrolled pattern of video game use that was associated with increasing aggressive tendencies.Przybylski, Ryan, and Rigby (2009) found that enjoyment, value, and desire to play in the future were strongly related to competence in the game. Players who were high in trait aggression, though, were more likely to prefer violent games, even though it didn’t add to their enjoyment of the game, while violent content lent little overall variance to the satisfactions previously cited.

Tear and Nielsen (2013) failed to find evidence that violent video game playing leads to a decrease in pro-social behavior (Szycik et al, 2017 also show that video games do not affect empathy). Gentile et al (2014) show that “habitual violent VGP increases long-term AB [aggressive behavior] by producing general changes in ACs [aggressive cognitions], and this occurs regardless of sex, age, initial aggressiveness, and parental involvement. These robust effects support the long-term predictions of social-cognitive theories of aggression and confirm that these effects generalize across culture.” The APA (2015) even states that “scientific research has demonstrated an association between violent video game use and both increases in aggressive behavior, aggressive affect, aggressive cognitions and decreases in prosocial behavior, empathy, and moral engagement.” How true is all of this, though? Does playing violent video games truly increase aggression/aggressive behavior? Does it have an effect on violence in America and shootings overall?

No.

Whitney (2015) states that the video-games-cause-violence paradigm has “weak support” (pg 11) and that, pretty much, we should be cautious before taking this “weak support” as conclusive. He concludes that there is not enough evidence to establish a truly causal connection between violent video game playing and violent and aggressive behavior. Cunningham, Engelstatter, and Ward (2016) tracked the sale of violent video games and criminal offenses after those games were sold. They found that violent crime actually decreased the weeks following the release of a violent game. Of course, this does not rule out any longer-term effects of violent game-playing, but in the short term, this is good evidence against the case of violent games causing violence. (Also see the PsychologyToday article on the matter.)

We seem to have a few problems here, though. How are we to untangle the effects of movies and other forms of violent media that children consume? You can’t. So the researcher(s) must assume that video games and only video games cause this type of aggression. I don’t even see how one can logically state that out of all other types of media that violent video games—and not violent movies, cartoons, TV shows etc—cause aggression/violent behavior.

Back in 2011, the Supreme Court case Brown vs. Entertainment Merchants Association concluding that since the effects on violent/aggressive behavior were so small and couldn’t be untangled from other so-called effects from other violent types of media. Ferguson (2015) found that violent video game playing had little effect on children’s mood, aggression levels, pro-social behavior or grades. He also found publication bias in this literature (Ferguson, 2017). Contrary to what those say about video games causing violence/aggressive behavior, video game playing was associated with a decrease in youth crime (Ferguson, 2014; Markey, Markey, and French, 2015 which is in line with Cunningham, Engelstatter, and Ward, 2016). You can read more about this in Ferguson’s article for The Conversationalong with his and others’ responses to the APA who state that violent video games cause violent behavior (with them stating that the APA is biased). (Also read a letter from 230 researchers on the bias in the APA’s Task Force on Violent Media.)

How would one actually untangle the effects of, say, violent video game playing and the effects of such other ‘problematic’ forms of media that also show aggression/aggressive acts towards others and actually pinpoint that violent video games are the culprit? That’s right, they can’t. How would you realistically control for the fact that the child grows up around—and consumes—so much ‘violent’ media, seeing others become violent around him etc; how can you logically state that the video games are the cause? Some may think it logical that someone who plays a game like, say, Call of Duty for hours on end a day would be more likely to be more violent/aggressive or more likely to commit such atrocities like school shootings. But none of these studies have ever come to the conclusion that violent video games may/will cause someone to kill or go on a shooting spree. It just doesn’t make sense. I can, of course, see the logic in believing that it would lead to aggressive behavior/lack of pro-social behavior (let’s say the kid played a lot of games and had little outside contact with people his age), but of course the literature on this subject should be enough to put claims like this to bed.

It’s just about impossible to untangle the so-called small effects of video games on violent/aggressive behavior from other types of media such as violent cartoons and violent movies. Who’s to say it’s not just the violent video games and not the violent movies and violent cartoons, too, that ’cause’ this type of behavior? It’s logically impossible to distinguish this, so therefore the small relationship between video games and violent behavior should be safely ignored. The media seems to be getting this right, which is a surprise (though I bet if Trump said the opposite—that violent video games didn’t cause violent behavior/shootings—that these same people would be saying that they do), but a broken clock is right twice a day.

So Trump’s claim (even if he didn’t outright state it) is wrong, along with anyone else who would want to jump in and attempt to say that video games cause violence. In fact, the literature shows a decrease in violence after games are released (Ferguson, 2014; Markey, Markey, and French, 2015; Cunningham, Engelstatter, and Ward, 2016). The amount of publication bias (also see Copenhaver and Ferguson, 2015 where they show how the APA ignores bias and methodological problems regarding these studies) in this field (Ferguson, 2017) should lead one to question the body of data we currently have, since studies that find an effect are more likely to get published than studies that find no effect.

Video games do not cause violent/aggressive behavior/school shootings. There is literally no evidence that they are linked to the deaths of individuals, and with the small effects noted on violent/aggressive behavior due to violent video game playing, we can disregard those claims. (One thing video games are good for, though, is improving reaction time (Benoit et al, 2017). The literature is strong here; playing these so-called “violent video games” such as Call of Duty improved children’s reaction time, so wouldn’t you say that these ‘violent video games’ have some utility?)

Delaying Gratification and Social Trust

1900 words

Tests of delayed gratification, such as the Marshmallow Experiment, show that those who can better delay their gratification have better life outcomes than those who cannot. The children who succumbed to eating the treat while the researcher was out of the room had worse life outcomes than the children who could wait. This was chalked up to cognitive processes by the originator of the test, while individual differences in these cognitive processes also were used as explanations for individual differences between children in the task. However, it doesn’t seem to be that simple. I did write an article back in December of 2015 on the Marshmallow Experiment and how it was a powerful predictor, but after extensive reading into the subject, my mind has changed. New research shows that social trust has a causal effect on whether or not one would wait for the reward—if the individual trusted the researcher he or she was more likely to wait for the other reward than if they did not trust the researcher, in which they were more likely to take what was offered in the first place.

The famous Marshmallow Experiment showed that children who could wait with a marshmallow or other treat in front of them while the researcher was out of the room, they would get an extra treat. The children who could not wait and ate the treat while the researcher was out of the room had worse life outcomes than the children who could wait for the other treat. These lead researchers to the conclusion that the ability to delay gratification depended on ‘hot’ and ‘cold’ cognitive processes. According to Walter Mischel, the originator of the study method, the ‘cool’ system is the thinking one, the cognitive system, which reminds you that you get a reward if you wait, while the ‘hot’ system is the impulsive system, the system that makes you want the treat now and not want to wait for the other treat (Metcalfe and Mischel, 1999).

Some of these participants were followed up on decades later, and those who could better delay their gratification had lower BMIs (Schlam et al, 2014); scored better on the SAT (Shoda, Mischel, and Peake, 1990) and other tests of educational attainment (Ayduk et al, 2000); along with other positive life outcomes. So it seems that placing a single treat—whether it be a marshmallow or another sweet treat—would predict one’s success, BMI, educational attainment and future prospects in life and that there are underlying cognitive processes, between individuals that lead to differences between them. But it’s not that simple.

After Mischel’s studies in the 50s, 60s and 70s on delayed gratification and positive and negative life outcomes (e.g., Mischel, 1958; Mischel, 1961Mischel, Ebbeson, and Zeiss, 1972) it was pretty much an accepted fact that delaying gratification somehow was related to these positive life outcomes, while the negative life outcomes were partly a result of the lack of ability to delay gratification. Though in 2014, a study was conducted showing that ability to delay gratification depends on social trust (Michaelson et al, 2013).

Using Amazon’s Mechanical Turk, (n = 78, 34 male, 39 female and 5 who preferred not to state their gender) completed online surveys and read three vignettes in order—trusty, untrustworthy and neutral—while using a scale of 1-7 to note how likeable, trustworthy, and how sharing their likelihood of sharing. Michaelson et al (2013) write:

Next, participants completed intertemporal choice questions (as in Kirby and Maraković, 1996), which varied in immediate reward values ($15–83), delayed reward values ($30–85), and length of delays (10–75 days). Each question was modified to mention an individual from one of the vignettes [e.g., “If (trustworthy individual) offered you $40 now or $65 in 70 days, which would you choose?”]. Participants completed 63 questions in total, with 21 different questions that occurred once with each vignette, interleaved in a single fixed but random order for all participants. The 21 choices were classified into 7 ranks (using the classification system from Kirby and Maraković, 1996), where higher ranks should yield higher likelihood of delaying, allowing a rough estimation of a subject’s willingness to delay using a small number of trials. Rewards were hypothetical, given that hypothetical and real rewards elicit equivalent behaviors (Madden et al., 2003) and brain activity (Bickel et al., 2009), and were preceded by instructions asking participants to consider each choice as if they would actually receive the option selected. Participants took as much time as they needed to complete the procedures.

When one’s trust was manipulated in the absence of a reward, within the group of subjects influenced their ability to delay gratification, along with how trustworthy one was perceived to be, influenced their ability to delay gratification. So this suggests that, in the absence of rewards, when social trust is reduced, ability to delay gratification would be lessened. Due to the issues of social trust manipulation due to the order of how the vignettes were read, they did a second experiment using the same model using 172 participants (65 males, 63 females, and 13 who chose not to state their gender). Though in this experiment, a computer-generated trustworthy, untrustworthy and neutral face was presented to the participants. They were only paid $.25 cents, though it has been shown that the compensation only affects turnout, not data quality (Burhmester, Kwang, and Gosling, 2011).

In this experiment, each participant read a vignette and there was a particular face attached to it (trustworthy, untrustworthy and neutral), which were used in previous studies on this matter. They found that when trust was manipulated in the absence of a reward between the subjects, this influenced the participants’ willingness and to delay gratification along with the perceived trustworthiness influencing it as well.

Michaelson et al (2013) conclude that the ability to delay gratification is predicated on social trust, and present an alternative hypothesis for all of these positive and negative life outcomes:

Social factors suggest intriguing alternative interpretations of prior findings on delay of gratification, and suggest new directions for intervention. For example, the struggles of certain populations, such as addicts, criminals, and youth, might reflect their reduced ability to trust that rewards will be delivered as promised. Such variations in trust might reflect experience (e.g., children have little control over whether parents will provide a promised toy) and predisposition (e.g., with genetic variations predicting trust; Krueger et al., 2012). Children show little change in their ability to delay gratification across the 2–5 years age range (Beck et al., 2011), despite dramatic improvements in self-control, indicating that other factors must be at work. The fact that delay of gratification at 4-years predicts successful outcomes years or decades later (Casey et al., 2011; Shoda et al., 1990) might reflect the importance of delaying gratification in other processes, or the importance of individual differences in trust from an early age (e.g., Kidd et al., 2012).

Another paper (small n, n = 28) showed that the children’s perception of the researchers’ reliability predicted delay of gratification (Kidd, Palmeri, and Aslin, 2012). They suggest that “children’s wait-times reflected reasoned beliefs about whether waiting would ultimately pay off.” So these tasks “may not only reflect differences in self-control abilities, but also beliefs about the stability of the world.” Children who had reliable interactions with the researcher waited about 4 times as long—12 minutes compared to 3 minutes—if they thought the researcher was trustworthy. Sean Last over at the Alternative Hypothesis uses these types of tasks (and other correlates) to show that blacks have lower self-control than whites, citing studies showing correlations with IQ and delay of gratification. Though, as can be seen, alternative explanations for these phenomena make just as much sense, and with the new experimental evidence on social trust and delaying gratification, this adds a new wrinkle to this debate. (He also shortly discusses ‘reasons’ why blacks have lower self-control, implicating the MAOA alleles. However, I have already discussed this and blaming ‘genes for’ violence/self-control doesn’t make sense.)

Michaelson and Munakata (2016) show more evidence for the relationship between social trust and delaying gratification. When children (age 4 years, 5 months, n = 34) observed an adult as trustworthy, they were able to wait for the reward, compared to when they observed the adult as untrustworthy they ate the treat thinking that, since they observed the adult as untrustworthy, they were not likely to get the second marshmallow than if they waited for the adult to return if they believed him to be untrustworthy. Ma et al (2018) also replicated these findings in a sample of 150 Chinese children aged 3 to 5 years old. They conclude that “there is more to delay of gratification than cognitive capacity, and they suggest that there are individual differences in whether children consider sacrificing for a future outcome to be worth the risk.” Those who had higher levels of generalized trust waited longer, even when age and level of executive functioning were controlled for.

Romer et al (2010) show that people who are more willing to take risks may be more likely to engage in risky behavior that provides insights to that specific individual on why delaying gratification and having patience leads to longer-term rewards. This is a case of social learning. However, people who are more willing to take risks have higher IQs than people who do not. Though SES was not controlled for, it is possible that the ability to delay gratification in this study came down to SES, with lower class people taking the money, while higher class people deferred. Raine et al (2002) showed a relationship between sensation seeking in 3-year-old children from Mauritius, which then was related to their ‘cognitive scores’ at age 11. As usual, parental occupation was used as a measure of ‘social class’, and since SES does not capture all aspects of social class then controlling for the variable does not seem to be too useful. Because a confound here could be that children from higher classes have more of a chance to sensation seek which may cause higher IQ scores due to cognitive enrichment. Either way, you can’t say that IQ ’causes’ delayed gratification since there are more robust predictors such as social trust.

Though the relationship is there, what to make of it? Since exploring more leads to, theoretically, more chances to get things wrong and take risks by being impulsive, those who are more open to experience will have had more chances to learn from their impulsivity, and so learn to delay gratification through social learning and being more open. ‘IQ’ correlating with it, in my opinion, doesn’t matter too much; it just shows that there is a social learning component to delaying gratification.

In conclusion, there are alternative ways to look at the results from Marshmallow Experiments, such as social trust and social learning (being impulsive and seeing what occurs when an impulsive act is carried out may have one learn, in the future, to wait for something). Though these experiments are new and the research is young, it’s very promising that there are other explanations for delayed gratification that don’t have to do with differences in ‘cognitive ability’, but depend on social trust—trust between the child and the researcher. If the child sees the researcher is trustworthy, then the child will wait for the reward, whereas if they see the researcher is not trustworthy, they ill take the marshmallow or whatnot, since they believe the researcher is not trustworthy and therefore won’t stick to their word. (I am also currently reading Mischel’s 2014 book Marshmallow Test: Mastering Self-Control and will have more thoughts on this in the future.)

Race and Medicine: Is Race a Useful Category?

2450 words

The New York Times published an article on December the 8th titled What Doctors Should Ignore: Science has revealed how arbitrary racial categories are. Perhaps medicine will abandon them, too. It is an interesting article and while I do not agree with all of it, I do agree with some.

It starts off by talking about sickle cell anemia (SCA) and how was once thought of as a ‘black disease’ because blacks were, it seemed, the only ones who were getting the disease. I recall back in high-school having a Sicilian friend who said he ‘was black’ because Sicilians can get SCA which is ‘a black disease’, and this indicates ‘black genes’. However, when I grew up and actually learned a bit about race I learned that it was much more nuanced than that and that whether or not a population has SCA is not based on race, but is based on the climate/environment of the area which would breed mosquitoes which carry malaria. SCA still, to this day, remains a selective factor in the evolution of humans; malaria selects for the sickle cell trait (Elguero et al, 2015).

This is a good point brought up by the article: the assumption that SCA was a ‘black disease’ had us look over numerous non-blacks who had the sickle cell trait and could get the help they needed, when they were overlooked due to their race with the assumption that they did not have this so-called ‘black disease’. Though it is understandable why it got labeled ‘a black disease’; malaria is more prevalent near to the equator and people whose ancestors evolved there are more likely to carry the trait. In regards to SCA, it should be known that blacks are more likely to get SCA, but just because someone is black does not automatically mean that it is a foregone conclusion that one has the disease.

The article then goes on to state that the push to excise race from medicine may undermine a ‘social justice concept’: that is, the want to rid the medical establishment of so-called ‘unconscious bias’ that doctors have when dealing with minorities. Of course, I will not discount that this doesn’t have some effect—however small—on racial health disparities but I do not believe that the scope of the matter is as large as it is claimed to be. This is now causing medical professionals to integrate ‘unconscious bias training’, in the hopes of ridding doctors of bias—whether conscious or not—in the hopes to ameliorate racial health disparities. Maybe it will work, maybe it will not, but what I do know is that if you know someone’s race, you can use it as a roadmap to what diseases they may or may not have, what they may or may not be susceptible to and so on. Of course, only relying on one’s race as a single data point when you’re assessing someone’s possible health risks makes no sense at all.

The author then goes on to write that the terms ‘Negroid, Caucasoid, and Mongoloid’ were revealed as ‘arbitrary’ by modern genetic science. I wouldn’t say that; I would say, though, that modern genetic science has shown us the true extent of human variation, while also showing that humans cluster into 5 distinct geographic categories, which we can call ‘race’ (Rosenberg et al, 2002; but see Wills, 2017 for alternative view that the clusters identified by Rosenberg et al, 2002 are not races. I will cover this in the future). The author then, of course, goes on to use the continuum fallacy stating that since “there are few sharp divides where one set of traits ends and another begins“. A basic rebuttal would be, can you point out where red and orange are distinct? How about violet and blue? Blue and Cyan? Yellow and orange? When people commit the continuum fallacy then the only logical conclusion is that if races don’t exist because there are “few sharp divides where one set of traits ends and another begins“, then, logically speaking, colors don’t exist either because there are ‘few [if any] sharp divides‘ where one color ends and another begins.

colors

The author also cites geneticist Sarah Tishkoff who states that the human species is too young to have races as we define them. This is not true, as I have covered numerous times. The author then cites this study (Ng et al, 2008) in which Craig Venter’s genome was matched with the (in)famous [I love Watson] James Watson and focused on six genes that had to do with how people respond to antipsychotics, antidepressants, and other drugs. It was discovered that Venter had two of the ‘Caucasian’ variants whereas Watson carried variants more common in East Asians. Watson would have gotten the wrong medicine based on the assumption of his race and not on the predictive power of his own personal genome.

The author then talks about kidney disease and the fact that blacks are more likely to have it (Martins, Agodoa, and Norris, 2012). It was assumed that environmental factors caused the disparity of kidney disease in blacks when compared to whites, however then the APOL1 gene variant was discovered, which is related to worse kidney outcomes and is in higher frequencies in black Americans, even in blacks with well-controlled blood pressure (BP) (Parsa et al, 2013). The author then discusses that black kidneys were seen as ‘more prone to failure’ than white kidneys, but this is, so it’s said, due to that one specific gene variant and so, race shouldn’t be looked at in regards to kidney disease but individual genetic variation.

In one aspect of the medical community can using medicine based on one’s race help: prostate cancer. Black men are more likely to be afflicted with prostate cancer in comparison to whites (Odedina et al, 2009; Bhardwaj et al, 2017) with it even being proposed that black men should get separate prostate screenings to save more lives (Shenoy et al, 2016). Then he writes that we still don’t know the genes responsible, however, I have argued in the past that diet explains a large amount—if not all of the variance. (It’s not testosterone that causes it like Ross et al, 1986 believe).

The author then discusses another medical professional who argues that racial health disparities come down to the social environment. Things like BP could—most definitely—be driven by the social environment. It is assumed that the darker one’s skin is, the higher chance they have to have high BP—though this is not the case for Africans in Africa so this is clearly an American-only problem. I could conjure up one explanation: the darker the individual, the more likely he is to believe he is being ‘pre-judged’ which then affects his state of mind and has his BP rise. I discussed this shortly in my previous article Black-White Differences in PhysiologyWilliams (1992) reviewed evidence that social, not genetic, factors are responsible for BP differences between blacks and whites. He reviews one study showing that BP is higher in lower SES, darker-skinned blacks in comparison to higher SES blacks whereas for blacks with higher SES no effect was noticed (Klag et al, 1991). Sweet et al (2007) showed that for lighter-skinned blacks, as SES rose BP decreased while for darker-skinned blacks BP increased as SES did while implicating factors like ‘racism’ as the ultimate causes.

There is evidence for the effect of psychosocial factors and BP (Marmot, 1985). In a 2014 review of the literature, Cuffee et al (2014) identify less sleep—along with other psychosocial factors—as another cause of higher BP. It just so happens that blacks average about one hour of sleep less than whites. This could cause a lot of the variation in BP differences between the races, so clearly in the case of this variable, it is useful to know one’s race, along with their SES. Keep in mind that any actual ‘racism’ doesn’t have to occur; the person only ‘needs to perceive it’, and their blood BP will rise in response to the perceived ‘racism’ (Krieger and Sidney, 1996). Harburg et al (1978) write in regards to Detroit blacks:

For 35 blacks whose fathers were from the West Indies, pressures were higher than those with American-born fathers. These findings suggest that varied gene mixtures may be related to blood pressure levels and that skin color, an indicator of possible metabolic significance, combines with socially induced stress to induce higher blood pressures in lower class American blacks.

Langford (1981) shows that when SES differences are taken into account that the black-white BP disparity vanishes. So there seems to be good evidence for the hypothesis that psychosocial factors, sleep deprivation, diet and ‘perceived discrimination’ (whether real or imagined) can explain a lot of this gap so race and SES need to be looked at when BP is taken into account. These things are easily changeable; educate people on good diets, teach people that, in most cases, no, people are not being ‘racist’ against you. That’s really what it is. This effect holds more for darker-skinned, lower-class blacks. And while I don’t deny a small part of this could be due to genetic factors, the physiology of the heart and how BP is regulated by even perceptions is pretty powerful and could have a lot of explanatory power for numerous physiological differences between races and ethnic groups.

Krieger (1990) states that in black women—not in white women—“internalized response to unfair treatment, plus non-reporting of race and gender discrimination, may constitute risk factors for high blood pressure among black women“. This could come into play in regards to black-white female differences in BP. Thomson and Lip (2005) show that “environmental influence and psychosocial factors may play a more important role than is widely accepted” in hypertension but “There remain many uncertainties to the relative importance and contribution of environmental versus genetic influences on the development of blood pressure – there is more than likely an influence from both. However, there is now evidence to necessitate increased attention in examining the non-genetic influences on blood pressure …” With how our physiology evolved to respond to environmental stimuli and respond in real time to perceived threats, it is no wonder that these types of ‘perceived discrimination’ causes higher BP in certain groups with lower SES.

Wilson (1988) implicates salt as the reason why blacks have higher BP than whites. High salt intake could affect the body’s metabolism by causing salt retention which influences blood plasma volume, cardiac output. However, whites have a higher salt intake than blacks, but blacks still ate twice the recommended amounts from the dietary guidelines (all ethnic subgroups they analyzed from America over-consumed salt as well) (Fulgoni et al, 2014). Blacks are also more ‘salt-sensitive’ than whites (Sowers et al 1988Schmidlin et al, 2009; Sanada, Jones, and Jose, 2014) which is also heritable in blacks (Svetke, McKeown, and Wilson, 1996). A slavery hypothesis does exist to explain higher rates of hypertension in blacks, citing salt deficiency in the parts of Africa that supplied the slaves to the Americas, to the trauma of the slave trade and slavery in America. However, historical evidence does not show this to be the case because “There is no evidence that diet or the resulting patterns of disease and demography among slaves in the American South were significantly different from those of other poor southerners” (Curtin, 1992) whereas Campese (1996) hypothesizes that blacks are more likely to get hypertension because they evolved in an area with low salt.

The NYT article concludes:

Science seeks to categorize nature, to sort it into discrete groupings to better understand it. That is one way to comprehend the race concept: as an honest scientific attempt to understand human variation. The problem is, the concept is imprecise. It has repeatedly slid toward pseudoscience and has become a major divider of humanity. Now, at a time when we desperately need ways to come together, there are scientists — intellectual descendants of the very people who helped give us the race concept — who want to retire it.

Race is a useful concept. Whether in medicine, population genetics, psychology, evolution, physiology, etc it can elucidate a lot of causes for differences between races and ethnic groups—whether or not they are genetic or psychosocial in nature. That just attests to both the power of suggestion along with psychosocial factors in regards to racial differences in physiological factors.

Finally let’s see what the literature says about race in medicine. Bonham et al (2009) showed that both black and white doctors concluded that race is medically relevant but couldn’t decide why however they did state that genetics did not explain most of the disparity in relation to race and disease aside from the obvious disorders like Tay Sachs and sickle cell anemia. Philosophers accept the usefulness of race in the biomedical sciences (Andreason, 2009Efstathiou, 2012; Hardimon, 2013Winther, Millstein, and Nielsen, 2015; Hardimon, 2017) whereas Risch et al (2002) and Tang et al (2002) concur that race is useful in the biomedical sciences. (See also Dorothy Roberts’ Ted Talk The problem with race-based medicine which I will cover in the future). Richard Lewontin, naturally, has hang-ups here but his contentions are taken care of above. Even if race were a ‘social construct‘, as Lewontin says, it would still be useful in a biomedical sense; but since there are differences between races/ethnic groups then they most definitely are useful in a biomedical sense, even if at the end of the day individual variation matters more than racial variation. Just knowing someone’s race and SES, for instance, can tell you a lot about possible maladies they may have, even if, utltimately, individual differences in physiology and anatomy matter more in regards to the biomedical context.

In conclusion, race is most definitely a useful concept in medicine, whether race is a ‘social construct’ or not. Just using Michael Hardimon’s race concepts, for instance, shows that race is extremely useful in the biomedical context, despite what naysayers may say. Yes, individual differences in anatomy and physiology trump racial differences, but just knowing a few things like race and SES can tell a lot about a particular person, for instance with blood pressure, resting metabolic rate, and so on. Denying that race is a useful concept in the biomedical sciences will lead to more—not less—racial health disparities, which is ironic because that’s exactly what race-deniers do not want. They will have to accept a race concept, and they would accept Hardimon’s socialrace concept because that still allows it to be a ‘social construct’ while acknowledging that race and psychosocial factors interact to cause higher physiological variables. Race is a useful concept in medicine, and if the medical establishment wants to save more lives and actually end the racial disparities in health then they should acknowledge the reality of race.