Answer the following questions based on the enclosed (attached) readings:
1. What is the issue the author is concerned with and the main point the authors are trying to make about that issue? What are the reasons the authors give for thinking that this point is worth serious consideration?
2. What do they agree or disagree about?
3. Identify an example from the news, controversial issue, or personal experience that the reading(s) made you think about. How does it relate to the reading and/or illuminate the issues the reading raises?
4. What do you find interesting or confusing about the reading(s)?
5. How do you think, in the light of the evidence presented, intuitive/affective and rational processes interact in the ways people make moral judgments and decisions?
6. What insights about the structure and social functions of moral judgments, can we learn from their similarities with the structure (grammar) and social functions of language?
Multi‐system Moral Psychology
Page 1 of 26
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
The Moral Psychology Handbook
John M. Doris and The Moral Psychology Research Group
Print publication date: 2010
Print ISBN-13: 9780199582143
Published to Oxford Scholarship Online: September 2010
DOI: 10.1093/acprof:oso/9780199582143.001.0001
Multi‐system Moral Psychology
Fiery Cushman
Liane Young
Joshua D. Greene
DOI:10.1093/acprof:oso/9780199582143.003.0003
This chapter argues for a synthesis of two recent projects in moral psychology.
One has proposed a division between “emotional” versus “cognitive” moral
judgments, while another has proposed a division between “automatic” versus
“controlled” moral judgments. These appear to describe the same underlying
psychological systems: one that appears to give rise to automatic, rapid, and
emotionally forceful moral intuitions, and another that appears to use controlled,
effortful cognition to apply explicit moral principles. These psychological
systems can explain a lot about the basic form of competing philosophical
normative theories. Some core philosophical deontic principles are mirrored in
ordinary people’s emotional intuitions; meanwhile, people often use controlled
cognitive processes to think about moral problems in utilitarian terms. The
chapter concludes by arguing that this division is not hard-and-fast. There is
evidence for reasoning from explicit deontic principles in ordinary people;
meanwhile, utilitarian thought must depend on some underlying affective
currency. This analysis is used to motivate several new questions facing the field
of moral psychology.
Keywords: dual process theory, emotion, cognition, utilitarianism, deontology, intuition, reasoning,
moral judgment
In the field of moral psychology, a number of theoretical proposals that were at
one time regarded as unconnected at best—and, at worst, contradictory—are
showing signs of reconciliation. At the core of this emerging consensus is a
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143
https://www.universitypressscholarship.com/search?f_0=keywords&q_0=dual process theory
https://www.universitypressscholarship.com/search?f_0=keywords&q_0=emotion
https://www.universitypressscholarship.com/search?f_0=keywords&q_0=cognition
https://www.universitypressscholarship.com/search?f_0=keywords&q_0=utilitarianism
https://www.universitypressscholarship.com/search?f_0=keywords&q_0=deontology
https://www.universitypressscholarship.com/search?f_0=keywords&q_0=intuition
https://www.universitypressscholarship.com/search?f_0=keywords&q_0=reasoning
https://www.universitypressscholarship.com/search?f_0=keywords&q_0=moral judgment
Multi‐system Moral Psychology
Page 2 of 26
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
recognition that moral judgment is the product of interaction and competition
between distinct psychological systems. Our goal is to describe these systems
and to highlight directions for further inquiry.
Recent research in moral psychology has focused on two challenges to the long‐
dominant cognitive development paradigm conceived by Piaget and nurtured by
Kohlberg (Kohlberg, 1969; Piaget, 1965/1932; Turiel, 1983, 2005). The first
challenge claims that moral judgment takes the form of intuition, accomplished
by rapid, automatic, and unconscious psychological processes (Haidt, 2001;
Hauser, 2006: 1; Mikhail, 2000; Schweder & Haidt, 1993; see also Damasio,
1994: 165–201), contra the cognitive developmentalists’ assumption that moral
judgment is the product of conscious, effortful reasoning. A central motivation
for this challenge comes from studies demonstrating people’s inability to
articulate a rational basis for many strongly held moral convictions (Haidt, 2001;
Cushman, Young, & Hauser, 2006; Hauser, Cushman, Young, Jin, & Mikhail,
2007; Mikhail, 2000). The second and related challenge claims that moral
judgment is driven primarily by affective responses (Blair, 1995; Damasio, 1994;
Greene & Haidt, 2002; Schweder & Haidt, 1993), contra the cognitive
developmentalists’ focus on the deliberate application of explicit moral theories
and principles to particular cases. Evidence for the role of affect is largely
neuroscientific (Shaich Borg, Hynes, Van Horn, Grafton, & Sinnott‐Armstrong,
2006; Ciaramelli, Muccioli, Ladavas, & di Pellegrino, 2007; Damasio, 1994;
Greene, Nystrom, Engell, Darley, & Cohen, 2004; Greene, Sommerville,
Nystrom, Darley, & Cohen, 2001; Koenigs et al., 2007; Mendez, Anderson, &
Shapria, 2005), but also includes behavioral studies of moral judgment using
affective manipulations (Valdesolo (p.48) & DeSteno, 2006; Schnall, Haidt,
Clore, & Jordan, 2008; Wheatley & Haidt, 2005).
The evidence that moral judgment is driven largely by intuitive emotional
responses is strong, but it does not follow from this that emotional intuition is
the whole story. Concerning the role of intuition, the research of Kohlberg and
others indicates a truly astonishing regularity in the development of explicit
moral theories and their application to particular dilemmas (Kohlberg, 1969).
Recent studies show that while people cannot offer principled justifications for
some of their moral judgments, they are quite able to do so for others (Cushman
et al., 2006), and that people alter some moral judgments when asked to engage
in conscious reasoning (Pizarro, Uhlmann, & Bloom, 2003) or presented with
appropriate opportunities (Lombrozo, 2008). Others studies implicate reasoning
processes in moral judgment using brain imaging (Greene et al., 2004) and
reaction time data (Greene et al., 2008). Together, these studies seem to capture
an important and relatively common experience: deliberation about right and
wrong, informed by an awareness of one’s explicit moral commitments.
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-203
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-213
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-222
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-223
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-189
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-193
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-209
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-217
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-177
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-189
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-193
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-194
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-209
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-171
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-177
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-184
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-217
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-215
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-173
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-177
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-187
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-188
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-202
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-208
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-224
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-216
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-226
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-203
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-175
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-214
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-205
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-187
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-186
Multi‐system Moral Psychology
Page 3 of 26
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
In fact, the claim that moral judgment depends on affective responses at all has
been met with skepticism, for instance by champions of “universal moral
grammar” (Huebner, Dwyer, & Hauser, 2008; Hauser, 2006; Mikhail, 2000). They
observe that moral judgment requires computations performed over
representations of agents, intentions, and causal relationships in order to output
judgments of “right” and “wrong.” Information processing of this sort
necessarily precedes an affective response, the argument goes, and thus it must
be that moral judgment leads to affect, not the other way around.
Reconciling these apparent alternatives—intuitive versus rational,1 affective
versus cognitive2—is therefore a focal point of research. In our view, the most
successful attempts share a common insight: moral judgment is accomplished by
multiple systems. Here, we pursue a dual‐process approach in which moral
judgment is the product of both intuitive and rational psychological processes,
and it is the product of what are conventionally thought of as “affective” and
“cognitive” mechanisms. As we shall see, a dual‐process model of moral
judgment can explain features of the data that (p.49) unitary models cannot:
dissociations in clinical populations, cognitive conflict in healthy individuals, and
so on.
The new challenge we face is to understand the specific features of distinct
systems and the processes of integration and competition among them. In this
chapter, we begin by reviewing the evidence in favor of a division between a
cognitive system and an affective system for moral judgment. Next, we argue
that the cognitive system operates by “controlled” psychological processes
whereby explicit principles are consciously applied, while affective responses
are generated by “automatic” psychological processes that are not available to
conscious reflection. Thus, we suggest, the cognitive/affective and conscious/
intuitive divisions that have been made in the literature in fact pick out the same
underlying structure within the moral mind. Finally, we consider a set of
Humean hypotheses according to which moral principles deployed in conscious
cognition have affective origins.
The present chapter focuses exclusively on two psychological systems that shape
moral judgments concerning physically harmful behavior. Psychological
researchers have often noted, however, that the moral domain encompasses
much more than reactions to and prohibitions against causing bodily harm
(Darley & Shultz, 1990; Gilligan, 1982/1993: 19; Haidt, 2007; Schweder & Haidt,
1993). Other sub‐domains of morality might include the fair allocation of
resources, sexual deviance, altruism and care, respect for social hierarchy, and
religious devotion, for instance. Several authors have suggested that
independent psychological systems are responsible for judgments in several of
these domains, and we regard such conjectures as plausible. Our focus on two
systems that are important for judgments concerning harm is not presented as a
complete account of moral psychology. On the contrary, we hope that the dual‐
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-195
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-193
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-209
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-178
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-181
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-190
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-217
Multi‐system Moral Psychology
Page 4 of 26
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
process model explored here will eventually be understood as part of a larger
constellation of psychological systems that enable the human capacity for moral
judgment.
Moral dilemmas come in many flavors, and a perennial favorite in moral
philosophy forces a choice between harming one person and letting many people
die, as in the classic trolley dilemmas (Foot, 1967; Thomson, 1985). In a case
that we shall call the switch dilemma, a runaway trolley threatens to run over
and kill five people. Is it morally permissible to flip a switch that will redirect the
trolley away from five people and onto one person instead, (p.50) thus saving
five lives at the cost of one? In a large web survey, most people said that it is
(Hauser et al., 2007.) This case contrasts with the footbridge dilemma. Here, one
person is standing next to a larger person on a footbridge spanning the tracks,
in between the oncoming trolley and the five. In this case, the only way to save
the five is to push the large person off of the footbridge and into the trolley’s
path, killing him, but preventing the trolley from killing the five. (You can’t stop
the trolley yourself because you’re not big enough to do the job.) Most people
surveyed said that in this case trading one life for five is not morally permissible.
It appears that most ordinary people, like many philosophers, endorse a
characteristically consequentialist judgment (“maximize the number of lives
saved”) in first case and a characteristically deontological judgment (“harm is
wrong, no matter what the consequences are”) in the second. (For a discussion
of why we consider it legitimate to refer to these judgments as
“characteristically deontological” and “characteristically consequentialist,” see
Greene, 2007.) This pair of dilemmas gives rise to the “trolley problem,” which,
for decades, philosophers have attempted to solve (Fischer & Ravizza, 1992;
Kamm, 1998, 2006). What makes people judge these cases differently?
Greene and colleagues’ dual‐process theory of moral judgment (Greene et al.,
2001, 2004; Greene, 2007) attempted to characterize the respective roles of
affect and controlled cognition in people’s responses to these dilemmas (Greene
et al., 2001). Specifically, Greene and colleagues proposed that the thought of
harming someone in a “personal” way, as in the footbridge dilemma, triggers a
negative emotional responses that says, in effect, “That’s wrong, don’t do it!”
According to their theory, this emotional alarm bell overrides any
consequentialist inclination to approve of the five‐for‐one trade off. In contrast,
people tend to say that redirecting the trolley in the switch case is morally
permissible because the “impersonal” nature of this action fails to trigger a
comparable emotional response. In the absence of such a response,
consequentialist moral reasoning (“Five lives are worth more than one”)
dominates the decision. The core insight of this proposal is that our discrepant
judgments in the switch and push cases may not come from the operation of a
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-180
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-221
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-194
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-183
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-179
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-197
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-198
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-188
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-187
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-183
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-188
Multi‐system Moral Psychology
Page 5 of 26
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
single psychological system, but rather from competition between two distinct
psychological systems.
Putting this proposal to the empirical test, Greene and colleagues examined the
neural activity of people responding to various “personal” and “impersonal”
moral dilemmas. As predicted, they found that brain regions associated with
emotion (and social cognition more broadly) exhibited increased activity in
response to “personal” moral dilemmas such as the footbridge case. These brain
regions included a region of the medial prefrontal cortex (Brodmann’s (p.51)
area 9/10) that was damaged in the famous case of Phineas Gage, the
nineteenth‐century railroad foreman whose moral behavior became severely
disordered after a tragic accident sent a metal tamping iron through his eye
socket and out of the top of his head (Damasio, 1994: 3; Macmillan, 2000). In
contrast, and also as predicted, Greene and colleagues found that brain regions
associated with controlled cognitive processes such as working memory and
abstract reasoning exhibited increased activity when people were responded to
“impersonal” moral dilemmas such as the switch case.
Building on this finding, Greene and colleagues conducted a second study
(Greene et al., 2004). Whereas their first study identified patterns of neural
activity associated with the kind of dilemma in question (“personal” vs.
“impersonal”), the second study identified patterns of neural activity associated
with the particular judgments people made (“acceptable” vs. “unacceptable”).
They focused their analysis on difficult dilemmas in which harming someone in a
“personal” manner would lead to a greater good. Here is an example of a
particularly difficult case, known as the crying baby dilemma:
Enemy soldiers have taken over your village. They have orders to kill all
remaining civilians. You and some of your townspeople have sought refuge
in the cellar of a large house. Outside, you hear the voices of soldiers who
have come to search the house for valuables. Your baby begins to cry
loudly. You cover his mouth to block the sound. If you remove your hand
from his mouth, his crying will summon the attention of the soldiers who
will kill you, your child, and the others hiding out in the cellar. To save
yourself and the others, you must smother your child to death. Is it
appropriate for you to smother your child in order to save yourself and the
other townspeople?
Subjects tend to take a long time to respond to this dilemma, and their
judgments tend to split fairly evenly between the characteristically
consequentialist judgment (“Smother the baby to save the group”) and the
characteristically deontological judgment (“Don’t smother the baby”). According
to Greene and colleagues’ dual‐process theory, the characteristically
deontological responses to such cases are driven by prepotent emotional
responses that nearly everyone has. If that’s correct, then people who deliver
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-177
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-207
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-187
Multi‐system Moral Psychology
Page 6 of 26
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
characteristically consequentialist judgments in response to such cases must
override their emotional responses.
This theory makes two predictions about what we should see in people’s brains
as they respond to such dilemmas. First, we would expect to see increased
activity in a part of the brain called the anterior cingulate cortex (ACC), which,
in more dorsal subregions, reliably responds when two or more incompatible
behavioral responses are simultaneously activated, i.e. under conditions of
“response conflict” (Botvinick, Braver, Barch, Carter, & Cohen, 2001). For
example, if one is presented with the word “red” written in green (p.52) ink,
and one’s task is to name the color of the ink, then one is likely to experience
response conflict because the more automatic word‐reading response (“red”)
conflicts with the task‐appropriate color‐naming response (“green”) (Stroop,
1935). According to this dual‐process theory, the aforementioned difficult
dilemmas elicit an internal conflict between a prepotent emotional response that
says “No!” and a consequentialist cost–benefit analysis that says “Yes.”
Consistent with this theory, Greene and colleagues found that difficult
“personal” dilemmas like the crying‐baby case elicit increased ACC activity,
relative to easier “personal” dilemmas, such as whether to kill your boss because
you and others don’t like him, in which reaction times are shorter and judgments
are more overwhelmingly negative. Second, we would expect to see increased
activity in a part of the brain known as the dorsolateral prefrontal cortex
(DLPFC). This part of the brain is the seat of “cognitive control” (Miller &
Cohen, 2001) and is necessary for overriding impulses and for “executive
function” more broadly. Once again, if the characteristically deontological
judgment is based on an intuitive emotional response, then giving a
characteristically consequentialist response requires overriding that response, a
job for the DLPFC. As predicted, Greene and colleagues found that
consequentialist judgments in response to difficult “personal” dilemmas
(“Smother the baby in the name of the greater good”) are associated with
increased activity in the DLPFC relative to the activity associated with trials on
which deontological judgments were made.
These neuroimaging results support a dual‐process theory of moral judgment in
which distinct “cognitive” and emotional processes sometimes compete. But a
consistent association between neural activity and behavior cannot provide
conclusive evidence that the neural activity is a cause of the behavior. To provide
stronger evidence for such causal relationships, one must intervene on the
neural process. In a recent study, Greene and colleagues (2008) did this by
imposing a “cognitive load” on people responding to difficult “personal” moral
dilemmas like the crying‐baby dilemma. People responded to the moral
dilemmas while simultaneously monitoring a string of digits scrolling across the
screen. The purpose of this manipulation is to disrupt the kind of controlled
cognitive processes that are hypothesized to support consequentialist moral
judgments. They found, as predicted, that imposing a cognitive load slowed
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-172
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-219
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-211
Multi‐system Moral Psychology
Page 7 of 26
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
down characteristically consequentialist judgments, but had no effect on
characteristically deontological judgments. (Deontological judgments were in
fact slightly faster under cognitive load, but this effect was not statistically
significant.)
A complementary tactic is to manipulate the relevant emotional responses,
rather than the capacity for cognitive control. Valdesolo and DeSteno (2006) (p.
53) did this by presenting either comedic video clips or affectively neutral video
clips to two groups of subjects who then responded to versions of the switch and
footbridge dilemmas. They reasoned as follows: if people judge against pushing
the man in front of the trolley because of a negative emotional response, then a
dose of positive emotion induced by watching a short comedic sketch from
Saturday Night Live might counteract that negative response and make their
judgments more consequentialist. As predicted, the researchers found that
people who watched the funny video were more willing to endorse pushing the
man in front of the trolley.
Yet another method for establishing a causal relationship between emotion and
moral judgment is to test individuals with selective deficits in emotional
processing. This approach was first taken by Mendez and colleagues (2005) in a
study of patients with frontotemporal dementia (FTD), a disease characterized
by deterioration of prefrontal and anterior temporal brain areas. FTD patients
exhibit blunted emotion and diminished regard for others early in the disease
course. Behavioral changes include moral transgressions such as stealing,
physical assault, and unsolicited or inappropriate sexual advances. Mendez and
colleagues presented FTD patients with versions of the switch and footbridge
dilemmas and found, as predicted, that most FTD patients endorsed not only
flipping the switch but also pushing the person in front of the trolley in order to
save the five others. Mendez and colleagues suggest that this result is driven by
the deterioration of emotional processing mediated by the ventromedial
prefrontal cortex (VMPC). Since neurodegeneration in FTD affects multiple
prefrontal and temporal areas, however, firm structure–function relationships
cannot be ascertained from this study.
To fill this gap, moral judgment has since been investigated in patients with
more focal VMPC lesions—that is, lesions involving less damage to other
structures. Like FTD patients, VMPC lesion patients exhibit reduced affect and
diminished empathy, but unlike FTD patients, VMPC lesion patients retain
broader intellectual function. Thus VMPC patients are especially well suited to
studying the role of emotion in moral judgment. Koenigs, Young, and colleagues
(2007) tested a group of six patients with focal, adult‐onset, bilateral lesions of
VMPC to determine whether emotional processing subserved by VMPC is in fact
necessary for deontological moral judgment. In this study patients evaluated a
series of impersonal and personal moral scenarios, used by Greene and
colleagues in the neuroimaging work discussed above. VMPC patients responded
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-224
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-208
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-202
Multi‐system Moral Psychology
Page 8 of 26
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
normally to the impersonal moral scenarios, but for the personal scenarios the
VMPC patients were significantly more likely to endorse committing an
emotionally aversive harm (e.g. smothering the baby) if a greater number of
people would benefit. That is, they were (p.54) more consequentialist. A second
lesion study conducted by Ciaramelli and colleagues (2007) produced consistent
results. Thus these lesion studies lend strong support to the theory that
characteristically deontological judgments are—in many people, at least—driven
by intuitive emotional responses that depend on the VMPC, while
characteristically consequentialist judgments are supported by controlled
cognitive processes based in the DLPFC.
It is worth noting that this dual‐process theory, and the body of evidence that
supports it, run counter to the traditional philosophical stereotypes concerning
which normative positions are most closely allied with emotion vis‐à‐vis
controlled cognition. Historically, consequentialism is more closely allied with
the “sentimentalist” tradition (Hume, 1739/1978), while deontology is more
closely associated with Kantian (1785/1959) rationalism. According to the dual‐
process theory, this historical stereotype gets things mostly backwards. The
evidence suggests that characteristically deontological judgments are driven
primarily by automatic emotional responses, while characteristically
consequentialist judgments depend more on controlled cognition (Greene, 2007).
Of course, this association between emotion and deontology, on the one hand,
and consequentialism and controlled cognition, on the other, is bound to be
overly simple. In what follows, we present a more nuanced picture of the
interactions between cognition and affect that support both characteristically
consequentialist and characteristically deontological moral judgments.
The division between affective and cognitive systems of moral judgment
proposed by Greene and colleagues is by now well supported. Several critics
have noted, however, that early formulations of the theory left many details of
the affective system unspecified (Cushman et al., 2006; Hauser, 2006: 8; Mikahil,
2007). Taking the specific example of the trolley problem, what is it about the
footbridge dilemma that makes it elicit a stronger emotional responses than the
switch dilemma? As Mikhail (2000) observed, this question can be answered in
at least two complementary ways. On a first pass, one can provide a descriptive
account of the features of a given moral dilemma that reliably produce
judgments of “right” or “wrong.” Then, at a deeper level, one can provide an
explanatory account of these judgments with reference to the specific cognitive
processes at work.
A natural starting point for the development of a descriptive account of our
moral judgments is the philosophical literature, which in the last half‐century
has burgeoned with principled accounts of moral intuitions arising from (p.55)
hypothetical scenarios (e.g. Fischer & Ravizza, 1992). The most prominent
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-173
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-196
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-199
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-183
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-175
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-193
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-210
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-209
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-179
Multi‐system Moral Psychology
Page 9 of 26
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
philosophical account of the trolley problem is a moral principle called the
“Doctrine of Double Effect,” or DDE. According to proponents of the DDE, the
critical difference between the switch and footbridge cases is that the large man
in the footbridge case is used to stop the train from hitting the five, whereas the
death of the victim in the switch dilemma is merely a side‐effect of diverting the
trolley away from the five. In its general form, the DDE states that it is
impermissible to use a harm as the means to achieving a greater good, but
permissible to cause a harm as a side‐effect of achieving a greater good. Setting
aside its validity as a moral principle, we may ask whether the DDE is an
accurate descriptive account of the moral judgments of ordinary people.
The characteristic pattern of responses elicited by the trolley problem provides
some evidence for this view, but the DDE is certainly not the only account for
subjects’ judgments of these cases. Another obvious distinction between the
footbridge and switch cases is that the former involves physical contact with the
victim, while the latter occurs through mechanical action at a distance
(Cushman et al., 2006). At a more abstract level, the footbridge case requires an
intervention on the victim (pushing the large man), while the bystander case
merely requires an intervention on the threat (turning the trolley) (Waldman &
Dieterich, 2007).
In order to isolate the DDE from these sorts of alternative accounts, Mikhail
(2000) tested subjects on a modified version of the switch dilemma. Both cases
involve a “looped” side track that splits away from, and then rejoins, the main
track (Figure 2.1). Five people are threatened on the main track, and they are
positioned beyond the point where the side track rejoins. Thus, if the train were
to proceed unimpeded down either the main track or the side track, the five
would be killed. In the analog to the footbridge case, there is a man standing on
the side track who, if hit, would be sufficiently large to stop the train before it
hit the five. In the analog to the switch case, there is a weighted object on the
side track that, if hit, would be sufficiently large to stop the train before it hit the
five. However, standing in front of the weighted object is a man who would first
be hit and killed. These cases preserve the distinction between harming as a
means to saving five (when the man stops the train) and harming as a side‐effect
of saving five (when an object stops the train and a man dies incidentally).
However, neither involves physical contact, and both require a direct
intervention on the trolley rather than the victim.
Mikhail found that subjects reliably judged the looped means case to be morally
worse than the looped side‐effect case, although the size of the effect was
markedly smaller than the original switch/footbridge contrast. These results
suggest that the DDE is an adequate descriptive account for at least (p.56)
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-175
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-225
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-209
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-figureGroup-4
Multi‐system Moral Psychology
Page 10 of 26
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
Figure 2.1.
some part of the moral distinction
between the fat man and
bystander cases. A follow‐up study
involving several thousand
subjects tested online replicated
this effect, and found it to be
remarkably consistent across
variations in age, gender,
educational level, exposure to
moral philosophy, and religion
(Hauser et al., 2007). While the
specific wording of Mikhail’s
“loop” cases has been criticized
(Greene et al., 2009; Waldman &
Dieterich, 2007), subsequent
research has demonstrated use of
the DDE across multiple
controlled pairs of moral dilemmas
(Cushman et al., 2006).
These initial studies establish
that DDE is at least part of an
accurate description of subjects’
moral judgments, but leave
open the explanatory question: how is the means/side‐effect distinction
manifested in the psychological systems that give rise to moral judgment?
Several related proposals have been offered (Cushman, Young, & Hauser, in
preparation; Greene et al., 2009; Mikahil, 2007), but we will not discuss these in
detail. In short, all three share the claim that when harm is used as the means to
an end it is represented as being intentionally inflicted to a greater degree than
when harm is produced as a side‐effect. Cushman and colleagues (in
preparation) provide evidence that that this relationship holds even in non‐moral
cases. That is, when somebody brings about a consequence as a means to
accomplishing a goal, participants are more likely to judge that the effect was
brought about intentionally than (p.57) when somebody brings about a
consequence as a foreseen side‐effect of their goal. Many details remain to be
worked out in providing a mechanistic account of the DDE in producing moral
judgments, but it is likely to involve representations of others’ mental states.
Several studies have explored another dimension of the cognitive processes
underlying the DDE: specifically, whether they operate at the level of conscious
awareness. Hauser et al. (2007) selected a subset of participants who judged
killing the one to be impermissible in the loop means case but permissible in the
loop side‐effect case and asked them to provide a justification for their pattern of
judgments. Of twenty‐three justifications analyzed, only three contained a
principled distinction between the two cases (e.g. “the man’s death is a by‐
product rather than direct goal of trying to save the five men”). Subsequent
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-194
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-185
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-225
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-175
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-185
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-210
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-194
Multi‐system Moral Psychology
Page 11 of 26
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
research has replicated this result with a much larger group of participants,
demonstrating the general inability of a majority of individuals to provide a
sufficient justification for a variety of double effect cases (Cushman et al., 2006).
These results illustrate a phenomenon that Haidt has termed “moral
dumbfounding” (Bjorklund et al., 2000). Moral dumbfounding occurs when
individuals make moral judgments that they confidently regard as correct, but
then cannot provide a general moral principle that accounts for their specific
judgment. One example is the moral prohibition against incest among
consenting adults. Haidt and Hersh (2001) told subjects about a brother and
sister who make love, but who use highly reliable contraception to avoid
pregnancy, are rightly confident that they will suffer no negative emotional
consequences, are able to keep their activities private, and so on. Presented with
such cases, subjects often insist that the siblings’ actions are morally wrong,
despite the fact that they cannot provide more general justification for this
judgment.
Haidt and colleagues argue that these difficult‐to‐justify moral judgments are
generated by rapid, automatic, unconscious mental processes—in short,
intuitions. Moreover, research indicates that these intuitions are supported by
affective processing. For instance, subjects’ moral judgments are harsher when
made in a physically dirty space (Schnall, Haidt, Clore, & Jordan, 2008). It
appears that subjects’ feelings of disgust brought about by the dirty space were
misattributed to the moral vignettes they judged, implicating disgust as an
affective contributor to intuitive moral judgments. In another study, Wheatley
and Haidt (2005) hypnotized subjects to experience disgust when they heard a
key target word, such as “often.” When subjects were told apparently innocent
stories that contained this word—for example, a story about cousins who often
visit the zoo, or a student council officer who often picks discussion topics—
some made moralistic allegations against the protagonists, saying (p.58) for
instance, “It just seems like he is up to something,” or that the he is a
“popularity‐seeking snob.”
In summary, there is research suggesting that the DDE characterizes patterns of
moral intuition, and research suggesting that affective responses underlie
certain moral intuitions. This raises a clear question: does contemplating harm
used as the means to an end trigger an affective response that in turn generates
moral judgments consistent with the DDE? We suggest that the answer is yes:
although the DDE alone cannot be fashioned into a complete descriptive account
of the conditions under which affective processes are engaged in moral
judgments of harmful acts, there are good reasons to suppose that it captures
part of the story. The DDE is among the features that distinguish “personal”
from “impersonal” dilemmas used to validate the dual‐process model (Greene et
al., 2001; Koenigs et al., 2007; Mendez et al., 2005; Valdesolo & DeSteno, 2006).
A subsequent study by Shaich Borg and colleagues (2006) also revealed
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-175
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-170
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-191
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-216
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-226
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-188
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-202
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-208
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-224
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-215
Multi‐system Moral Psychology
Page 12 of 26
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
increased neural activity in brain regions associated with emotion, such as the
VMPC, during the judgment of cases involving the use of harm as a means to a
greater good. These data suggest that the use of harm as a means may play a
specific role in engaging affective processes of moral judgment. However, the
stimuli used have often not been sufficiently well controlled, leaving room for
multiple interpretations.
We draw several conclusions from this collective body of data. First, the
processes that shape moral judgment often operate below the level of conscious
awareness, in the sense that individuals are often unable to articulate the basis
for moral judgments derived from them. Second, these unconscious mechanisms
operate over representations of agents and actions, causes and intentions, and
so on, in ways that can be captured by fairly general principles such as the DDE.
Third, although aspects of these unconscious mechanisms may involve the
processing of information without emotional valence, evidence suggests that
emotion plays some causal role in generating the ultimate moral judgment.
Finally, the very sorts of features that seem to guide intuitive moral judgments
also seem to guide affective moral judgments. A parsimonious account of these
data clearly stands out: the research programs into the affective basis of moral
judgment and research programs into the intuitive basis of moral judgment have
been investigating the same kind of psychological process.
While the research described above associates conscious, principled reasoning
with characteristically consequentialist moral judgment and emotional intuition
(p.59) with characteristically deontological moral judgment, other evidence
suggests that this pattern need not hold in all cases. Consider, first,
deontological philosophy. It seems that philosophical reasoning can lead to
judgments that are not consequentialist and that are even strikingly
counterintuitive. (See, e.g., Kant’s [1785/1983] infamous claim that it would be
wrong to lie to a would‐be murderer in order to save someone’s life.) Even when
intuitions inspire putative deontological distinctions, reasoning can then
determine their actual role in a normative theory, determining whether and the
extent to which we should take them seriously. To adapt a dictum from
evolutionary biology: intuition proposes, reason disposes.
But philosophers aren’t the only ones providing evidence of non‐consequentialist
reasoning that appears to be conscious, principled, and deliberate. Let’s return
briefly to the patients with emotional deficits due to VMPC damage (Koenigs et
al., 2007). The sketch we provided of their performance on personal scenarios
was just that—a sketch. The full picture is both richer and messier. The personal
moral scenarios on which VMPC patients produced abnormally consequentialist
judgments could in fact be subdivided into two categories: “low‐conflict” or easy
scenarios and “high‐conflict” or difficult scenarios. Low‐conflict scenarios
elicited 100% agreement and fast reaction times from healthy control subjects;
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-200
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-202
Multi‐system Moral Psychology
Page 13 of 26
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
high‐conflict scenarios elicited disagreement and slow responses. All high‐
conflict scenarios featured a “consequentialist” option in which harm to one
person could serve to promote the welfare of a greater number of people. Low‐
conflict scenarios, by contrast, typically described situations in which harm to
one person served another person’s purely selfish ends, for example, throwing
one’s baby in a dumpster to avoid the financial burden of caring for it. In these
cases, the VMPC patients judged the actions to be wrong, just as normal
individuals do.
How do VMPC patients arrive at the “appropriate” answer on low‐conflict or
easy personal moral scenarios? One proposal is that low‐conflict scenarios pit a
strong emotional response to the harmful action against a weak case for the
alternative. According to this proposal, VMPC subjects could have generated the
normal pattern of judgments on low‐conflict scenarios because they retained
sufficiently intact emotional processing to experience an aversion to the harm.
This proposal isn’t entirely plausible, however, in light of the fact that the VMPC
subjects tested show abnormal processing of even highly charged emotional
stimuli.
According to an alternative proposal, VMPC patients reasoned their way to
conclusions against causing harm. The difference between low‐conflict and high‐
conflict scenarios is driven by the absence of conflicting moral norms in the low‐
conflict scenarios and the presence of conflicting moral norms in the (p.60)
high‐conflict scenarios. As described above, high‐conflict scenarios described
situations that required harm to one person to help other people. The decision of
whether to endorse such harm presents participants with a moral dilemma, in
the sense that they have distinct moral commitments demanding opposite
behaviors. Regardless of the decision, either a norm against harming or a norm
against not helping is violated. Low‐conflict scenarios, on the other hand,
typically described situations that required harm to one person in order to help
only oneself. Thus it is possible that an uncontested moral norm against harming
someone purely for self‐benefit guides judgment. Alternatively, the VMPC
patients could be applying the same utilitarian principles that they applied to the
high‐conflict cases, judging, for example, that the financial benefits to the young
mother are outweighed by the harm done to her infant.
There were some low‐conflict dilemmas featuring situations in which harm to
one person was required to save other people. These scenarios, however, feature
actions that violate widely understood moral norms, for instance norms against
child prostitution, cannibalism, and the violation of a patient’s rights by his
doctor. The pattern of data thus suggests that patients with compromised
emotional processing are able to use their intact capacity for reasoning to apply
clear moral norms to specific situations.
Multi‐system Moral Psychology
Page 14 of 26
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
Finally, ordinary people appear capable of bringing explicit moral principles to
bear on particular cases in service of moral judgment. In contrast to the
distinction between intended and foreseen harm, which appears to be intuitively
generated and then justified post hoc, the distinction between killing and letting
die (or, more generally, between action and omission) may commonly be
consciously deployed in the first place. Cushman and colleagues (2006) asked
people to justify patterns of their own judgments that were consistent with these
two distinctions. They found that many people were able to explicitly produce a
version of the action/omission distinction, while very few were able to explicitly
produce a version of the distinction between intended and foreseen harm. Thus
it is at least possible that ordinary people engage in moral reasoning from some
version of the action/omission distinction, just as some philosophers do. Even
more revealing is a reanalysis of that data showing that participants who were
able to articulate the action/omission distinction showed significantly greater use
of the distinction in their prior judgments. This link between having a principle
and using that principle provides stronger evidence that explicit moral reasoning
can play a role in the process of moral judgment.
Further evidence for the role of moral reasoning—albeit a limited role—comes
from a study by Lombrozo (2008). Participants were asked (p.61) several
questions designed to reveal whether they adhered to a more consequentialist or
more deontological moral theory. For instance, they were asked whether it is
“never permissible” to murder (Lombrozo designated this a deontological
response), “permissible” if it would bring about a greater good
(consequentialist), or “obligatory” if it would bring about a greater good (strong
consequentialist). They were also asked to make a moral judgment of two
specific cases: the “switch” and “push” variants of the trolley problem. Subjects’
general commitments with their judgment of particular cases, but this effect was
limited in a revealing way. As predicted, subjects whose general commitments
were consequentialist exhibited greater consistency in their moral judgments of
the “push” and “switch” cases. This effect only emerged, however, when the
“push” and “switch” cases were judged side by side. When specific judgments
were made sequentially, there was no effect of subjects’ general moral
commitments. This suggests that moral reasoning can be triggered in some
specific contexts, but not in others.
We expect that further research will uncover further evidence for conscious,
principled reasoning in moral judgments. The examples offered here serve to
support its potential role in non consequentialist moral judgments—adding
important detail to the dual‐process model.
So far we have considered the general properties of two different processes that
shape moral judgment: a deliberate, effortful process that reasons about specific
cases from explicit abstract principles, and a rapid, automatic process of moral
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-175
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-205
Multi‐system Moral Psychology
Page 15 of 26
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
judgment that generates affective responses to specific cases on the basis of
mental processes inaccessible to conscious reflection. We have also begun to
characterize these systems at a more computational level, specifying the content
of their moral rules: a general principle favoring welfare‐maximizing behaviors
appears to be supported by controlled cognitive processes, while a principle
prohibiting the use of harm as a means to a greater good appears to be part of
the process that generates intuitive emotional responses. We conclude by
turning to a crucial, unanswered question: how do these principles get into our
heads?
Elsewhere Greene has suggested that some of our emotionally driven moral
judgments have an innate and evolutionarily adaptive basis (Greene & Haidt,
2002; Greene, 2003, 2007). Recent research demonstrating sophisticated social
(p.62) evaluation in preverbal infants has begun to lend credence to this view
(Hamlin, Wynn, & Bloom, 2007). At the very least, it points the way towards
research paradigms that could support a nativist hypothesis. While we look
forward eagerly to progress on this front, here we will not pursue the question
of the developmental or adaptive origins of the principles at work on the
intuitive emotional side of the dual‐process divide.
Our aim in this section is instead to consider the origins of the cognitive
commitment to a utilitarian principle favoring welfare‐maximizing choices
(hereafter, the “welfare principle”). We explore the Humean (1739/1978)
hypothesis that utilitarian judgment, despite its association with controlled
cognition (Greene et al., 2004) and its prominence in the presence of emotional
deficits (Mendez et al., 2005; Koenigs et al., 2007; Ciaramelli et al., 2007), itself
has an affective basis. Put simply, we suggest that affect supplies the primary
motivation to regard harm as a bad thing, while “cognition” uses this core
statement of value to construct the utilitarian maximum that we ought to act so
as to minimize harm. The result is the welfare principle. Below, we sketch out
two versions of this hypothesis in greater detail.
They begin with a proposal by Greene (2007) distinguishing two kinds of
emotional responses: those that that function like alarm bells and those that
function like currencies. This distinction is meant to capture the difference
between deontological intuitions and utilitarian reasoning, but it is also offered
as a broader distinction that operates beyond the moral domain. The core idea is
that alarm‐bell emotions are designed to circumvent reasoning, providing
absolute demands and constraints on behavior, while currency emotions are
designed to participate in the process of practical reasoning, providing
negotiable motivations for and against different behaviors. For example, the
amygdala, which has been implicated in responses to personal moral dilemmas,
reliably responds to threatening visual stimuli such as snakes and faces of out‐
group members (LeDoux, 1996: 166; Phelps et al., 2000). Thus the amygdala is a
good candidate for a region that is critical for supporting at least some alarm‐
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-184
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-182
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-183
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-192
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-196
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-187
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-208
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-202
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-173
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-183
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-204
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-212
Multi‐system Moral Psychology
Page 16 of 26
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
bell emotions. In contrast, Knutson and colleagues (2005) have identified a set of
meso‐limbic brain regions that appear to represent expected monetary value in a
more graded fashion, with distinct regions tracking a stimulus’s reward
magnitude, reward probability, and expected value. These regions, in a rather
transparent way, support currency‐like representations.
Above, we argued that strongly affective deontological intuitions are triggered
when types of harms occur, perhaps those that involve physical contact and use
harm as a means to accomplishing a goal. The emotional response is like an
alarm bell because it is designed to make a clear demand that is extremely
difficult to ignore. This response can be overridden with substantial effort (p.63)
(cognitive control), but it is not designed to be negotiable. The prohibition of
certain physical harms seems to be just an example of a broader class of alarm‐
bell emotions. For instance, research on “protected values” (Baron & Spranca,
1997) and “sacred values” (Tetlock, Kristel, Elson, Green, & Lerner 2000) in
moral judgment suggests that people sometimes treat harm to the environment
as non‐negotiable as well. Outside the moral domain, the strong disgust
response to eating contaminated food also seems to have an alarm‐like
character: most people are not willing to negotiate eating feces.
Whereas alarm‐bell emotions treat certain types of physical harms as non‐
negotiable, it is precisely the defining feature of the “welfare principle” that
harms are negotiable. One person’s life counts for one‐fifth as much as five
people’s lives, and the relative costs and benefits of lives, among other
outcomes, are weighed against each other to determine the best course of
action. Here, again, the case of moral decision‐making is presumed to be one
case of a much broader phenomenon. The desire for ice cream on a hot summer
day is an example of a currency emotion: it supplies a reason to pursue the Good
Humor truck, but this reason can be traded off against others, such as
maintaining a slim poolside profile. Currency‐like emotions function by adding a
limited measure of motivational weight to a behavioral alternative, where this
weighting is designed to be integrated with other weightings in order to produce
a response. Such emotional weightings say, “Add a few points to option A” or
“Subtract a few points from Option B,”, rather than issuing resolute commands.
Above, we suggested that affect supplies the primary motivation to regard harm
as bad. Once this primary motivation is supplied, reasoning proceeds in a
currency‐like manner. But we still must explain the origin of the primary
motivation to regard harm as bad. Is it located in the alarm‐bell response to
personal harms, or in an independent currency response? Each hypothesis has
pros and cons.
According to the “alarm‐bell” hypothesis, the primary motivation not to harm is
ultimately derived from the alarm‐bell emotional system that objects to things
like pushing a man in front of a train. When people attempt to construct general
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-201
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-169
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-220
Multi‐system Moral Psychology
Page 17 of 26
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
principles that account for their particular “alarm‐bell” moral intuitions, one of
the first things they notice is that their intuitions respond negatively to harm.
This gives rise to a simple moral principle: “harm is bad.” Combined with a
general cognitive strategy of minimizing bad things (i.e. practical reasoning), the
result is the welfare principle. Represented as such, the welfare principle can
become a basis for controlled, conscious processes of reasoning. Critically,
reasoning on the basis of an explicit welfare principle can occur in a currency‐
like manner. The alarm‐bell hypothesis locates the origins (p.64) of the welfare
principle in a process of theory‐construction over alarm‐bell responses, but
maintains the view that the operation of the welfare principle occurs in a
currency‐like manner that engages controlled reasoning processes.
According to this hypothesis, the welfare principle takes hold not because it
offers a fully adequate descriptive account of our intuitive moral judgments
(which it does not), but because it is simple, salient, and accounts for a large
proportion of our intuitive judgments (which it does). Ultimately, the same
mechanism can also give rise to more complex moral principles. For instance,
Cushman and colleagues (in preparation) have explored this hypothesis in the
particular cases of the doctrine of double effect and doctrine of doing and
allowing.
The principle virtue of the alarm‐bell hypothesis is its parsimony: by explaining
the origins of the welfare principle in terms of an independently hypothesized
alarm‐bell aversion to harm, it can provide a motivational basis for controlled
cognitive moral reasoning without invoking any additional primary affective
valuation. It can also explain why utilitarian moral judgment is preserved in
individuals who experience damage to frontal affective mechanisms: the welfare
principle has already been constructed on the basis of past affective responses.
But one shortcoming of the alarm‐bell hypothesis is that it leaves unexplained
how a theory of one’s own moral intuitions gives rise to practical reasoning.
When an individual regards her own pattern of moral intuitions and notes, “I
seem to think harm is bad,” will this lead automatically to the conclusion “Harm
is a reason not to perform a behavior”? At present, we lack a sufficient
understanding of the interface between cognition and affect to answer this
difficult question. It seems plausible that a descriptive theory of one’s
motivations could become a motivational basis for practical reasoning, but it also
seems plausible that it might not.
Philosophers will note that the alarm‐bell hypothesis paints an unflattering
portrait of philosophical utilitarianism because it characterizes the welfare
principles as a sort of crude first pass, while characterizing deontological
principles as more subtle and sophisticated. People’s intuitive judgments are
often consistent with the welfare principle, but it is clear that in many cases they
are not—for instance, in the footbridge version of the trolley problem. If the
welfare principle is designed to capture our intuitive moral judgments as a
Multi‐system Moral Psychology
Page 18 of 26
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
whole, then it would receive about a B+, getting things descriptively right much
of the time, but getting things descriptively wrong much of the time, too.
The currency hypothesis is more friendly toward utilitarianism/consequentialism
as a normative approach. According to the currency hypothesis, we are
furnished with certain currency‐like affective responses independently (p.65) of
our alarm‐bell responses. For instance: harm is bad, regardless of who
experiences it. Benefits are good, regardless of who experiences them. More
harm is worse than less harm. More benefits are better than fewer benefits.
Small harms can be outweighed by large benefits. Small benefits can be
outweighed by large harms. And so on.
Notice that these premises assign a valence to outcomes (harm is bad), and even
ordinal relationships among outcomes (more harm is worse), but they do not
assign cardinal values to particular outcomes. Thus they specify the general
structure of utilitarian thinking, but do not specify how, exactly, various harms
and benefits trade off against one another. According to the currency process,
utilitarian thinking is the product of a rational attempt to construct a set of
practical principles that is consistent with the constraints imposed by the
aforementioned premises. It is an idealization based on the principles that
govern the flow of emotional currency. One might say that it is a union between
basic sympathy and basic math.
Note that the operational mechanisms upon which the currency hypothesis
depends—the math, so to speak—overlap substantially with the mechanisms
upon which the alarm‐bell hypothesis depends. Both accounts posit that explicit
utilitarian principles arise from a combination of abstract reasoning and affect.
The key difference is the source of the affect: generalization from alarm‐bell
responses prohibiting harm versus an independent set of currency‐like
valuations of harms and benefits.
The principal virtue of the currency hypothesis is that utilitarian cost–benefit
reasoning looks very much like cost–benefit reasoning over other currency‐like
responses. The welfare principle functions very much as if there were a
negotiable negative value placed on harm—and, for that matter, a negotiable
positive value placed on benefits. Also, in contrast to the alarm‐bell hypothesis,
it is apparent how the currency‐like weighting of costs and benefits directly and
necessarily enters into practical reasoning. This tight fit comes with a slight
expense in parsimony, however. The currency hypothesis demands two separate
types of affective response to harm of independent origin: an alarm‐bell
response to a select set of harms, and a currency‐like response to a larger set of
harms. It also implies that currency‐like responses are preserved in individuals
with frontal‐lobe damage, since they continue to reason from the welfare
principle.
Multi‐system Moral Psychology
Page 19 of 26
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
As noted earlier, the currency hypothesis is more friendly toward philosophical
utilitarianism. According to this view, utilitarians are not simply doing a poor job
of generalizing over the body of their alarm‐bell moral intuitions. Instead, their
judgments are based indirectly on a distinct set of currency‐like emotional
responses. Is it better to rely on currency‐like emotions to the (p.66) exclusion
of alarm‐like emotions? Perhaps. The alarm‐like emotions that drive people’s
non‐utilitarian judgments in response to trolley dilemmas appear to be sensitive
to factors that are hard to regard as morally relevant, such as whether the
action in question involves body‐contact between agent and victim (Cushman et
al., 2006). Taking the charge of bias one step further, Greene et al. (2009)
hypothesizes that the DDE is a by‐product of the computational limitations of the
processes that govern our intuitive emotional responses. Borrowing some
computational machinery from Mikhail (2000), he argues that the controlled
cognitive system based in the DLPFC has the computational resources necessary
to represent the side‐effects of actions, while the appraisal system that governs
our emotional responses to actions like the one in the footbridge dilemma lacks
such resources. As a result, our emotional responses have a blind spot for
harmful side‐effects, leading us to draw a moral distinction between intended
and foreseen harms, i.e. the DDE.
Although the alarm‐bell and currency hypotheses vary in detail, they both reject
philosophical moral rationalism in that they (a) require a specification of
primitive goods before practical reasoning (including utilitarian reasoning) can
proceed and (b) locate these primitives in our affective responses. Moreover,
these hypotheses are not mutually exclusive: the welfare principle may be
supported both by theory‐building based on alarm‐bell responses as well as a
distinct set of currency responses. As we argued above, there is ample evidence
that a distinction between cognitive and affective processes of moral judgment is
warranted. We strongly suspect, however, that when the origins of the cognitive
process are understood, we shall find a pervasive influence of affect.
As we hope this chapter attests, it no longer makes sense to engage in debate
over whether moral judgment is accomplished exclusively by “cognition” as
opposed to “affect,” or exclusively by conscious reasoning as opposed to
intuition. Rather, moral judgment is the product of complex interactions between
multiple psychological systems. We have focused on one class of moral
judgments: those involving tradeoffs between avoiding larger harms and causing
smaller ones. Cases such as these engage at least two distinct systems: an
intuitive/affective response prohibiting certain kinds of basic harm, and a
conscious/cognitive response favoring the welfare‐maximizing response. The
underlying psychology of moral judgment in these cases helps to explain why
they strike us as particularly difficult moral dilemmas: we are forced to reconcile
the conflicting output of competing brain systems. (p.67)
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-175
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-185
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-209
Multi‐system Moral Psychology
Page 20 of 26
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
We have also identified several aspects of this account that are in need of further
investigation. First, there is much to be learned about the evaluative processes
that operate within each of the systems we have identified. In the case of the
intuitive/affective system, we have suggested that one component of the
evaluative process mirrors the doctrine of double effect. But the evidence is not
conclusive on this point. Moreover, this single principle alone cannot account for
the full pattern of data associated with intuitive/affective moral judgments. In
the case of the conscious/cognitive system, the data strongly suggest that
ordinary people typically reason from a principle favoring welfare‐maximizing
choices. But there is also good reason to believe that people, at least in some
circumstances, explicitly reason from deontological moral principles. Further
research in this area will be critical.
Second, the origins of each system of moral judgment remain unknown. In this
chapter we have explored two hypotheses concerning the origins of explicit
moral principles within individuals, both of which ultimately implicate affective
systems. Others have proposed diverse origins for moral principles such as
metaphor (Lakoff & Johnson, 1999: 290–334) and construction from social
experience (Kohlberg, 1969). A further question is the origin of moral intuitions,
which several authors have suggested may have an innate basis (Hauser, 2006:
1; Mikhail, 2007; Greene & Haidt, 2002). Several of these proposals operate at
distinct levels of analysis (ontogenetic versus evolutionary) and are specific to
one of the two systems we have discussed (cognitive versus affective), and thus
may be broadly compatible.
Third, at present we know little about how the intuitive/affective and conscious/
cognitive systems interact on‐line in the production of moral judgments. This is a
topic we have left largely untouched in the present chapter, and somewhat out of
necessity. Until the contours of individual systems of moral judgment are better
understood, it will be difficult to make much progress towards understanding
the interactions between systems. One aspect of this problem that we suspect
will be of interest are the standards for the normative plausibility of putative
moral principles. Certain factors that reliably shape moral judgments, such as
the physical proximity of an agent to a victim, are sometimes rejected as prima
facie invalid criteria for moral judgment. Others, such as the doctrine of double
effect, are more commonly accepted (Cushman et al., 2006). The bases on which
particular explicit moral rules are accepted or rejected are poorly understood,
and such meta‐ethical intuitions are a key area for investigation at the interface
between the intuitive/affective and conscious/cognitive systems.
Finally, it remains to be seen how generally the dual‐process model developed
for harm‐based moral dilemmas can be extended to other domains of (p.68)
morality. There are at least two ways that this issue can be framed. On the one
hand, we have argued that in the specific case of tradeoffs in harms, conflict
between distinct psychological systems gives rise to the phenomenon of a
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-206
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-203
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-193
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-210
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-184
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-175
Multi‐system Moral Psychology
Page 21 of 26
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
dilemma (Cushman & Young, 2009; Greene, 2007; Sinnott‐Armstrong, in
preparation). One important question is whether this multi‐system account can
be employed to understand the phenomenon of a moral dilemma beyond the
domain of physically harmful actions. That is, are there distinct systems that
give rise to potentially conflicting moral judgments in domains such as the
division of economic resources, the establishment of conventional standards of
conduct, sexual taboo, and so forth?
On the other hand, we have argued for the operation of two psychological
processes: an intuitive/affective response to intentional harms and a conscious/
cognitive response favoring welfare maximization. This raises a second question:
to what extent do other types of moral judgment depend on the operation of
these specific processes? For instance, when evaluating allocations of resources,
financial losses could be coded as “harms” and then processed via the operation
of one of the systems we have explored in this chapter. Whether these particular
systems have such broad applicability is presently unknown.
It is, of course, our hope that the small corner of moral judgment we have
explored in this chapter—a corner strictly limited to the occurrence of physical
harms and preternaturally concerned with railway operations—will teach lessons
with broad applicability. The extent of this applicability remains to be
determined, but we feel confident in asserting at least this much: there is no
single psychological process of moral judgment. Rather, moral judgment
depends on the interaction between distinct psychological systems. Notably,
these systems tend to be reflected in the principles characteristic of competing
moral philosophies (Greene, 2007).
We thank Tim Schroeder for his comments on an earlier version of this chapter.
We also thank the members of the Moral Psychology Research Group, and
especially John Doris, for their valuable input.
References
Bibliography references:
Baron, J., & Spranca, M. (1997). Protected values. Behavior and Human Decision
Processes, 70: 1–16. (p.69)
Bjorklund, F., Haidt, J., & Murphy, S. (2000). Moral dumbfounding: When
intuition finds no reason. Lund Psychological Reports, 2: 1–23.
Blair, R. J. R. (1995). A cognitive developmental approach to morality:
Investigating the psychopath. Cognition, 57: 1–29.
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-174
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-183
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-183
Multi‐system Moral Psychology
Page 22 of 26
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
Botvinick, M. M., Braver, T. S., Barch, D. M., Carter, C. S., & Cohen, J. (2001).
Conflict monitoring and cognitive control. Psychological Review, 108 (3): 624–
652.
Ciaramelli, E., Muccioli, M., Ladavas, E., & di Pellegrino, G. (2007). Selective
deficit in personal moral judgment following damage to ventromedial prefrontal
cortex. Social Cognitive Affective Neuroscience, 2: 84–92.
Cushman, F. A., & Young, A. W. (2009). The psychology of dilemmas and the
philosophy of morality. Ethical Theory and Moral Practice, 12 (1): 9.
Cushman, F. A., Young, L., & Hauser, M. D. (2006). The role of conscious
reasoning and intuitions in moral judgment: testing three principles of harm.
Psychological Science, 17 (12): 1082–1089.
—— (in preparation). Patterns of moral judgment derive from non‐moral
psychological representations.
Damasio, A. (1994). Descartes’ Error. Boston, MA: Norton.
Darley, J. M., & Shultz, T. R. (1990). Moral rules—their content and acquisition.
Annual Review of Psychology, 41: 525–556.
Fischer, J. M., & Ravizza, M. (1992). Ethics: Problems and Principles. New York:
Holt, Rinehart & Winston.
Foot, P. (1967). The problem of abortion and the doctrine of double effect. Oxford
Review, 5: 5–15.
Gilligan, C. (1982/1993). In a Different Voice: Psychological Theory and Women’s
Development. Cambridge, MA: Harvard University Press.
Greene, J. D. (2003). From neural ‘is’ to moral ‘ought’: what are the moral
implications of neuroscientific moral psychology? Nature Reviews Neuroscience,
4: 847–850.
—— (2007). The secret joke of Kant’s soul. In W. Sinnott‐Armstrong (ed.), Moral
Psychology (Vol. 3). Cambridge, MA: MIT Press, 35–117.
Greene, J. D., & Haidt, J. (2002). How (and where) does moral judgment work?
Trends in Cognitive Science, 6: 517–523.
Greene, J. D., Cushman, F. A., Stewart, L. E., Lowenberg, K., Nystrom, L. E., &
Cohen, J. D. (2009). Pushing moral buttons: The interaction between personal
force and intention in moral judgment. Cognition, 111 (3): 364–371.
Multi‐system Moral Psychology
Page 23 of 26
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
Greene, J. D., Morelli, S. A., Lowenberg, K., Nystrom, L. E., & Cohen, J. D.
(2008). Cognitive load selectively interferes with utilitarian moral judgment.
Cognition, 107: 1144–1154.
Greene, J. D., Nystrom, L. E., Engell, A. D., Darley, J. M., & Cohen, J. D. (2004).
The neural bases of cognitive conflict and control in moral judgment. Neuron,
44: 389–400.
Greene, J. D., Sommerville, R. B., Nystrom, L. E., Darley, J. M., & Cohen, J. D.
(2001). An fMRI investigation of emotional engagement in moral judgment.
Science, 293: 2105–2108. (p.70)
Haidt, J. (2001). The emotional dog and its rational tail: A social intuitionist
approach to moral judgment. Psychological Review, 108: 814–834.
—— (2007). The new synthesis in moral psychology. Science, 316 (5827): 998–
1002.
Haidt, J., & Hersh, M. (2001). Sexual morality. The cultures and emotions of
conservatives and liberals. Journal of Applied Social Psychology, 31: 191–221.
Hamlin, K., Wynn, K., & Bloom, P. (2007). Social evaluation by preverbal infants.
Nature, 450: 557–559.
Hauser, M. D. (2006). Moral Minds: How Nature Designed a Universal Sense of
Right and Wrong. New York: HarperCollins.
Hauser, M. D., Cushman, F. A., Young, L., Jin, R., & Mikhail, J. M. (2007). A
dissociation between moral judgment and justification. Mind & Language, 22 (1):
1–21.
Huebner, B., Dwyer, S., & Hauser, M. (2008). The role of emotion in moral
psychology. Trends in Cognitive Science, 13 (1): 1–6.
Hume, D. (1739/1978). A Treatise of Human Nature. In L. Selby‐Bigge and P. H.
Nidditch (eds.). New York: Oxford University Press.
Kamm, F. M. (1998). Morality, Mortality: Death and Whom to Save from it. New
York: Oxford University Press.
—— (2006). Intricate Ethics: Rights, Responsibilities, and Permissible Harm.
New York: Oxford University Press.
Kant, I. (1785/1959). Foundations of the Metaphysics of Morals. New York:
Macmillan.
—— (1785/1983). On a Supposed Right to Lie because of Philanthropic
Concerns. Indianapolis, IN: Hackett.
Multi‐system Moral Psychology
Page 24 of 26
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
Knutson, B., Taylor, J., Kaufman, M., Peterson, R., & Glover, G. (2005).
Distributed neural representation of expected value. Journal of Neuroscience, 25
(19): 4806–4812.
Koenigs, M., Young, L., Adolphs, R., Tranel, D., Cushman, F. A., Hauser, M. D., et
al. (2007). Damage to the prefrontal cortex increases utilitarian moral
judgments. Nature, 446: 908–911.
Kohlberg, L. (1969). Stage and sequence: The cognitive‐developmental approach
to socialization. In D. A. Goslin (ed.), Handbook of Socialization Theory and
Research. New York: Academic Press, 151–235.
LeDoux, J. (1996). The Emotional Brain. New York: Simon & Schuster.
Lombrozo, T. (2008). The role of moral commitments in moral judgment.
Cognitive Science, 33: 273–286
Lakoff, G., & Johnson, M. (1999). Philosophy in the Flesh. New York: Basic
Books.
Macmillan, M. (2000). An Odd Kind of Fame. Cambridge, MA: MIT Press.
Mendez, M. F., Anderson, E., & Shapria, J. S. (2005). An investigation of moral
judgment in frontotemporal dementia. Cognitive and Behavioral Neurology, 18
(4): 193–197.
Mikhail, J. M. (2000). Rawls’ linguistic analogy: A study of the “generative
grammar” model of moral theory described by John Rawls in A theory of justice.
Unpublished PhD, Cornell University, Ithaca, NY. (p.71)
—— (2007). Universal moral grammar: Theory, evidence, and the future. Trends
in Cognitive Science, 11 (4): 143–152.
Miller, E. K., & Cohen, J. D. (2001). An integrative theory of prefrontal cotex
function. Annual Review of Neuroscience, 24: 167–202.
Phelps, E. A., O’Connor, K. J., Cunningham, W. A., Funayama, E. S., Gatenby, J. C.,
Gore, J. C., et al. (2000). Performance on indirect measures of race evaluation
predicts amygdala activation. Journal of Cognitive Neuroscience, 12 (5): 729–
738.
Piaget, J. (1965/1932). The Moral Judgment of the Child. New York: Free Press.
Pizarro, D. A., Uhlmann, E., & Bloom, P. (2003). Causal deviance and the
attribution of moral responsibility. Journal of Experimental Social Psychology,
39: 653–660.
Multi‐system Moral Psychology
Page 25 of 26
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
Schaich Borg, J., Hynes, C., Van Horn, J., Grafton, S. T., & Sinnott‐Armstrong, W.
(2006). Consequences, action, and intention as factors in moral judgments: An
fMRI investigation. Journal of Cognitive Neuroscience, 18 (5): 803–837.
Schnall, S., Haidt, J., Clore, G. L., & Jordan, A. H. (2008). Disgust as embodied
moral judgment. Personality and Social Psychology Bulletin, 34: 1096–1109.
Schweder, D., & Haidt, J. (1993). The future of moral psychology: truth, intuition,
and the pluralist way. Psychological Science, 4: 360–365.
Sinnott‐Armstrong, W. (2008). Abstract + Concrete = Paradox. In S. Nichols & J.
Knobe (eds.), Experimental Philosophy. New York: Oxford University Press, 209–
230.
Stroop, J. R. (1935). Studies of interference in serial verbal reactions. Journal of
Experimental Psychology, 12: 643–662.
Tetlock, P., Kristel, O., Elson, S., Green, M., & Lerner, J. (2000). The psychology
of the unthinkable. Journal of Personality and Social Psychology, 78 (5): 853–870.
Thomson, J. J. (1985). The trolley problem. The Yale Law Journal, 279: 1395–
1415.
Turiel, E. (1983). The Development of Social Knowledge: Morality and
Convention. Cambridge: Cambridge University Press.
—— (2005). Thoughts, emotions, and social interactional processes in moral
development. In M. Killen and J. Smetana (eds.), Handbook of Moral
Development. Mahwah, NJ: Lawrence Erlbaum, 1–36.
Valdesolo, P., & DeSteno, D. (2006). Manipulations of Emotional Context Shape
Moral Judgment. Psychological Science, 17 (6): 476–477.
Waldman, M. R., & Dieterich, J. H. (2007). Throwing a bomb on a person versus
throwing a person on a bomb: intervention myopia in moral intuitions.
Psychological Science, 18 (3): 247–253.
Wheatley, T., & Haidt, J. (2005). Hypnotic disgust makes moral judgments more
severe. Psychological Science, 16 (10): 780–784.
Notes:
(1) By contrasting intuitive versus “rational” processes of moral judgment we
aim to capture the common social psychological distinction between automatic
(rapid, effortless, involuntary) and controlled (slow, effortful, voluntary)
processes. Our purpose in choosing the term “rational” is not to imply the
normative optimality of a particular decision, but rather to imply the use of
deliberative reasoning in reaching that decision.
Multi‐system Moral Psychology
Page 26 of 26
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
(2) For more on the somewhat artificial, but undeniably useful distinction
between affect and cognition, see Greene (2007).
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-3#acprof-9780199582143-bibItem-183
John M. Doris and The Moral Psychology Research Group
Multi‐system Moral Psychology
Fiery Cushman
Liane Young
Joshua D. Greene
Abstract and Keywords
Multi‐system Moral Psychology
Multi‐system Moral Psychology
Multi‐system Moral Psychology
1. A Dual‐Process Model of Moral Judgment
Multi‐system Moral Psychology
Multi‐system Moral Psychology
Multi‐system Moral Psychology
Multi‐system Moral Psychology
2. Intuition and Affect: A Common System?
Multi‐system Moral Psychology
Multi‐system Moral Psychology
Multi‐system Moral Psychology
Multi‐system Moral Psychology
3. Reasoning from Deontological Principles
Multi‐system Moral Psychology
Multi‐system Moral Psychology
4. The Pervasive Role of Affect
Multi‐system Moral Psychology
Multi‐system Moral Psychology
Multi‐system Moral Psychology
Multi‐system Moral Psychology
Multi‐system Moral Psychology
5. Conclusion
Multi‐system Moral Psychology
Multi‐system Moral Psychology
Acknowledgments
Multi‐system Moral Psychology
Multi‐system Moral Psychology
Multi‐system Moral Psychology
Multi‐system Moral Psychology
Notes:
Multi‐system Moral Psychology
Linguistics and Moral Theorys 1
Page 1 of 24
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
The Moral Psychology Handbook
John M. Doris and The Moral Psychology Research Group
Print publication date: 2010
Print ISBN-13: 9780199582143
Published to Oxford Scholarship Online: September 2010
DOI: 10.1093/acprof:oso/9780199582143.001.0001
Linguistics and Moral Theorys 1
Erica Roedder
Gilbert Harman (Contributor Webpage)
DOI:10.1093/acprof:oso/9780199582143.003.0009
This chapter describes various issues suggested by the analogy between
morality and linguistics. The first part discusses what might be involved in a
generative “moral grammar”. The second part discusses what might be involved
in a universal moral grammar that parallels universal linguistic grammar.
Keywords: acquisition, animal morality, categorization, constraints, core language, core morality,
generative grammar, moral grammar, recursive embedding, universal grammar
Analogies are often theoretically useful. Important principles of electricity are
suggested by an analogy between water current flowing through a pipe and
electrical current “flowing” through a wire. A basic theory of sound is suggested
by an analogy between waves caused by a stone being dropped into a still lake
and “sound waves” caused by a disturbance in air. At the very least, analogies
suggest questions and shape the direction of research; at most, analogies may
be crucial to explaining and understanding natural phenomena (Holyoak &
Thagard, 1995). Of course, principles suggested by such analogies need to be
confirmed by relevant evidence. Even where analogies are fruitful, they are only
partial. Sound waves and electricity are not wet; electricity does not spill out
when a wire is cut.
In this chapter we consider what is suggested by taking seriously an analogy
between language and morality. Recently there have been a number of striking
claims made about such a linguistic analogy—claims that, if true, have profound
implications for longstanding debates in moral philosophy about the innateness
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143
http://www.princeton.edu/~harman/
https://www.universitypressscholarship.com/search?f_0=keywords&q_0=acquisition
https://www.universitypressscholarship.com/search?f_0=keywords&q_0=animal morality
https://www.universitypressscholarship.com/search?f_0=keywords&q_0=categorization
https://www.universitypressscholarship.com/search?f_0=keywords&q_0=constraints
https://www.universitypressscholarship.com/search?f_0=keywords&q_0=core language
https://www.universitypressscholarship.com/search?f_0=keywords&q_0=core morality
https://www.universitypressscholarship.com/search?f_0=keywords&q_0=generative grammar
https://www.universitypressscholarship.com/search?f_0=keywords&q_0=moral grammar
https://www.universitypressscholarship.com/search?f_0=keywords&q_0=recursive embedding
https://www.universitypressscholarship.com/search?f_0=keywords&q_0=universal grammar
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-613
Linguistics and Moral Theorys 1
Page 2 of 24
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
of moral capacities, and the existence of moral universals. (For example, the title
of Hauser, 2006 is Moral Minds: How Nature Designed a Universal Sense of
Right and Wrong.) We shall indicate what support pursuit of the linguistic
analogy might provide for such claims. Perhaps more importantly, we shall
discuss implications the linguistic analogy might have for the methodology and
structure of moral theory.
Rules of morality have sometimes been described as analogous to rules of
grammar, with the occasional further suggestion that moral theory may be at
least in some respects analogous to linguistic theory. (p.274)
Adam Smith appeals to this analogy when he says, “The rules of justice may be
compared to the rules of grammar; the rules of the other virtues, to the rules
which critics lay down for the attainment of what is sublime and elegant in
composition” (Smith, 1817: 281).2
More recently, starting in the 1950s (Chomsky, 1957, 1965), a few moral
theorists began to see analogies between moral theory and generative linguistic
theory. According to John Rawls,
[O]ne may think of moral theory at first (and I stress the provisional nature
of this view) as the attempt to describe our moral capacity; or, in the
present case, one may regard a theory of justice as describing our sense of
justice. . . . A conception of justice characterizes our moral sensibility when
the everyday judgments we do make are in accord with its
principles. . . . We do not understand our sense of justice until we know in
some systematic way covering a wide range of cases what these principles
are.
A useful comparison here is with the problem of describing the sense of
grammaticalness that we have for the sentences of our native language
[here Rawls refers to Chomsky, 1965: 3–9]. In this case the aim is to
characterize the ability to recognize well‐formed sentences by formulating
clearly expressed principles which make the same discriminations as the
native speaker. This undertaking is known to require theoretical
constructions that far outrun the ad hoc precepts of our explicit
grammatical knowledge. A similar situation presumably holds in moral
theory. There is no reason to assume that our sense of justice can be
adequately characterized by familiar common sense precepts, or derived
from the more obvious learning principles. A correct account of moral
capacities will certainly involve principles and theoretical constructions
which go much beyond the norms and standards cited in everyday life . . .–
(Rawls, 1971: sec. 9)
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-612
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-646
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-590
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-592
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-592
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-641
Linguistics and Moral Theorys 1
Page 3 of 24
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
Moreover, the appeal of the linguistic analogy has not been limited to
philosophers in the Western “analytic tradition.” Pope Benedict XVI uses a
similar analogy in his message for the 2007 World Day of Peace:
The transcendent “grammar,” that is to say the body of rules for individual
action and the reciprocal relationships of persons in accordance with
justice and solidarity, is inscribed on human consciences, in which the wise
plan of God is reflected. As I recently had occasion to reaffirm: “we believe
that at the beginning of everything is the eternal word, reason and not
unreason.” Peace is thus also a task demanding of everyone a personal
response consistent with God’s plan. The criterion inspiring this response
can only be respect for the “grammar” written on human hearts by the
divine creator.–(Benedict, 2006)3
The authors above are explicitly concerned with analogies between principles of
grammar and principles of justice. In addition, other authors invoke an analogy
between rules of grammar and more general moral principles. (p.275)
For instance, three years before Rawls (1971), Nozick (1968) offered an account
of “moral complications and moral structures” in general, not limited to
principles of justice. Nozick (1997: 4–5) reports that one “model [for the 1968
paper] was Noam Chomsky’s Syntactic Structures, and I meant ‘Moral
Complications’ to begin to uncover the structure underlying our moral
judgments.” Nozick (1968) mentions that “one needs a distinction similar to that
which linguists make between linguistic competence and linguistic
performance,” referring to Chomsky (1965), a distinction we discuss below.
As another example, Kamm (2007) describes the methodology she uses in that
book and in Kamm (1992, 1993, 1996) as follows:
[P]eople who have responses to cases are a natural source of data from
which we can isolate the reasons and principles underlying their
responses. The idea [is] that the responses come from and reveal some
underlying psychologically real structure, a structure that was always
(unconsciously) part of the thought processes of some people. Such people
embody the reasoning and principles (which may be thought of as an
internal program) that generates these responses. . . . (Unlike the deep
structure of the grammar of a language, at least one level of the deep
structure of moral intuitions about cases seems to be accessible upon
reflection by those who have intuitive judgements . . . ) If the same “deep
structure” is present in all persons—and there is growing psychological
evidence that this is true (as in the work of Professor Marc Hauser)—this
would be another reason why considering the intuitive judgements of one
person would be sufficient, for each person would give the same
response.–(Kamm, 2007: 8)
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-583
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-641
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-636
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-637
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-636
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-592
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-619
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-616
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-617
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-618
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-619
Linguistics and Moral Theorys 1
Page 4 of 24
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
We return below to the issue whether the same moral deep structure is present in all
persons. At the moment we are content to point to Kamm’s acceptance of an analogy
between the grammar of a language and an individual’s moral principles.
Finally, Mikhail (2000, 2007) provides an account of “universal moral grammar,”
based on legal categories. We eventually return to the idea of a universal moral
grammar. But we want first to consider what proposals about “moral grammar”
are suggested by the analogy with the grammar of a language, appealing to
developments in contemporary linguistics starting with Chomsky (1957, 1965).
I‐grammar
The primary object of study of linguistics so conceived is not language in the
ordinary sense in which English, German, Mohawk, and Chinese are languages.
Any ordinary language in this sense has different dialects that blend into one
another in ways that do not correspond to national or geographical (p.276)
boundaries. There is a well‐known saying that “a language (in the ordinary
sense) is a dialect with an army and a navy.” What counts as a dialect of French
rather than Flemish is a social or political issue, not an issue in the science of
language. It may be that any two speakers have at least somewhat different
dialects, with at least somewhat different vocabularies. Chomsky (2000) and
other linguists (e.g. Freidin, 1992; Isac and Reiss, 2008) are concerned with
language as a property of a particular person, assumed to be abstractly
specifiable by an internal grammar, or I‐grammar.
This is one of a number of significant idealizations or abstractions linguistics
makes in the course of arriving at a theoretically useful account of an otherwise
unwieldy phenomenon. Others include the distinction, mentioned below,
between “competence” and “performance” and a more controversial distinction
some linguists make between “core” and other aspects of grammar.
In pursuing possible analogies between linguistics and moral theory, we might
consider whether similar idealizations or abstractions are theoretically useful to
moral theory. Should moral theory be in the first instance concerned with an
individual’s I‐morality—that particular person’s moral standards? This might be
captured by some sort of (possibly unconscious) abstract representation, an I‐
moral‐grammar. I‐morality in this sense would be distinguished from the
principles or conventions of a social group; it would also be distinguished from
the correct moral principles, if there are any.
“Generative” Grammar
A generative grammar of a language (a generative I‐grammar) would fully and
explicitly characterize the relevant linguistic properties of expressions of the
language (Chomsky, 1965; Freidin, 1992). It is very important to recognize that
the word “generative” is here used as a technical term. There is no implication
that the principles of a generative grammar are followed by speakers in
generating (in a more ordinary sense) what they say. To suppose that there is a
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-628
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-629
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-590
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-592
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-594
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-603
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-615
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-592
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-603
Linguistics and Moral Theorys 1
Page 5 of 24
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
passive “transformation”—a function that associates the structure underlying an
active sentence like Bill gave Mabel a book with the structure underlying a
passive sentence like Mabel was given a book by Bill—is not necessarily to
suppose that a speaker first forms the active structure and then, by some causal
process, converts it into the passive structure.
A generative grammar of a particular person’s language would specify the
linguistic structure of expressions of that language, indicating the nouns, verbs,
adjectives, prepositions, etc., as well as the phrases of which these are the
“heads,” as when a noun is the head of a noun phrase and a preposition the head
of a prepositional phrase. A generative grammar might specify important (p.
277) aspects of the meanings of expressions, depending on how the expressions
are structured and possibly indicating the scope of quantifiers and other such
operators. The grammar would also specify aspects of the pronunciation of
expressions (for spoken languages—something similar would be true of the
grammar of a sign language). So, a generative grammar would relate
pronunciations of expressions to possible interpretations, in this way indicating
certain sound–meaning (phonetic–semantic) relationships.
An analogous moral grammar would attempt explicitly to characterize an
individual’s moral standards. Just as a grammar specifies the structure of a well‐
formed linguistic sentence by using a specialized linguistic vocabulary, an I‐
moral grammar might specify the structure of impermissible actions, using a
specialized moral vocabulary.4 That is, just as English grammar might specify
how a noun phrase and verb phrase combine to form a grammatical sentence, an
I‐moral grammar might specify that that certain actions, in virtue of their
structure, are impermissible, e.g. intentional harm directed towards an innocent
person without reason. In this way, an I‐moral grammar might specify certain
action‐assessment relationships—or, indeed, character‐assessment or situation‐
assessment relationships. Thus the rules of an I‐moral‐ grammar would generate
a set of impermissible actions, just as a linguistic grammar generates a set of
grammatical sentences. An I‐moral grammar would have to specify the relevant
structures of the objects of moral assessment (e.g. acts, character traits,
situations, etc.) using a specialized moral vocabulary, perhaps including terms
like ought, obligation, duty, justice, fairness, right, wrong, responsibility, excuse,
justification, etc.
Relatedly, principles of generative grammar specify how the structure of a
sentence might be transformed, grammatically, into another. For instance, the
structure underlying a declarative sentence The book is on the chair is
transformable into the structure underlying a question, What is on the chair?
Characterized abstractly, such principles allow the theoretician to move from the
known grammaticality of one sentence structure to the grammaticality of a
transformed sentence structure. Similarly, there might be found certain
Linguistics and Moral Theorys 1
Page 6 of 24
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
transformative principles within I‐morality; we suggest an example of such a
principle below.
Given such a picture, it may be tempting to assume certain causal structures.
However, this would be a mistake. For instance, let us assume that the moral
grammar specifies action‐assessment relationships. If this is right, one might (p.
278) assume the following causal structure: first, some non‐moral brain system
assigns a structural description of the action. Second, the structural description
is provided as input to the moral system, and the moral system generates a
moral assessment. As in the linguistic case, however, there is no guarantee that
the brain’s causal pathway parallels that of the grammatical principles. While
the existence of such a moral grammar guarantees that there is, in a
mathematical sense, a mapping of action to assessments, the brain’s method of
processing information may be quite different. (Cf. Marr’s [1982] distinction
between computational and algorithmic descriptions of mental processing.)
This point about the non‐causal character of moral grammars is important for at
least two reasons, which we discuss immediately below. The first reason is a
theoretical point about recursive embedding; the second reason draws on some
recent empirical studies.
Recursive Embedding
The rules of a generative grammar specify recursive embedding of grammatical
structures. For example, a larger sentence can contain a smaller sentence within
it, which can contain another even smaller sentence, and so on without limit:
Jack told Betty that Sally believes that Albert no longer thinks Arnold
wants to leave Mabel.
Similarly, the rules of grammar imply that a noun phrase can contain a noun phrase
within it, which can contain another noun phrase, and so on:
the man next to the girl near the door with the broken hinge.
So, a particular generative grammar might imply rules that could be represented as
follows:
S => NP + (V + S)
NP => Art + N + (P + NP)
This point about linguistic recursive embedding is a point about grammatical
rules. It is not just that a sentence can occur within a larger sentence or a noun
phrase within a larger noun phrase. This sort of “recursion” is common enough:
many artifacts can contain smaller things of the same kind, e.g. a large raft can
be made of smaller rafts which have been tied together.5 Thus there is a sense in
which recursion is not particularly uncommon. However, there is a (p.279)
substantial difference between recursion found in linguistic grammars and this
sort of everyday “recursion.” In the case of linguistics, recursion is an explicit
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-627
Linguistics and Moral Theorys 1
Page 7 of 24
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
feature of rules that have proved theoretically useful. In contrast, recursion is
not explicitly implied by theoretically fruitful rules either for artifacts in general,
or for rafts in particular. Were we to construct rules to answer the question,
“What objects count as rafts?” the answer would presumably focus on how the
object functions (it bears weight as it floats on water), its similarity to certain
prototypes, etc. without making use of recursion.
The analogy between morality and language suggests that a “generative moral
grammar” might also imply recursive embedding. It might, for example, imply
that it is wrong to encourage someone to do something that is wrong, which is
recursive in the sense that it implies that it is wrong to encourage someone to
encourage someone to do something that is wrong, etc. Similarly, a generative
moral grammar might imply that it is wrong to promise to do something that it
would be wrong to do.6
There are two points to be made about recursive embedding. First, true
recursion—as opposed to the natural embedding that occurs with rafts—is an
important feature of grammatical rules. This recursive character allows a
relatively small number of rules to generate a potentially infinite set of
sentences. Recursion is an important, and perhaps distinctive, feature of
generative grammars; thus if moral judgments can only be modeled using a
recursive theory, that is a point in favor of the linguistic analogy.
Second, recursive embedding emphasizes that, in a generative I‐moral grammar,
the normative evaluation assigned to structural descriptions can depend on
normative evaluations embedded within those structural descriptions. The point
does not yet by itself rule out a processing model that first constructs
nonnormative action and situation descriptions in terms of causes and effects,
ends and means, distinguishing what is done intentionally or not, etc., and then,
when that is done, adds normative labels to various parts of the structure,
perhaps starting with smaller parts, and then labeling larger parts to handle the
recursive aspects of the I‐moral grammar. But such a processing model faces the
further worry that seemingly nonnormative aspects of relevant descriptive
structure themselves seem to depend on normative considerations.
Structural Dependence on Normative Assessments
There is considerable evidence (summarized in Knobe, 2008) that whether
people accept certain seemingly descriptive claims can depend on whether (p.
280) they accept certain normative claims about side effects of actions. For
example, suppose Tom increases his profits by doing something that, as a side
effect, either (a) harms or (b) helps the environment, where Tom does not care
either way about the effect on the environment. People who think that it is good
to help the environment and bad to harm it tend to say (a) that in so acting Tom
intentionally harmed the environment but will tend not to say (b) that in so
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-622
Linguistics and Moral Theorys 1
Page 8 of 24
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
acting Tom intentionally helped the environment (Knobe, 2003; for further
discussion, see Knobe & Doris, Chapter 10 in this volume).
They also tend to say (a) that Tom increased profits by harming the environment
but tend not to say (b) that he increased profits by helping the environment.
They tend to say (a) that Tom harmed the environment in order to increase
profits but tend not to say (b) that he helped the environment in order to
increase profits (Knobe, 2007). And they are more inclined to say (a) that the
harming of the environment was the same thing as the implementing of the
policy than they are to say (b) that the helping of the environment was the same
thing as the implementing of the policy (Ulatowski, 2008).
Furthermore, sometimes whether one will judge that a particular person caused
a certain result can depend on whether one judges that the person is morally at
fault for doing what he or she did (Mackie, 1955; Dray, 1957; Alicke, 1992;
Thomson, 2003). For example, whether one judges that Tom’s omitting to water
the plants caused their death can depend in part on whether one thinks Tom was
obligated to water the plants.
So, it seems that relevant structural descriptions of the situations or actions
must themselves sometimes involve normative assessments.
Categorization versus Grammar
Many moral judgments offer a moral categorization of an action, person, or
situation. So, instead of drawing an analogy between moral theory and
generative grammar, some theorists (e.g. Stich, 1993) propose instead to draw
an analogy between moral theory and psychological theories of categorization,
especially theories in which an instance to be categorized is compared with a
number of standard prototypes (or exemplars) of various categories and is then
assigned the category of the prototype to which it is most similar (theories of
this sort are described in section 6 in Harman, Mason, & Sinnott‐Armstrong,
Chapter 6, this volume).
We can offer two preliminary comments on the categorization proposal: first, it
is not clear that a categorization approach competes with a moral grammar
approach. The two approaches may be apples and oranges. Categorization is a
mental process, i.e. the term refers to a form of cognitive processing. In
contrast, the proposal that there is a mental grammar is not, first and foremost,
(p.281) about mental processes at all. Rather, it is a proposal about the
structures of the contents of moral assessments, just as a linguistic grammar is a
proposal about the structures of linguistic expressions. It might both be true that
there is a moral grammar, and that we make moral judgments by applying
certain prototypes. This might be the case if, for instance, all of a person’s
“murder” prototypes shared certain basic features, e.g. intentional harm of an
innocent.
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-620
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-11#
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-621
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-651
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-626
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-597
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-582
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-650
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-647
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-7#
Linguistics and Moral Theorys 1
Page 9 of 24
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
Second, modern theories of concepts and categorization often emphasize the
multi‐faceted nature of concepts: they are not just prototypes; they also
embedded in theories (Murphy, 2004; Machery, 2009). Prototypes and exemplars
alone seem insufficient to explain certain features of categorization. If moral
concepts also have a theoretical component, as seems likely to be demonstrated
empirically, then that would leave the door open to a moral grammar approach
as part of a theory of mental processing.
“Competence–Performance Distinction”
In developing a theory of linguistic grammar, Chomsky (1965) distinguishes
between “competence” and “performance”—where these are technical terms
that are not used in their ordinary senses. By “competence” Chomsky means an
individual’s internalized grammar. He uses the term “performance” to refer to all
other factors that may affect the use of language. In Chomsky’s framework, an
individual’s linguistic intuitions may or may not reflect the individual’s
competence, in Chomsky’s sense.
For example, compare the following sentences:
This is the cheese that the mouse that the cat that the dog that the
man owns bothered chased ate.
This is the man that owns the dog that bothered the cat that chased
the mouse that ate the cheese.
Intuitively, the first sentence is hard to understand—deviant in some way—while the
second is much easier to understand. The relative deviance of the first sentence
appears to be the result of “center embedding.” When this center embedding is made
explicit, it can appear that the semantic content of the two sentences is similar:
This is the cheese that the mouse ate.
This is the cheese that the mouse that the cat chased ate.
This is the cheese that the mouse that the cat [that the dog bothered]
chased ate.
This is the cheese that the mouse that the cat [that the dog (that the
man owns)] bothered chased ate. (p.282)
The intuitive deviance of the first sentence may fail to reflect something about one’s
linguistic competence in the sense of an internalized grammar. It may instead be due,
for example, to limitations of one’s parsing strategies, in which case it may be purely
the result of performance limitations. The sentence may fit one’s I‐grammar, but
performance factors affect one’s ability to comprehend the sentence as fully
grammatical.
The linguistic analogy suggests that it might be fruitful to suppose that there is a
distinction between moral competence and performance in something like
Chomsky’s technical sense of the terms “competence” and “performance.” As in
the linguistic case, we might suppose that other considerations beyond one’s
assumed I‐morality affect moral intuitions. In addition to memory and other
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-634
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-625
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-592
Linguistics and Moral Theorys 1
Page 10 of 24
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
processing limitations, moral intuitions might be affected by “heuristics and
biases” (Gilovich, Griffin, & Kahneman, 2002; see Sinnott‐Armstrong et al.,
Chapter 7 in this volume; also Sunstein, 2004; Horowitz, 1998; Doris & Stich,
2005; Sinnott‐Armstrong, 2008), prejudices, emotional responses, self‐interest,
etc.7 For example, utilitarians such as Greene (2008) have argued that intuitions
that appear to conflict with utilitarianism are due to morally irrelevant emotional
factors (see also Baron, 1993).
Of course, in most cases, performance factors are not distorting; they are an
integral part of proper functioning. One needs an internal parsing strategy in
order to comprehend what others say, so that strategy is not necessarily
distorting, as it may be in the above example. Similarly, pace Greene, emotions
like disgust may play an important role in getting one to appreciate the moral
character of a situation—even though, in other cases, they can lead one astray.
Someone with full moral “competence” (that is, with an internalized moral
grammar) who is unable to experience relevant emotions may have trouble with
moral evaluations in the same way that someone with full linguistic
“competence” but with a damaged parser might be unable to understand other
speakers. Here we have in mind work on so‐called “psychopaths” or other brain‐
damaged patients who do not exhibit normal emotive responses. Research has
shown that such individuals make abnormal moral judgments (Blair, 1995, 1997;
Koenigs et al., 2007). Some of these patients might possess moral
“competence” (an internalized moral grammar) but lack certain critical
performance capabilities. Of course, whether or not this is a useful way to
understand such patients depends on future empirical work. We offer this only
as an example of how the distinction between performance and competence
might be put into play. (p.283)
More generally, our point is that in moral theory, as in linguistics, we can
consider whether it makes sense to postulate a distinction between an
individual’s internal moral grammar or moral “competence” in contrast with
various other factors that determine the person’s intuitions and actions. For
linguistics, the corresponding assumption has been theoretically fruitful and
illuminating. With respect to moral theory, the assumption amounts to assuming
that there is a distinctive moral competence or I‐moral‐grammar. A contrary
view is that one’s morality is determined entirely by performance factors, so that
for example one’s emotional responses are actually constitutive of one’s moral
view rather than merely enabling or distorting performance factors.
It may be that a distinction between competence and performance applies only
to certain parts of morality and not other parts. We come back to this possibility
below in considering whether to distinguish aspects of morality we share with
other animals and aspects that are distinctively human.
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-605
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-8#
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-648
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-614
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-596
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-645
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-609
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-585
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-586
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-587
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-623
Linguistics and Moral Theorys 1
Page 11 of 24
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
Summary of Suggested Issues about Generative Moral Grammar
We have so far discussed one idea from linguistics that might be relevant to the
theory of morality, namely, generative grammar—an explicit specification of the
grammatical properties of expressions of a language. The conception of
grammar as generative or explicit in this way has proved quite productive for
the study of language and some moral theorists try to see an analogous
conception of moral grammar that might be productive for the study of morality.
Among the issues such an analogy raises for moral theory are (1) whether the
useful unit of analysis for moral theory is an individual’s I‐grammar, in contrast,
for example, with the moral conventions of a group; (2) whether and how such a
moral grammar might associate structural descriptions of actions, situations,
etc. with normative assessments; (3) whether and how the rules of such a moral
grammar might involve recursive embedding of normative assessments; and (4)
whether it is useful to distinguish moral “competence” from moral
“performance,” using these terms in the technical senses employed in linguistic
theory.
We now turn to a related issue about linguistic and moral universals. Features of
language are universal if they occur in all natural human languages that
children can acquire as first languages. Universal grammar is concerned in part
(p.284) with such linguistic universals and also with limits on ways in which
natural human languages can differ. To what extent might it be fruitful to
develop an analogous universal moral theory which would seek to describe
common features of any moralities children can acquire as first moralities—
along with an account of how moralities can differ?
Some people believe that various familiar moral principles are universal in the
sense that they are cross‐culturally accepted: “You should not steal.” “It is wrong
to murder or harm others.” “You should not tell lies.” “You should do your part in
useful group activities.” “You should help those worse off than yourself.” “You
should not treat others as you would not want to be treated.” “You should do
unto others what you would want them to do unto you.”
Instead of directly discussing whether such familiar principles are universally
accepted, we shall focus on two other sorts of universality suggested by the
analogy with linguistics. (That is, we shall not discuss whether everyone, cross‐
culturally, believes that it is wrong to murder; instead, we shall consider
universality of the sort that linguists find.) The first sort of universality is the
existence of universal constraints on possible I‐moral grammars; the second is
the notion of universal parameters (with local variance in parameter settings).
Linguistics and Moral Theorys 1
Page 12 of 24
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
Constraints on Linguistic Grammars
Universal linguistic grammar is concerned in part with constraints on the rules
of particular I‐grammars. Such grammars might contain two sorts of rules—
phrase structure rules for forming phrases out of their parts and transformations
or movement rules for reorganizing these parts (Chomsky, 1957, 1965; Freidin,
1992; van Riemsdijk & Williams, 1986).
For example, an adjective like green can combine or “merge” with a noun like
apple to form a noun phrase green apple; a determiner like a or the can merge
with a noun phrase to form a determiner phrase like a green apple; a preposition
like from can merge with a determiner phrase to form a prepositional phrase like
from a green apple; and so on. These are instances of phrase structure rules.
Universal grammar places constraints on the phrase structure rules for a given
I‐language. One such constraint implies that, if there are prepositions and so
prepositional phrases in which prepositions appear before their objects, then the
same must be true (p.285) of verb phrases and noun phrases: verbs must
appear before their objects and nouns before their objects. On the other hand, if
instead of prepositions there are postpositions and postpositional phrases in
which postpositions appear after their objects, then the same must be true of
verb phrases and noun phrases: verbs and nouns must appear after their
objects.
Transformations allow items in phrase structures to be rearranged. For example,
a transformation (wh‐movement) might allow the derivation of the structure
underlying a relative clause like (2) from something like the structure underlying
(1).
(1) Bob gave which to Bill.
(2) which Bob gave to Bill.
A different transformation (topicalization) might allow the derivation of the structure
underlying (4) from something like the structure underlying (3).
(3) Bob gave a book to Bill.
(4) A book, Bob gave to Bill.
Chomsky (1964) observed that there must be constraints of various sorts on such
transformations. He proposed an “A over A” constraint that would rule out moving an
item of a given type A from a larger phrase of type A. That turned out to be overly
simple and Ross (1967) proposed to replace it with a number of more specific “island
constraints,” where “islands” are structures that items cannot be moved out of. Ross’s
constraints include a “coordinate structure constraint” and a “complex NP constraint,”
among others; here we describe only his coordinate structure constraint.
Coordinate structures are special cases of the A over A principle in which an
item of type A immediately contains two or more coordinate items of type A, as in
the noun phrase: a man, a woman, and a child; or the verb phrase: kissed Jack or
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-590
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-592
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-603
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-642
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-591
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-643
Linguistics and Moral Theorys 1
Page 13 of 24
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
hugged Harry. These larger phrases themselves immediately contain smaller
coordinate phrases of the same type.
Ross’s coordinate structure constraint prevents transformations from moving
any coordinate item out of a coordinate structure and also prevents
transformations from moving anything contained in a coordinate item out of that
item. Suppose, for example, there is a transformation of topicalization that starts
with (5) and moves a book to the front to yield (6):
(5) Bob gave a book to Bill.
(6) A book, Bob gave to Bill.
Ross’s constraint would prevent topicalization from moving a book from a position in a
coordinate structure like (7) to yield (8).
(7) Bob gave a record and a book to Bill.
(8) * A book, Bob gave a record and to Bill.
(p.286)
Observe that it is unclear how a coordinate structure constraint of this sort
could be learned by children either from explicit instruction or by trial and error.
No one had heard about this constraint before Ross (1967), so no one could have
told children about it before that time. Furthermore, children never seem to
make errors that consist in violating the constraint, so they don’t seem to
acquire it by trial‐and‐error learning.
One way to explain the above facts about acquisition is to postulate that children
are predisposed towards certain linguistic constraints. If there is such a
predisposition, however, it would be present in all normally developing children.
Of course, it is possible to invent and use an artificial language that is not
subject to the constraint. But such a language might not be easily acquired as a
first language by children. We might expect them to acquire a variant that is
subject to the constraint. Furthermore, we might expect that children not
exposed to human language (e.g. because they were deaf) who developed a
language among themselves (presumably a sign language) would develop a
language that was subject to the constraint. There is evidence that this
expectation is correct. Padden (1988) claims that coordinate clauses in American
Sign Language (ASL) adhere to the coordinate structure constraint.
Constraints on Moralities?
A number of moral philosophers (Foot, 1978; Quinn, 1993) have suggested that
ordinary moral intuitions obey certain non‐obvious rules. For instance, it may be
that moral intuitions reflect some version of the Principle of Double Effect. Here
is one version of that principle:
Double Effect: It is worse knowingly to harm one person X in saving
another Y when the harm to X is intended as part of one’s means to
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-643
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-638
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-602
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-640
Linguistics and Moral Theorys 1
Page 14 of 24
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
saving Y as opposed to when the harm to X is merely a foreseen
unintended side‐effect of one’s attempt to save Y.
For another example, Thomson (1978) suggests that some ordinary moral
intuitions might reflect a principle of the following sort:
Deflection: It is better to save a person X by deflecting a harmful
process onto another person Y than by originating a process that
harms Y.
As with Ross’s coordinate structure constraint, these principles are not generally
recognized and it is therefore unlikely that they are explicitly taught to children.
If children do acquire moralities containing such principles, it would seem that
they must be somehow predisposed to do so. If so, we might (p.287) expect
such principles to be found in all moralities that children naturally acquire.8
As in the linguistic case, even if there are such constraints, it may be possible to
construct more or less artificial moral systems not subject to them—utilitarian
moralities, for example. But we might then expect that children of utilitarians
would not easily and naturally acquire their parents’ morality but would acquire
a version subject to those constraints. John Stuart Mill (1859, 1863, 1873) might
be an example. His version of utilitarianism appears to incorporate various
principles that do not fit with the version taught him by his father, James Mill
(1828). In particular, John Stuart Mill’s views about the importance of personal
creativity, self‐development, and personal liberty led him to adopt a distinction
between higher and lower qualities of pleasure that conflicts with the purely
quantitative conception of amount of pleasure accepted by Bentham and James
Mill.
Linguistic Principles and Parameters
We now turn to a discussion of linguistic principles and parameters and possible
analogies for moral theory.
Baker (2001) begins a highly readable account of current linguistic theory with
the story of eleven Navajo Code Talkers. During World War II, Allied forces in
the Pacific were in difficulty because Japanese cryptographers were able to
decode all their messages. The US Marine Corps solved the problem by using
the Code Talkers to transmit and receive military communications in Navajo.
Japanese cryptographers were unable to “decode” these messages.
Baker notes that this illustrates two important points about languages. On the
one hand, Navajo is sufficiently different from English and Japanese that skilled
cryptographers were unable to decode the Navajo messages. On the other hand,
Navajo is sufficiently similar to English that the Code Talkers could immediately
translate messages from English into Navajo and from Navajo back into English.
How can we account for both the similarities and the differences? Indeed, how
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-649
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-631
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-632
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-633
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-630
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-584
Linguistics and Moral Theorys 1
Page 15 of 24
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
can we account for the fact that a child exposed to a language when young will
easily acquire that language, when the best cryptographers are unable to make
head or tail of it?
Part of the answer why a child can so easily learn a language may be that the
child is somehow prepared to acquire a language whose syntax is largely
determined by the setting of a small number of parameters—a parameter that
(p.288) indicates whether a sentence must have a subject as in English, or not
as in Italian; a parameter that indicates whether the head of a phrase precedes
its complements, as in English, or follows, as in Japanese; and so on. Baker
suggests that it may be possible to give something like a structured periodical
table of languages, much as there is a structured periodical table of chemical
elements. (Hence his title, The Atoms of Language.)
A child picks up language from its interactions with others; there is no need for
explicit teaching. According to the principles and parameters theory, the child
uses its limited experience to set a relatively small number of parameters and
acquire the “core” syntax of a language. Other “non‐core” aspects of syntax,
perhaps involving stylistic variation, are learned as exceptions.
Universal grammar includes an account of principles and parameters. To repeat
a point mentioned above in connection with universal constraints on rules, it is
possible to devise and use a language that is not in accord with the principles
and parameters of universal grammar, and people can learn to use that
language. But children will not naturally acquire that language in the way they
acquire languages that are in accord with universal grammar.
Another aspect of the principles and parameters universal grammar approach in
linguistics is that a theorist’s conception of the generative grammar of a given
language can be affected by considerations involving other languages and ease
of language acquisition. The task of the child learner is easier to the extent that
as much as possible of grammar is part of universal grammar and so is built into
the child’s innate language faculty. So, for example, instead of detailed rules for
noun phrases, verb phrases, adjective phrases, and prepositional or
postpositional phrases, the basic theory of phrase structure might become part
of universal grammar, perhaps with a single phrase structure rule (merge X with
something to form a phrase), so that the grammar of a particular language like
English would be limited to such matters as whether the head of a phrase (noun,
verb, adjective, preposition or postposition) comes before or after its
complement. The grammar of a particular language like English might contain
no special rearrangement or transformational rules; instead there might be a
single rule of universal grammar (move X somewhere) that says something can
be moved anywhere, subject to universal constraints (Chomsky, 1981; van
Riemsdijk & Williams, 1986).
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-593
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-642
Linguistics and Moral Theorys 1
Page 16 of 24
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
One possible parameter might be whether or not the grammar allows such
movement or rearrangement. At least one currently existing language appears
not to allow this, the Amazonian Pirahã (Everett, 2005). Bolender (2007) argues
that all or most languages existing more than 40,000 years ago would probably
not have allowed such movement. (p.289)
The linguistic analogy suggests considering whether it is possible to develop a
universal theory of morality involving relevant principles and parameters.
How Moralities Differ: Principles and Parameters?
Moralities differ about abortion and infanticide, about euthanasia, about slavery,
about the moral status of women, about the importance of chastity in women
and in men, about caste systems, about cannibalism, about eating meat, about
how many wives a man can have at the same time, about the relative importance
of equality versus liberty, the individual versus the group, about the extent to
which one is morally required to help others, about duties to those outside one
or another protected group, about the importance of religion (and which religion
is important), about the importance of etiquette (and which standards of
etiquette), about the relative importance of personal virtues, and so on.
What are the prospects of a principles and parameters approach to explaining
these and other moral differences? Perhaps all moralities accept principles of
the form: avoid harm to members of group G, share your resources with
members of group F, etc., where G and F are parameters that vary from one
morality to another. However, Prinz (2008) points out that such parameters
would differ from those envisioned in linguistics by having many possible values,
whereas the parameters in linguistics have a small number of values, typically
two.
Dworkin (1993) argues that all intuitive moralities take human life to have what
he calls a “sacred value” that increases during pregnancy and in childhood,
reaching a maximum at the beginning of adulthood, and slowly declining
thereafter, so that the death of 20‐year‐old person is worse (“more tragic”) than
the death of an infant or someone who has reached old age. However, according
to Dworkin, individual moralities differ concerning whether an early embryo or
fetus has significant sacred value. Dworkin argues that this difference is what
lies behind ordinary (as opposed to theoretical) disagreements about abortion.
Perhaps there are different settings of a “sacred value” parameter. In this case
there might be only two relevant settings, indicating whether there is significant
sacred value in the embryo or fetus immediately after conception.
A somewhat different appeal to parameters can perhaps be found in the work of
certain anthropologists who see a small number of ways of organizing societies,
with differences in which ways organize which aspects. Fiske (1991, 1992)
distinguishes four general types of social organization: communal sharing,
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-599
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-589
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-639
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-598
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-600
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-601
Linguistics and Moral Theorys 1
Page 17 of 24
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
authority ranking, equality matching, and market pricing. Shweder et al. (1997)
distinguish the “big three” of morality: autonomy, community, and divinity. In
these cases, the relevant “parameters” might determine which models fit which
aspects of organization in a given society. (p.290)
However, we shall need to have more explicit generative accounts of a variety of
moralities before being able to think clearly about moral principles and
parameters.
Does Morality have to be Taught?
Acquisition of language does not require explicit instruction. Children pick it up
from the locals even if parents do not know the local language. The linguistic
analogy suggests we consider whether morality is like that.
How does a child acquire an I‐morality? Does the child pick up morality from
others in something like the way in which language is picked up? Does morality
have to be taught? Can it be taught?
Perhaps all normal children pick up the local morality much as they pick up the
local dialect, whether or not they receive explicit instruction in right and wrong.
Children with certain brain defects might have trouble acquiring morality, just as
children with certain other brain defects have trouble acquiring language.
Prinz (2008) suggests that moral conventions can be picked up in the way other
conventions are, with no need for moral universals, principles and parameters,
and other aspects of moral grammar. Similar suggestions were raised as
objections against universal grammar in linguistics in the 1960s (e.g. Harman,
1967, later retracted in Harman, 1973). But today universal grammar is a
central part of contemporary linguistics (Baker, 2001).
Core Morality?
As mentioned above, some linguists distinguish “core” aspects of grammar from
other “peripheral” aspects. In their view, the “core” grammar of an I‐language is
constituted by those aspects of grammar acquired just by setting parameters.
Peripheral aspects of the I‐language are special language‐specific rules and
exceptions that are acquired in other ways. Examples might include irregular
verbs and special “constructions” as in the sentence John sneezed the tissue
across the table, where the normally intransitive verb sneeze is used in a special
way. However, the distinction between core and peripheral aspects of grammar
is controversial, with some linguists (Newmeyer, 2005) treating it as a matter of
degree and others rejecting it altogether (Culicover and Jackendoff, 2005;
Goldberg, 1995).
The linguistic analogy suggests there might be a similar controversy within
moral theory about the usefulness of a distinction between core and peripheral
aspects of morality, where the core aspects are universal or are anyway acquired
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-644
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-639
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-610
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-611
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-584
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-635
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-595
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-607
Linguistics and Moral Theorys 1
Page 18 of 24
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
via parameter setting and other aspects are acquired in other ways. For
example, (p.291) it might be that people explicitly accept certain moral
principles on the basis of authority, much as they might accept certain
grammatical principles on the basis of authority. Where some people accept,
purely on the basis of authority, that it is ungrammatical to end an English
sentence with a preposition, others might accept, purely on the basis of
authority, that euthanasia is always wrong even to prevent prolonged terrible
suffering when someone is dying. This is just an example; we do not know
whether a distinction between core and peripheral moral principles will prove to
be theoretically useful in thinking about morality.
Humans and Other Animals
Certain aspects of language are similar to aspects of communication systems in
other animals—the use of certain sounds and gestures, for example. Other
aspects are particular to human language, arguably the existence of recursive
embedding rules of grammar of the sort discussed above. An analogy between
language and morality suggests that it might be useful to consider what aspects
of morality are shared with animals and what aspects are specific to human
moralities.
Recall Fiske’s four models of social relations: communal sharing, authority
ranking, equality ranking, and market pricing. Bolender (2003) notes that the
first three of these models, or something like them, can be found among non‐
human social animals. Furthermore, only two of Fiske’s models, communal
sharing and egalitarian equality matching, seem to have existed in most groups
of hunter‐gatherers, a mode of life “which represents over ninety‐nine percent of
human existence” (Bolender, 2003: 242). Bolender offers “evidence that humans
do not need hierarchy in groups of 150 or less” (245) so that this form of social
relation “was little used before the advent of plowing . . . [which] forced people
to allocate meager resources in large groups” (245). Similarly, there was no
previous need for market pricing.
Some questions arise here. Are the moralities of humans who use only these two
models (communal sharing and egalitarian equality matching) significantly
different from the moralities of social animals who use these models? Do such
humans have moralities with anything like the recursive properties mentioned
above, so that they accept such principles as that it is wrong to encourage
someone to do something that it is wrong to do? (Do non‐human animals accept
such principles?)
One way to test speculation in this area might be to look at the morality of the
Pirahã mentioned above, a small group of people with limited contact with
others until recently. The Pirahã have a very limited sense of absent (p.292)
items; their language does not appear to use any sort of quantification or other
constructions that involve variable binding operators (Everett, 2005; Bolender,
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-588
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-588
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-599
Linguistics and Moral Theorys 1
Page 19 of 24
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
2003, 2007). Presumably they cannot even express in their language the
principle that it is wrong to encourage someone to do what is wrong. A Pirahã
infant who was removed from that culture and brought up by adoptive parents in
New Jersey would acquire normal human language, an ability to use
quantification, and similar moral principles to those accepted by others in New
Jersey. On the other hand, nothing similar is true of a non‐human social animal.
In this chapter, we have described various issues suggested by the analogy
between morality and linguistics. In the first part of the chapter, we discussed
what might be involved in a generative “moral grammar.” In the second part we
discussed what might be involved in a universal moral grammar that parallels
universal linguistic grammar.
We believe that the issues suggested by such analogies are worth pursuing. In
particular, we think that looking at the developed science of linguistics suggests
new sorts of research that might be useful in thinking about the complicated
structure of morality. First, if moral theory is modeled on linguistics, its primary
focus would be on an individual’s I‐morality. Second, it may be useful to
distinguish performance and competence. Third, as in linguistics, moral theory
may benefit by distinguishing between the principles characterizing a moral
grammar and the algorithms actually used in the mental production of
grammatical sentences. Fourth, the linguistic analogy suggests ways to think
systematically about how moralities differ from each other and whether there
are moral universals. Finally, the analogy frames fruitful questions regarding
innateness of morality, the existence of a core morality, and the comparison of
human and animal morality.
In summary, linguistics has successfully developed into a mature science of
human language. In considering the linguistic analogy, then, we have been
exploring one possible way to develop a mature science of human morality.
References
Bibliography references:
Alicke, M. D. (1992). “Culpable Causation,” Journal of Personality and Social
Psychology 63: 368–378. (p.293)
Benedict XVI, Pope (2006). “The Human Person, the Heart of Peace,” Vatican
City. Available online at
Baker, M. C. (2001). The Atoms of Language: The Mind’s Hidden Rules of
Grammar. New York: Basic Books.
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-588
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-589
http://www.catholic.org/international/internationalstory.php?id=22342
http://www.catholic.org/international/internationalstory.php?id=22342
Linguistics and Moral Theorys 1
Page 20 of 24
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
Baron, J. (1993). “Heuristics and biases in equity judgments: a utilitarian
approach.” In B. A. Mellers and J. Baron (eds.), Psychological Perspectives on
Justice: Theory and Applications. New York: Cambridge University Press, 109–
137.
Blair, Robert James (1995). “A Cognitive Developmental Approach to Morality:
Investigating the Psychopath,” Cognition 57: 1–29.
—— (1997). “Moral Reasoning and the Child with Psychopathic Tendencies,”
Personality and Individual Differences 26: 731–739.
Bolender, J. (2003). “The Genealogy of the Moral Modules,” Minds and Machines
13: 233–255.
—— (2007). “Prehistoric Cognition by Description: A Russellian Approach to the
Upper Paleolithic,” Biology and Philosophy 22: 383–399.
Chomsky, N. (1957). Syntactic Structures. The Hague: Mouton.
—— (1964). Current Issues in Linguistic Theory. The Hague: Mouton.
—— (1965). Aspects of the Theory of Syntax. Cambridge, MA: MIT Press.
—— (1981). Lectures on Government and Binding. Dordrecht: Foris.
—— (2000). New Horizons in the Study of Language and the Mind. Cambridge:
Cambridge University Press.
Culicover, P. W. & Jackendoff, R. (2005). Simpler Syntax. New York: Oxford
University Press.
Doris, J. M. & Stich, S. P. (2005). “As a Matter of Fact: Empirical Perspectives on
Ethics.” In F. Jackson & M. Smith (eds.), The Oxford Handbook of Contemporary
Philosophy. Oxford: Oxford University Press, 114–152.
Dray, W. (1957) Laws and Explanation in History. London: Oxford University
Press.
Dworkin, R. (1993). Life’s Dominion: An Argument about Abortion, Euthanasia,
and Individual Freedom. New York: Knopf.
Everett, D. (2005). “Cultural Constraints on Grammar and Cognition in Pirahã,”
Current Anthropology 46: 621–634.
Fiske, A. P. (1991). Structures of Social Life: The Four Elementary Forms of
Human Relations. New York: Free Press.
—— (1992). “The Four Elementary Forms of Sociality: Framework for a Unified
Theory of Social Relations,” Psychological Review 99: 689–723.
Linguistics and Moral Theorys 1
Page 21 of 24
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
Foot, P. (1978). Virtues and Vices and Other Essays in Moral Philosophy.
Berkeley, CA: University of California Press; Oxford: Blackwell.
Freidin, R. (1992). Foundations of Generative Syntax. Cambridge, MA: MIT
Press.
Geach, P. T. (1965). “Assertion,” Philosophical Review 69: 449–465.
Gilovich, T., Griffin, D., & Kahneman, D. (eds.) (2002). Heuristics and Biases.
Cambridge: Cambridge University Press.
Gigerenzer, Gerd (2007). Gut Feelings. New York: Viking Penguin. (p.294)
Goldberg, A. (1995). Constructions: A Construction Grammar Approach to
Argument Structure. Chicago, IL: University of Chicago Press.
Goldman, A. I. (1970). A Theory of Human Action. Englewood Cliffs, NJ: Prentice‐
Hall.
Greene, J. (2008). “The Secret Joke of Kant’s Soul.” In W. Sinnott‐Armstrong
(ed.), Moral Psychology, Vol. 3: The Neuroscience of Morality: Emotion, Disease,
and Development. Cambridge, MA: MIT Press, 35–79.
Harman, G. (1967). “Psychological aspects of the theory of syntax,” Journal of
Philosophy 64: 75–87.
—— (1973). Review of Noam Chomsky, Language and Mind, Revised Edition, in
Language 49: 453–464.
Hauser, M. (2006). Moral Minds: How Nature Designed a Universal Sense of
Right and Wrong. New York: Ecco Press/HarperCollins.
Holyoak, K. J. & Thagard, P. (1995). Mental Leaps: Analogy in Creative Thought.
Cambridge, MA: MIT Press.
Horowitz, T. (1998). “Philosophical Intuitions and Psychological Theory.” In M.
DePaul & W. Ramsey (eds.), Rethinking Intuition. Lanham, MD: Rowman &
Littlefield, 143–160.
Isac, D. & Reiss, C. (2008). I‐Language: An Introduction to Linguistics as a
Cognitive Science. New York: Oxford University Press.
Kamm, F. (1992). Creation and Abortion. New York: Oxford University Press.
—— (1993). Morality, Mortality, Vol. 1: Death and Who to Save from It. New York:
Oxford University Press.
—— (1996). Morality, Mortality, Vol. 2: Rights, Duties, and Status. New York:
Oxford University Press.
Linguistics and Moral Theorys 1
Page 22 of 24
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
—— (2007). Intricate Ethics: Rights, Responsibilities, and Permissible Harm.
New York: Oxford University Press.
Knobe, J. (2003). “Intentional Action and Side Effects in Ordinary Language,”
Analysis 63: 190–193.
—— (2007). “Reason Explanation in Folk Psychology.” Midwest Studies in
Philosophy 31: 90–106.
—— (2008). “Action Trees and Moral Judgment.” Unpublished manuscript.
University of North Carolina–Chapel Hill.
Koenigs. M., Young, L., Adolphs, R., Tranel, D., Hauser, M., Cushman, F., &
Damasio, A. (2007). “Damage to the prefrontal cortex increases utilitarian moral
judgments.” Nature, 19.
Liberman, M. (2006). “Chomsky and the Pope: Separated at Birth?” Posting on
Language Log, 20 December 2006 at
Machery, E. (2009). Doing Without Concepts. New York: Oxford University Press.
Mackie, J. L. (1955). “Responsibility and Language.” Australasian Journal of
Philosophy 33: 143–159. (p.295)
Marr, D. (1982). Vision: A Computational Approach. San Francisco, CA:
Freeman.
Mikhail, J. (2000). Rawls’ Linguistic Analogy: A Study of the “Generative
Grammar” Model of Moral Theory Discussed by John Rawls in A Theory of
Justice. PhD dissertation in Philosophy, Cornell University.
—— (2007). “Universal Moral Grammar: Theory, Evidence, and the Future,”
Trends in Cognitive Sciences 11: 143–152.
Mill, J. (1828). Essays on Government. London: J. Innes.
Mill, J. S. (1859). On Liberty. London: J. W. Parker.
—— (1863). Utilitarianism. London: Parker, Son & Bourn.
—— (1873). Autobiography. London: Longmans, Green, Reader, & Dyer.
Murphy, Gregory M. (2004). The Big Book of Concepts. Cambridge, MA: MIT
Press.
Newmeyer, F. J. (2005). Possible and Probable Languages. Oxford: Oxford
University Press.
http://itre.cis.upenn.edu/~myl/languagelog/archives/003940.html
http://itre.cis.upenn.edu/~myl/languagelog/archives/003940.html
Linguistics and Moral Theorys 1
Page 23 of 24
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
Nozick, R. (1968). “Moral Complications and Moral Structures,” Natural Law
Forum 13: 1–50.
—— (1997). Socratic Puzzles. Cambridge, MA: Harvard University Press.
Padden, C. A. (1988). “Grammatical Theory and Signed Languages.” In F.
Newmeyer (ed.), Linguistics: The Cambridge Survey (Vol. II). Cambridge & New
York: Cambridge University Press.
Prinz, J. (2008). “Is Morality Innate?” Moral Psychology, Volume 1: The Evolution
of Morality: Adaptations and Innateness, W. Sinnott‐Armstrong (ed.). Cambridge,
MA: MIT Press, 367–406.
Quinn, W. (1993). Morality and Action. Cambridge: Cambridge University Press.
Rawls, J. (1971). A Theory of Justice. Cambridge, MA: Harvard University Press.
van Riemsdijk, H. & Williams, E. (1986). Introduction to the Theory of Grammar.
Cambridge, MA: MIT Press.
Ross, J. R. (1967). Constraints on Variables in Syntax. Cambridge, MA: MIT PhD
Dissertation. (Published as Infinite Syntax, Ablex, 1986.)
Shweder, R. A., Much, N. C., Mahapatra, M., & Park, L. (1997). “The ‘Big Three’
of Morality (Autonymy, Community, and Divinity) and the ‘Big Three’
Explanations of Suffering as Well.” In A. Brandt & P. Rozen (eds.), Morality and
Health. New York: Routledge, 119–169.
Sinnott‐Armstrong, W. (2008). “Framing Moral Intuitions.” In W. Sinnott‐
Armstrong (ed.), Moral Psychology, Volume 2: The Cognitive Science of Morality.
Cambridge, MA: MIT Press, 47–76.
Smith, A. (1817). The Theory of Moral Sentiments. Philadelphia, PA: Anthony
Finley.
Stich, S. (1993). “Moral Philosophy and Mental Representation.” In M. Hechter,
L. Nadel, & R. Michod (eds.), The Origin of Values. New York: Aldine de Gruyter,
215–228.
Sunstein, C. (2004). “Moral Heuristics and Moral Framing,” Minnesota Law
Review 88:1556. (p.296)
Thomson, J. J. (1978). “Killing, Letting Die, and the Trolley Problem,” Monist 59:
204–217.
—— (2003). “Causation: Omissions,” Philosophy and Phenomenological Research
66: 81–103.
Linguistics and Moral Theorys 1
Page 24 of 24
PRINTED FROM OXFORD SCHOLARSHIP ONLINE (oxford.universitypressscholarship.com). (c) Copyright Oxford University Press,
2021. All Rights Reserved. An individual user may print out a PDF of a single chapter of a monograph in OSO for personal use.
Subscriber: University of Central Florida; date: 15 February 2021
Ulatowski, J. (2008). “Action Under a Description.” Unpublished manuscript.
University of Wyoming.
Notes:
(1) We have benefited from comments on an earlier version of this chapter from
members of our moral psychology research group at a workshop in Pittsburgh
November 2007 and more recently from detailed comments from Stephen Stich.
(2) We are indebted here to Richard Holton.
(3) Here we are indebted to Liberman (2006).
(4) Mikhail (2000; see also 2007) suggests that relevant aspects of structure may
be arrived at through using Goldman’s (1970) theory of action and legal
concepts of battery, end, means, side effect, intention, etc. Abstractly
considered, the resulting structures are somewhat similar to the phrase
structures that figure in linguistic theory.
(5) We thank Jesse Prinz for bringing this point to our attention and John Doris
for the specific example.
(6) Such principles are famously problematic for expressivist theories that try to
explain the meaning of wrong simply by saying the word is used to express
disapproval (Geach, 1965).
(7) Of course, there is an ongoing debate whether such “biases” should or
should not be thought of as bad reasoning. See, for instance, Gigerenzer (2007).
(8) The linguistic analogy suggests that this remark is too strong. As we shall
explain, a principles and parameters approach of the sort we are about to
discuss allows for other possibilities.
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-624
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-628
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-629
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-608
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-604
https://oxford.universitypressscholarship.com/view/10.1093/acprof:oso/9780199582143.001.0001/acprof-9780199582143-chapter-9#acprof-9780199582143-bibItem-606
John M. Doris and The Moral Psychology Research Group
Linguistics and Moral Theorys 1
Erica Roedder
Gilbert Harman (Contributor Webpage)
Abstract and Keywords
Linguistics and Moral Theorys 1
1. Moral Grammar
Linguistics and Moral Theorys 1
Linguistics and Moral Theorys 1
I‐grammar
“Generative” Grammar
Linguistics and Moral Theorys 1
Linguistics and Moral Theorys 1
Recursive Embedding
Linguistics and Moral Theorys 1
Structural Dependence on Normative Assessments
Linguistics and Moral Theorys 1
Categorization versus Grammar
Linguistics and Moral Theorys 1
“Competence–Performance Distinction”
Linguistics and Moral Theorys 1
Linguistics and Moral Theorys 1
Summary of Suggested Issues about Generative Moral Grammar
2. Universal Morality
Linguistics and Moral Theorys 1
Constraints on Linguistic Grammars
Linguistics and Moral Theorys 1
Constraints on Moralities?
Linguistics and Moral Theorys 1
Linguistic Principles and Parameters
Linguistics and Moral Theorys 1
Linguistics and Moral Theorys 1
How Moralities Differ: Principles and Parameters?
Linguistics and Moral Theorys 1
Does Morality have to be Taught?
Core Morality?
Linguistics and Moral Theorys 1
Humans and Other Animals
Linguistics and Moral Theorys 1
3. Conclusion
Linguistics and Moral Theorys 1
Linguistics and Moral Theorys 1
Linguistics and Moral Theorys 1
Linguistics and Moral Theorys 1
Linguistics and Moral Theorys 1
Notes:
We provide professional writing services to help you score straight A’s by submitting custom written assignments that mirror your guidelines.
Get result-oriented writing and never worry about grades anymore. We follow the highest quality standards to make sure that you get perfect assignments.
Our writers have experience in dealing with papers of every educational level. You can surely rely on the expertise of our qualified professionals.
Your deadline is our threshold for success and we take it very seriously. We make sure you receive your papers before your predefined time.
Someone from our customer support team is always here to respond to your questions. So, hit us up if you have got any ambiguity or concern.
Sit back and relax while we help you out with writing your papers. We have an ultimate policy for keeping your personal and order-related details a secret.
We assure you that your document will be thoroughly checked for plagiarism and grammatical errors as we use highly authentic and licit sources.
Still reluctant about placing an order? Our 100% Moneyback Guarantee backs you up on rare occasions where you aren’t satisfied with the writing.
You don’t have to wait for an update for hours; you can track the progress of your order any time you want. We share the status after each step.
Although you can leverage our expertise for any writing task, we have a knack for creating flawless papers for the following document types.
Although you can leverage our expertise for any writing task, we have a knack for creating flawless papers for the following document types.
From brainstorming your paper's outline to perfecting its grammar, we perform every step carefully to make your paper worthy of A grade.
Hire your preferred writer anytime. Simply specify if you want your preferred expert to write your paper and we’ll make that happen.
Get an elaborate and authentic grammar check report with your work to have the grammar goodness sealed in your document.
You can purchase this feature if you want our writers to sum up your paper in the form of a concise and well-articulated summary.
You don’t have to worry about plagiarism anymore. Get a plagiarism report to certify the uniqueness of your work.
Join us for the best experience while seeking writing assistance in your college life. A good grade is all you need to boost up your academic excellence and we are all about it.
We create perfect papers according to the guidelines.
We seamlessly edit out errors from your papers.
We thoroughly read your final draft to identify errors.
Work with ultimate peace of mind because we ensure that your academic work is our responsibility and your grades are a top concern for us!
Dedication. Quality. Commitment. Punctuality
Here is what we have achieved so far. These numbers are evidence that we go the extra mile to make your college journey successful.
We have the most intuitive and minimalistic process so that you can easily place an order. Just follow a few steps to unlock success.
We understand your guidelines first before delivering any writing service. You can discuss your writing needs and we will have them evaluated by our dedicated team.
We write your papers in a standardized way. We complete your work in such a way that it turns out to be a perfect description of your guidelines.
We promise you excellent grades and academic excellence that you always longed for. Our writers stay in touch with you via email.