I’ve recently completed a research monograph on moral psychology and meta-ethics. Below you’ll find summaries of the book and its chapters. If you’d like to see any draft chapters, get in touch.
Regard for Reason in the Moral Mind
(forthcoming with Oxford University Press)
The burgeoning science of ethics has produced a trend toward pessimism. Ordinary moral thought and action, we’re told, are profoundly influenced by arbitrary factors and ultimately driven by unreasoned feelings. This book counters the current orthodoxy on its own terms by carefully engaging with the empirical literature. The resulting view, optimistic rationalism, maintains that reason plays a pervasive role in our moral minds and ultimately defuses sweeping debunking arguments in ethics. The science does suggest that moral knowledge and virtue don’t come easily. However, despite being heavily influenced by automatic and unconscious processes that have been shaped by evolutionary pressures, we needn’t reject ordinary moral psychology as fundamentally flawed or in need of serious repair. Reason can be corrupted in ethics just as in other domains, but a special pessimism about morality in particular is unwarranted.
Table of Contents
1. Empirical Pessimism
Part A: Moral Judgment & Knowledge
2. The Limits of Emotion
3. Reasoning Beyond Consequences
4. Defending Moral Judgement
5. The Difficulty of Moral Knowledge
Part B: Moral Motivation & Virtue
6. Beyond Self-Interest
7. The Motivational Power of Moral Beliefs
8. Freeing Reason from Desire
9. Defending Virtuous Motivation
10. Cautious Optimism
Scientifically-informed theories of ordinary moral thought and action are on the rise but trend toward pessimism. Many theorists argue that ordinary moral judgment involves little reasoning or not enough to yield justified belief, while others argue that we rarely act for the right reasons. This chapter describes such sources of empirical pessimism and outlines the remaining chapters that defend the alternative, optimistic rationalism, which allows for more virtue by according reason a central role in moral psychology. While the science doesn’t suggest that moral knowledge and virtuous motivation come easily, there is no reason to reject ordinary moral psychology as fundamentally flawed. This chapter also discusses some preliminaries, such as the reason/emotion dichotomy, non-cognitivism, and how to draw on empirical research.
Empirical research apparently suggests that emotions play an integral role in moral judgment. The evidence for sentimentalism is diverse, but it is rather weak and has generally been overblown. There is no evidence that our moral concepts themselves are partly comprised of or necessarily dependent on emotions. While the moral/conventional distinction may partly characterize the essence of moral judgment, moral norms needn’t be backed by affect in order to transcend convention. Priming people with incidental emotions like disgust doesn’t make them moralize actions. Finally, moral judgment can only be somewhat impaired by damage to areas of the brain that are generally associated with emotional processing. Psychopaths, for example, exhibit both emotional and rational deficits, and the latter alone can explain any minor defects in moral cognition.
Experimental research demonstrates that ordinary moral judgment involves both conscious and unconscious reasoning or inference. The evidence suggests in particular that we treat as morally significant more than the consequences of a person’s actions, including the distinctions between: intentional versus accidental outcomes, actions versus omissions, and harming as a means versus a byproduct. The result is an extremely minimal dual process model of moral judgment on which we at least compute both an action’s outcomes and the actor’s role in bringing them about. This view resembles the famous linguistic analogy in only its least controversial aspects, particularly the emphasis on unconscious reasoning in moral cognition.
Despite containing non-consequentialist elements and relying in part on automatic heuristics, ordinary moral thinking can rise to justified moral belief. Prominent, wide-ranging debunking arguments (based on evolutionary pressures, framing effects, automatic emotional heuristics, and disgust) are subject to a Debunker’s Dilemma: they can identify an influence on moral belief that is either substantial or defective, but not both. When one identifies a genuinely defective influence on a large class of moral beliefs (e.g. framing effects), this influence is insubstantial, failing to render the beliefs unjustified. When one identifies a main basis for belief (e.g. automatic heuristics), the influence is not roundly defective. There is ultimately a trade-off for sweeping debunking arguments in ethics: identifying a substantial influence on moral belief implicates a process that is not genuinely defective. We thus lack empirical reason to believe that moral judgment is fundamentally flawed.
The science does support more modest threats to moral knowledge. First, while wide-ranging debunking arguments are problematic, this does not hinder highly targeted attacks, particularly on specific sets of moral beliefs that experimental research reveals to be distinguished for morally irrelevant reasons. Second, controversial moral issues are disputed by many of one’s opponents. The science of political disagreement suggests that many ordinary people can’t claim to know what they believe about controversial moral issues that divide liberals and conservatives within a culture. Such controversial moral beliefs are disputed by what one should regard as epistemic peers, at least because others are just as likely to be wrong, even if not right, due to cognitive biases that affect proponents of all ideologies.
This chapter introduces the long-standing idea that inappropriate motives, such as self-interest, can militate against virtuous motivation. Some theorists have tried to show that we are universally egoistic by appeal to empirical research, from evolutionary theory to the neuroscience of learning. However, these efforts fail and instead decades of experiments provide powerful evidence that we are capable of genuine altruism. We can be motivated ultimately by a concern for others for their own sake, especially when empathizing with them. The evidence does not show that empathy blurs the distinction between self and other in a way that makes helping behavior truly egoistic or non-altruistic.
Even if we can rise above self-interest, we may just be slaves of our passions. But the motivational power of reason, via moral beliefs, has been understated, even in the difficult case of temptation. Experiments show that often when we succumb, it is due in part to a change in moral (or normative) judgment. Rationalization, perhaps paradoxically, reveals a deep regard for reason, to act in ways we can justify to ourselves and to others. The result is that we are very often morally motivated or exhibit moral integrity. Even when behaving badly, actions that often seem motivated by self-interest are actually ultimately driven by a concern to do what’s right.
The previous chapter showed that our beliefs about which actions we ought to perform frequently have an effect on what we do. But Humean theories, holding that all motivation has its source in desire, insist on connecting such beliefs with an antecedent motive. However, we can allow moral (or normative) beliefs a more independent role to generate intrinsic desires. An anti-Humean theory provides perfectly ordinary and intelligible explanations of actions, which places a burden on Humeans to justify a more restrictive account. However, they do not discharge this burden on empirical grounds, whether by appealing to research on neurological disorders (acquired sociopathy, Parkinson’s, and Tourette’s), the psychological properties of desire, or the scientific virtue of parsimony.
This chapter considers remaining empirical challenges to the idea that we’re commonly motivated to do what’s right for the right reasons. Two key factors threaten to defeat claims to virtuous motivation, self-interest and arbitrary situational factors. Both threats aim to identify defective influences on moral behavior that reveal us to be commonly motivated by the wrong reasons. However, there are limits to such wide-ranging skeptical arguments. Ultimately, like debunking arguments, defeater challenges succumb to a Defeater’s Dilemma: one can identify influences on many of our morally-relevant behaviors that are either substantial or arbitrary, but not both. The science suggests a familiar trade-off in which substantial influences on many morally-relevant actions are rarely defective.
This chapter briefly draws out some main lessons from the previous chapters but contains a discussion of some their implications. Ordinary moral psychology is capable of rising to knowledge and virtue, in part because we do have a regard for reason, but we do often fail. When we do fail, though, the problem is not with moral psychology in particular but the ways in which reason can be corrupted generally. One broad implication of cautious optimism is that the best method for increasing virtue won’t target our passions to the exclusion of our (often unconscious) reasoning. However, sound arguments aren’t enough, for human beings are fallible creatures with limited attention spans. An educated and well-informed populace is necessary, but so are environments that facilitate good reasoning, not rationalization, particularly during moral learning and development.