ORIGINAL RESEARCH article
Volume 11 - 2017 | https://doi.org/10.3389/fnint.2017.00034
Deontological Dilemma Response Tendencies and Sensorimotor Representations of Harm to Others
- 1Ahmanson-Lovelace Brain Mapping Center, Brain Research Institute, University of California, Los Angeles, Los Angeles, CA, United States
- 2Department of Psychiatry and Biobehavioral Sciences, Semel Institute for Neuroscience and Human Behavior, Brain Research Institute, David Geffen School of Medicine, University of California, Los Angeles, Los Angeles, CA, United States
- 3Edie and Lew Wasserman Center, University of California, Los Angeles, Los Angeles, CA, United States
- 4Psychology Department, Florida State University, Tallahassee, FL, United States
- 5Department of Psychology, Social Cognition Center Cologne, University of Cologne, Cologne, Germany
The dual process model of moral decision-making suggests that decisions to reject causing harm on moral dilemmas (where causing harm saves lives) reflect concern for others. Recently, some theorists have suggested such decisions actually reflect self-focused concern about causing harm, rather than witnessing others suffering. We examined brain activity while participants witnessed needles pierce another person’s hand, versus similar non-painful stimuli. More than a month later, participants completed moral dilemmas where causing harm either did or did not maximize outcomes. We employed process dissociation to independently assess harm-rejection (deontological) and outcome-maximization (utilitarian) response tendencies. Activity in the posterior inferior frontal cortex (pIFC) while participants witnessed others in pain predicted deontological, but not utilitarian, response tendencies. Previous brain stimulation studies have shown that the pIFC seems crucial for sensorimotor representations of observed harm. Hence, these findings suggest that deontological response tendencies reflect genuine other-oriented concern grounded in sensorimotor representations of harm.
Imagine watching a video of a hypodermic syringe slowly and deliberately piercing a human hand. Despite knowing that the hand belongs to someone else, would you wince and reflexively withdraw your hand, or shrug and stay put? Would you empathize with the person getting pierced? If so, what does your reaction reveal about your moral psychology? The dual-process model of dilemma judgments (Greene et al., 2001) suggests that when people face dilemmas where causing harm saves lives, rejecting such harm (despite not saving lives) reflects other-oriented affective processing. In contrast, judgments to accept harm (thereby maximizing outcomes) reflect cognitive evaluations of outcomes. On such dilemmas, rejecting harm is said to uphold deontological morality, where the morality of actions derives from their intrinsic nature (Kant, 1785/1959), whereas accepting harm is said to uphold utilitarian morality, where the morality of actions derives from their outcomes (Mill, 1861/1998). Although considerable evidence supports the dual process model (e.g., Bartels, 2008; Amit and Greene, 2012; Conway and Gawronski, 2013; Gleichgerrcht and Young, 2013; cf. Kahane, 2015), some theorists have questioned whether the affective processing involved in harm-rejection truly reflects other-oriented concern. Affective reactions to harm in dilemma judgments may reflect self-focused emotions centered on causing harm, rather than genuine concern generated by others in pain (Miller et al., 2014).
When people know that others are experiencing pain, sensorimotor and affective systems in their brains respond as though they are personally experiencing pain (Singer et al., 2004; Avenanti et al., 2005; Lamm et al., 2011). This phenomenon, called neural resonance (Zaki and Ochsner, 2012), has also been documented for disgust (Wicker et al., 2003; Jabbi et al., 2007), emotions (Carr et al., 2003; Pfeifer et al., 2008), and motor behavior (Fadiga et al., 1995). A key moderator for the presence or absence of activity in sensorimotor circuits seems to be the direct observation of others experiencing pain (Avenanti et al., 2005). When others’ pain is communicated through symbolic cues (e.g., lightning bolts) rather than direct observation, affective circuits are activated, but sensorimotor circuits are not (Singer et al., 2004).
Although researchers conceptualize neural resonance as a component of empathy (Zaki and Ochsner, 2012), they have primarily examined immediate reactions to real-time stimuli, such as smiling faces or video clips of hands getting pierced by needles, rather than personality traits or decision-making tendencies in unrelated contexts. Yet, recent findings suggest that individual differences in neural resonance predict other aspects of an individual’s traits and behavior. Specifically, neural resonance for pain correlates with peoples’ tendency to take others’ perspectives and feel distressed by harm to others (Avenanti et al., 2009). Neural resonance for pain also correlates with charitable donations (Ma et al., 2011), helping behavior (Hein et al., 2011; Morelli et al., 2014), and generosity in economic games where strategic giving does not play any role (Christov-Moore and Iacoboni, 2016). Since correlation is not causation, we have further tested the meaning of these correlations with disruptive brain stimulation, showing that it is possible to modulate generosity by stimulating brain areas whose activity correlates with offers (Christov-Moore et al., 2017). Furthermore, deontological response tendencies on moral dilemmas correlate with empathic concern and perspective-taking (e.g., Conway and Gawronski, 2013; Gleichgerrcht and Young, 2013). These findings suggest that the neural mechanisms associated with pre-reflective reactions to others’ internal states may also be relevant to moral decision-making.
We propose that people who show greater neural resonance with the pain of others should evince stronger tendencies to reject harm in moral dilemmas, in line with the dual-process model (Greene et al., 2001). This finding would clarify that internal representations of others’ states (rather than just self-focused emotions) contribute to moral dilemma judgments. Such effects should pertain only to harm-rejection (deontological) response tendencies, which are linked to affective processing, rather than outcome-maximization (utilitarian) response tendencies, which are linked to cognitive processing (Conway and Gawronski, 2013). To examine this possibility, we employed process dissociation (PD, Jacoby, 1991) to disentangle the impacts of harm-rejection and outcome-maximization tendencies on conventional relative dilemma judgments (for a review of PD, see Payne and Bishara, 2009).
In the current work, we recorded participants’ brain activation while they watched videos of needles piercing hands (versus hands gently touched with a Q-tip). A month later or more, participants completed the PD moral dilemma battery, containing 10 moral dilemmas, each with two versions (Conway and Gawronski, 2013). Each dilemma entails causing harm to achieve a particular outcome: incongruent dilemmas correspond to conventional relative moral dilemmas where causing harm maximizes outcomes (hence deontological and utilitarian considerations conflict); congruent dilemmas involve causing harm that does not maximize outcomes (so deontological and utilitarian considerations align – both suggest rejecting harm). Participants indicated whether each harmful action was appropriate or not appropriate. By applying participant responses to both kinds of dilemmas to a processing tree (see Figure 1), we can mathematically represent the probability of accepting or rejecting harm in each case, and algebraically combine these equations to solve for two previously unknown variables: deontological inclinations (reflecting a pattern of consistently rejecting harm regardless of whether doing so maximizes outcomes), and utilitarian inclinations (reflecting a pattern of maximizing outcomes, whether or not doing so entails causing harm).
FIGURE 1. Processing tree illustrating the underlying components leading to judgments that harmful action is either acceptable or unacceptable in congruent and incongruent moral dilemmas, allowing researchers to estimate utilitarian response tendencies (maximize outcomes regardless of whether doing so causes harm or not) and deontological response tendencies (reject causing harm regardless of whether doing maximizes outcomes or not).
Conway and Gawronski (2013) found that deontological response tendencies uniquely correlated with measures of other-oriented concern, such as empathic concern and perspective-taking (Davis, 1983) and were uniquely increased when viewing photos of the victims of harm, clarifying similar findings on conventional dilemma judgments (e.g., Bartels, 2008; Gleichgerrcht and Young, 2013). Conversely, utilitarian response tendencies uniquely correlated with measures of cognitive processing, such as need for cognition (Epstein et al., 1996), and are uniquely impaired by cognitive load, again clarifying similar findings on relative judgments (e.g., Greene et al., 2008; Moore et al., 2008). Meta-analytic results indicated that deontological and utilitarian response tendencies are typically uncorrelated, but each correlates with conventional relative dilemma judgments in opposite directions (Friesdorf et al., 2015), suggesting that the PD parameters reflect two independent response tendencies that jointly influence conventional relative dilemma judgments.
Given research suggesting that deontological dilemma response tendencies reflect relatively affective other-oriented concern, whereas utilitarian dilemma response tendencies reflect relatively cognitive outcome-focused processing, we predicted that when viewing others in pain, activation in brain regions associated with pain processing would predict deontological, but not utilitarian, response tendencies. However, if Miller et al. (2014) are correct that deontological decisions primarily reflect self-focused concerns over causing harm, rather than other-focused affective processing regarding witnessing harm, then brain activation while witnessing others in pain (pain one did not cause) should not predict dilemma response tendencies.
Moreover, we aimed to clarify the nature of other-oriented concern involved in deontological dilemma decisions. As mentioned above, neural resonance for pain recruits both affective (anterior cingulate, medial prefrontal cortex, and anterior insula) and sensorimotor (ventral premotor cortex, inferior parietal lobe, and primary motor cortex) neural systems to different extents depending on stimuli and context (reviewed in Lamm et al., 2011). We employed stimuli known to activate both sensorimotor and affective systems (Avenanti et al., 2005; Bufalari et al., 2007) – videos of painful stimuli applied to others (Avenanti et al., 2005) – to examine which system best predicts deontological dilemma response tendencies, rather than employing stimuli known to selectively activate affective but not sensorimotor networks (as in Singer et al., 2004), which would prevent us from testing these hypotheses against one another.
If affective systems best predict deontological tendencies, this may suggest that deontological decisions primarily reflect general affective responses to others’ pain. On the other hand, if sensorimotor circuits best predict deontological tendencies, this may suggest that deontological decisions are partially grounded in participants’ previous personal perceptual and motor experiences of pain. In other words, witnessing a needle pierce someone else’s hand may lead participants to imagine, albeit likely implicitly, how others experience the physical sensation of having their hand pierced. Such physiological representations of others’ pain may lead participants to reject actions that cause others pain – such as endorsing deontological moral dilemma judgments. Note that whereas previous functional magnetic resonance imaging (fMRI) dilemma research examined brain activation as participants completed the dilemma task itself (e.g., Greene et al., 2001, 2004), we examined brain activation in response to unrelated stimuli – the needle task – and used this activation to predict dilemma responses made over 1 month later.
To recapitulate, this study tested two hypotheses. First, we examined whether neural responses to witnessing another person experience pain predicted subsequent deontological (but not utilitarian) dilemma response tendencies. Second, we investigated the nature of the functional processes associated with deontological tendencies: whether they are best predicted by systems involved in general affective processing, or by systems involved in the sensorimotor processing of watching others in pain.
Materials and Methods
We recruited 19 ethnically diverse adults aged 18–35 (9 female), through community fliers. We required that participants were right handed, with no prior or concurrent diagnosis of any neurological (e.g., epilepsy, Tourette’s syndrome), psychiatric (e.g., schizophrenia), or developmental disorders (e.g., ADHD, dyslexia), and no history of drug or alcohol abuse. All recruitment and experimental procedures were performed under approval of UCLA’s institutional review board.
Functional MRI Procedure
Pain Video Task
We employed (with permission) the 27 full-color video stimuli from Bufalari et al. (2007). Each video depicted the same human hand being pierced by a hypodermic syringe in varying locations (Pain condition), being touched by a wooden Q-tip in the same locations (Touch condition), or in isolation (Hand condition). The run consisted of 12 trial blocks lasting 26 s each, plus 8 alternating rest blocks that lasted either 5 or 10 s. Each trial block consisted of four videos of a single condition (Pain, Touch, Hand), each 5 s in duration, with an interstimulus interval of 400 ms. Subjects were simply instructed to watch the video clips. They were assured that the hand in the video clip was a human hand and not a model, but they were not instructed to empathize with the model, nor were there any audiovisual cues to indicate pain from the hand’s owner. We used (and controlled for) three different block orders, and ensured an approximately equal proportion of male and female subjects viewed each order. We coded the task within Presentation (created by Neurobehavioral Systems).
MR Image Acquisition
The fMRI data were acquired on a Siemens Trio 3 Tesla system housed in the Ahmanson-Lovelace Brain Mapping Center at UCLA. Functional images were collected over 36 axial slices covering the whole cerebral volume using an echo planar T2∗-weighted gradient echo sequence (TR = 2500 ms; TE = 25 ms; flip angle = 90°; matrix size = 64 × 64; FOV 20 cm; in-plane resolution = 3 mm × 3 mm; slice thickness = 3 mm/1 mm gap). Additionally, a high-resolution T1-weighted volume was acquired in each subject (TR = 2300 ms; TE = 25 ms; TI = 100 ms; flip angle = 8°; matrix size = 192 × 192; FOV = 256 cm; 160 slices), with approximately 1 mm isometric voxels (1.3 mm × 1.3 × 1.0 mm).
Functional MRI Analysis
Analyses were performed in FMRI Expert Analysis Tool (FEAT), part of FMRIB’s Software Library (FSL)1. After motion correction using MCFLIRT, images were temporally high-pass filtered with a cutoff period of 70 s (approximately equal to one rest-task-rest-task period), and smoothed using a 6 mm Gaussian FHWM algorithm in three dimensions. Each participants’ functional data were co-registered to standard space (MNI 152 template) via registration of an averaged functional image to the high resolution T1-weighted volume using a six degree-of-freedom linear registration and of the high-resolution T1-weighted volume to the MNI 152 template via non-linear affine registration, implemented in FNIRT.
In order to remove non-neuronal sources of coherent oscillation in the relevant frequency band (0.01–0.1 Hz), preprocessed data were subjected to probabilistic independent component analysis as implemented in Multivariate Exploratory Linear Decomposition into Independent Components (MELODIC) Version 3.10, part of FSL2. Noise components corresponding to head motion, scanner noise, and aliasing of cardiac/respiratory signals were identified by observing their localization, time series, and spectral properties (after Kelly et al., 2010) and removed using FSL’s regfilt command.
We performed statistical analyses of fMRI data using FSL’s implementation of the general linear model. For first level analyses gauging task activation, the anticipated BOLD response to each condition was modeled using an explanatory variable (EV) consisting of a boxcar function describing the onset and duration of each relevant experimental condition (task conditions, rest, and instruction screen) convolved with a canonical double-gamma hemodynamic response function (HRF) to produce an expected BOLD response. The temporal derivative of each task EV was also included in the model to improve the model fit and accommodate regional variations in the BOLD signal. Functional data were then fitted to the modeled BOLD signal using FSL’s implementation of the general linear model. This produced parameter estimate maps describing the goodness-of-fit of the data to the modeled BOLD signal. Contrasts were then performed to isolate clusters of voxels that showed significantly different parameter estimates between each condition. Higher-level analyses were implemented by including subjects’ scores on the deontological and utilitarian parameters as separate EV’s against subjects’ parameter estimate maps (superimposed in standard space) for the contrast Pain > Touch. Resultant whole-brain parameter estimates representing the between-subject variance explained by each behavioral EV were converted to normalized Z-scores and corrected for multiple comparisons at the cluster level (using Gaussian random field theory) using a cluster-wise Z-threshold of 2.3 and p-value cutoff of 0.05, using FLAME 1 + 2. The resultant Z-statistic images from the higher-level analysis were then masked by the activation map for the contrast Pain > Touch. The rationale for this masking procedure is straightforward: we were testing the hypothesis that task-related activation of neural systems engaged in processing others’ pain contributes to decision-making in moral dilemmas. Task-irrelevant areas cannot speak to this hypothesis.
Process Dissociation Dilemma Battery
In a separate experimental session conducted at least 1 month following the neuroimaging protocol, participants completed a battery of 10 moral dilemmas, each with two versions: an incongruent and congruent version (Conway and Gawronski, 2013). Incongruent moral dilemmas correspond to traditional, high-conflict moral dilemmas commonly employed in research (e.g., Foot, 1967; Koenigs et al., 2007). In such dilemmas, participants read a scenario where a great deal of harm is impending, but participants could avoid this impending harm by accepting causing a lesser degree of harm. For example, in the crying baby dilemma, townspeople are hiding from murderous soldiers, but a baby is about to cry, which will summon the soldiers who will murder the townspeople. The actor could fatally smother the baby to prevent its cries, thereby saving the townspeople. Other dilemmas include torturing a person to prevent a bomb from killing several people, and causing severe harm to research animals to cure AIDS. Participants indicated whether causing each harm in order to achieve the specified outcome is appropriate or not appropriate (in line with Greene et al., 2001).
Congruent dilemmas are worded identically to incongruent dilemmas, except the outcome of causing harm has been minimized. For example, in the congruent version of the crying baby dilemma, the actor could kill the baby to prevent the townspeople from being forced to performing hard labor. In other dilemmas, the actor could torture a person to prevent a messy but non-lethal paint bomb, or cause severe harm to research animals to create a better facial cleanser. Again, participants indicated whether causing harm to achieve the specified outcome in acceptable or not acceptable. Participants responded to all dilemmas in the same fixed random order as Conway and Gawronski (2013). By considering participant responses to both incongruent and congruent dilemmas, it is possible to perform a PD analysis (Jacoby, 1991) that provides independent estimates of each participant’s inclination to reject causing harm (consistent with deontology), that appears to track affective reactions to harm, as well as inclinations to maximize outcomes (consistent with utilitarianism), that appears to track cognitive evaluations of outcomes (Conway and Gawronski, 2013).
We computed the deontological and utilitarian PD parameters using the six formulae described by Conway and Gawronski (2013). Consider Figure 1. The top path (U) illustrates the case where utilitarianism drives responses: this entails rejecting harm for congruent dilemmas but accepting harm for incongruent dilemmas (thus always maximizing outcomes). Representing the case where utilitarianism drives responses also allows representing the case where utilitarianism does not drive responses: (1-U). This case may be further subdivided into (1) the case where deontology drives responses (1-U ×D), which entails rejecting harmful actions in both congruent and incongruent dilemmas (thus always rejecting causing harm), as well as (2) the case where neither utilitarianism nor deontology drives responses (1-U × 1-D), which entails accepting harm for both congruent and incongruent dilemmas (suggesting at best an amoral insensitivity to the outcomes of one’s actions, or at worst general willingness to cause harm even when doing so makes the world worse overall).
Using the processing tree in Figure 1, researchers can algebraically represent each case: when participants accept or reject harm on congruent or incongruent dilemmas. For congruent dilemmas, participants may reject harm either when utilitarianism drives responses (U), or when utilitarianism does not, but deontology does (1-U ×D). Conversely, participants may accept harm only when neither utilitarianism nor deontology drives responses (1-U × 1-D). For incongruent dilemmas, participants reject harm when utilitarianism does not drive the response, but deontology does (1-U × D). Conversely, people may accept harm either when utilitarianism drives responses (U), or when neither utilitarianism nor deontology drives responses (1-U × 1-D). By combining these values, researchers can algebraically represent the cases when participants accept and reject harm on congruent or incongruent dilemmas. The probability of rejecting harm for congruent dilemmas is represented by the case where either utilitarianism drives the response, or when utilitarianism does not drive the response, but deontology does:
Conversely, the probability of accepting harm for congruent dilemmas is represented by the case that neither utilitarianism nor deontology drives the response:
For incongruent dilemmas, the probability of rejecting harm is represented by the case that deontology drives the response when utilitarianism does not:
Conversely, the probability of accepting harm for incongruent dilemmas is represented by the cases that utilitarianism drives the response, and neither deontology nor utilitarianism drives the response:
Next, researchers can enter the empirical distributions of participant’s acceptable and unacceptable responses to congruent and incongruent dilemmas into these equations, and then algebraically combine them to solve for the two parameters. By including Eq. 3 into Eq. 1, researchers can solve for U:
Finally, by including the value for U in Eq. 3, researchers can solve for D:
Together, these formulas enable researchers to independently estimate the degree to which participants systematically rejected causing harm (deontology parameter) and systematically maximize outcomes (utilitarian parameter).
Participants’ scores on the moral dilemma battery were process dissociated to produce the Deontological Parameter (M = 0.45, SD = 0.087), and the Utilitarian Parameter (M = -0.41, SD = 0.163). Neither scale violated the assumption of normality as verified via a Shapiro–Wilk’s test (Supplementary Figure S1).
The contrast Pain > Touch yielded clusters of activation consistent with prior studies on empathy for pain (Lamm et al., 2011). These clusters were located in visual cortex, inferior and superior parietal cortex, motor cortex, inferior premotor cortex, dorsolateral prefrontal cortex, anterior cingulate, thalamus, and others (see Figure 2 and Supplementary Table S1). The deontological PD parameter (reflecting harm-aversion dilemma judgments) correlated with BOLD signal changes in the left inferior frontal gyrus, pars opercularis, (see Figure 3 and Table 1). As predicted, there were no correlations between the utilitarian PD parameter (reflecting outcome-maximization judgments) and BOLD signal changes for the contrast Pain > Touch.
FIGURE 3. BOLD activation in the inferior premotor cortex for the contrast Pain > Touch predicts the deontological tendency in moral dilemmas. Heat map reflects Z-scores of normalized goodness-of-fit of normalized deontological process dissociation (PD) parameter to parameter estimates of BOLD activation for the contrast of interest. The descriptive scatterplot presents the relationship between the average BOLD parameter estimate in the premotor clusters and the normalized deontological PD parameter. The R-value reflects a Pearson’s correlation between the average beta-parameter estimate in the premotor clusters and the deontological parameter and is purely illustrative.
TABLE 1. Clusters of high goodness-of-fit when regressing the deontological parameter against parameter estimates for Pain > Touch across subjects.
This study is the first attempt to assess the relationship between the BOLD signal while people witness others in pain and those same peoples’ later response tendencies in hypothetical moral dilemmas. It belongs to a larger project on the relationships between type 1, pre-reflective responses to the internal states of other people, and type 2, reflective prosocial decisions that also includes two recent studies (Christov-Moore and Iacoboni, 2016; Christov-Moore et al., 2017). A broad array of brain regions responded to the pain video task, but of those, later harm-rejection (deontological) judgment tendencies were only predicted by activation in a sensorimotor area (IFG) associated with observing, imitating, and understanding the motor behavior of others (Iacoboni, 2009; Urgesi et al., 2014), their internal states (Carr et al., 2003; Lamm et al., 2011), and the broader construct of empathy (Carr et al., 2003; Lamm et al., 2011) and empathic accuracy (Paracampo et al., 2017). In contrast, areas associated with affective processing of empathy for pain, such as the anterior cingulate or anterior insula (Singer et al., 2004), did not predict deontological responses, despite those areas showing activation during the pain video task. As expected, no brain activation while witnessing pain correlated with outcome-maximization (utilitarian) judgment tendencies. These findings cannot reflect priming effects from the pain video task, as participants completed the dilemma battery at least 1 month following the scanner session. Together, results suggest that people who demonstrate greater sensorimotor empathy for observed pain in others are more likely to avoid causing harm (but not maximize outcomes) in hypothetical moral dilemmas.
These results, while correlational, support the hypothesis that deontological response tendencies primarily reflect genuine concern for others generated by witnessing pain, rather than self-focused emotional reactions to causing pain – after all, participants passively witnessed, rather than actively caused, the painful needle stimuli (Miller et al., 2014). These findings align with research linking deontological response tendencies to other-oriented processing, such as empathic concern and perspective-taking (e.g., Bartels, 2008; Conway and Gawronski, 2013; Gleichgerrcht and Young, 2013), thereby corroborating the dual-process model (Greene et al., 2001). However, these findings do not rule out the possibility that self-focused concerns regarding causing harm independently influence deontological response tendencies via mechanisms not captured in the current paradigm (Reynolds and Conway, in press).
Furthermore, these findings suggest the possible nature of the functional processes supporting other-oriented affective responses that lead to deontological choices. We employed stimuli known to elicit activation in circuits involved in both sensorimotor and general affective processing, and indeed found activation in both regions during the pain video task. However, only activation in the posterior inferior frontal cortex (pIFC) – associated with sensorimotor processing but not general affective processing – predicted subsequent deontological response tendencies. Past work shows that the pain video used here modulated transcranial magnetic stimulation (TMS)-induced motor evoked potentials (MEPs), an index of motor excitability, providing clear evidence for a sensorimotor component to empathy (Avenanti et al., 2005). TMS-induced MEP modulation while perceiving others’ hand actions, however, disappears when pIFC and neighboring areas are transiently disrupted by repetitive TMS (Avenanti et al., 2007). These findings suggest that activity in pIFC and neighboring areas is crucial for the sensorimotor component of empathy (Avenanti et al., 2005). Hence, we believe the correlation between deontological tendencies and pIFC activity in the current study is mainly driven by the pIFC’s role in the sensorimotor component of empathy. Furthermore, the role of the IFG in recognizing and understanding the behavior of others (Urgesi et al., 2014) as well as accurately assessing their internal states or empathic accuracy (Paracampo et al., 2017) is consistent with the notion that the pIFC integrates or provides links between high- and low-level empathic processes like neural resonance and moral decision-making.
We propose that the sensorimotor and affective mechanisms used to process one’s own pain also react to abstract and non-sensory harm to others, such as vivid images of pain – including mental images conjured when considering hypothetical moral dilemma scenarios. The more individuals recruit non-agent-specific pain representations in response to others’ suffering, the more likely they are to personally appreciate how others will experience harm. People who especially engage in such implicit representations appear especially likely to recoil from inducing pain on subsequent moral dilemma tasks (i.e., reject causing even harm that maximizes outcomes). Such findings are consistent with evidence that making harm vivid reduces willingness to directly inflict harm in moral dilemmas (e.g., Bartels, 2008), specifically through increased deontological inclinations (Conway and Gawronski, 2013), and that impairing visual processing reduces deontological judgments (Amit and Greene, 2012). However, the present findings are the first to suggest that the embodied consideration of other people’s pain at a sensorimotor level influence moral decision-making.
Obviously, these findings do not challenge the engagement of affective systems in moral dilemma judgment tendencies. Indeed, we should exercise caution extrapolating from the current limited sample. The lack of brain-behavior correlation with affective systems may reflect a false negative due to insufficient power, or else indicate that perhaps affective processing in moral dilemmas is relatively insensitive to individual differences in judgment tendencies. Furthermore, because no subjective evaluation was conducted regarding subject’s internal state while observing the needle stimuli, we cannot be sure that the task we used consistently tapped into concern for others, rather than self-focused emotional/sensorimotor experience. The relationship between these sensorimotor representations and dilemma responses may not be the one hypothesized here. However, the vicarious nature of this sensorimotor processing as well as past studies showing correlations between neural activity during similar tasks and other-oriented processes like empathy and pro-sociality mitigate this concern.
As predicted, these results failed to indicate an association between brain activation while witnessing harm and utilitarian response inclinations. This pattern meshes with previous findings suggesting the utilitarian responses primarily reflect abstract cognitive considerations of overall well-being, rather than a concrete focus on harm (e.g., Greene et al., 2008; Moore et al., 2008; Conway and Gawronski, 2013; cf. Kahane, 2015). Future research might profitably examine whether the utilitarian PD parameter correlates with brain activity when participants engage in theory of mind reasoning, mathematical calculations, or other abstract cognitive operations.
The current findings bolster the growing evidence that individual variability in sensorimotor and affective neural resonance predict differences in decision-making (Hein et al., 2011; Ma et al., 2011; Morelli et al., 2014; Christov-Moore and Iacoboni, 2016). Thus far, researchers have investigated this link only in the realm of prosocial decision-making (decisions involving potential gains); the current findings extend this relationship to moral dilemma judgments (decisions involving potential losses). Whereas some studies have implicated affective areas like the anterior insula in motivating prosocial decisions (Hein et al., 2011), others have implicated sensorimotor areas like the superior parietal lobe (Christov-Moore and Iacoboni, 2016) and SII/pIFC (Ma et al., 2011). One interpretation of this pattern suggests that general affective processing may be more important for motivating prosocial behavior rather than harm avoidance, whereas sensorimotor processing features for both. Alternatively, it remains possible that the current stimuli do not reveal the full extent to which general affective processing impacts harm-avoidance judgments. Consider that the videos we employed showed a close-up of hands getting pierced, without accompanying stimuli associated with witnessing others in pain (e.g., facial grimacing, cries of pain). Perhaps including such cues may provide increased evidence for the impact of general affective processing on dilemma decision-making. It remains likely that the relative contributions of affective or sensorimotor processes depend on multiple factors.
These findings are the first to demonstrate that, upon witnessing others in pain, brain activation in sensorimotor circuits involved in processing pain in the self predicts subsequent inclinations to avoid causing harm in moral dilemmas (even to maximize outcomes). Such findings clarify the importance of genuine concern for others in dilemma judgments, corroborate the dual process model, and provide further evidence for the utility of PD in dilemma decision-making. Moreover, these findings suggest a role for the embodied sensorimotor processing of harm when making moral dilemma decisions. People with stronger internal sensorimotor representations of others’ physical pain may be especially unwilling to visit harm on others, even for a good cause.
LC-M developed the study concept. All authors contributed to the study design. Testing and data collection were performed by LC-M and PC. LC-M and PC performed the data analysis and interpretation under the supervision of MI. LC-M and PC drafted the manuscript, and MI provided critical revisions. All authors approved the final version of the manuscript for submission.
The authors acknowledge funding from the German Research Foundation (DFG) via Thomas Mussweiler’s Gottfried Wilhelm Leibniz Award (MU 1500/5–1), the National Institute of Mental Health under Grant R21 MH097178 to MI, and a National Science Foundation Graduate Fellowship Grant (DGE-1144087) to LC-M.
Conflict of Interest Statement
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
The authors wish to thank the Brain Mapping Medical Research Organization, Brain Mapping Support Foundation, Pierson-Lovelace Foundation, The Ahmanson Foundation, William M. and Linda R. Dietel Philanthropic Fund at the Northern Piedmont Community Foundation, Tamkin Foundation, Jennifer Jones Simon Foundation, Capital Group Companies Charitable Foundation, Robson Family, and North Star Fund for their generous support.
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fnint.2017.00034/full#supplementary-material
FIGURE S1 | Distributions of deontological and utilitarian moral dilemma response tendencies. Solid line indicates ideal normal distribution. Mean and standard deviation are presented for both response tendency subscales.
TABLE S1 | MNI coordinates (in mm) of representative peaks of activation for the contrast Pain > Touch.
Amit, E., and Greene, J. D. (2012). You see, the ends don’t justify the means: visual imagery and moral judgment. Psychol. Sci. 23, 861–868. doi: 10.1177/0956797611434965
Avenanti, A., Bolognini, N., Maravita, A., and Aglioti, S. M. (2007). Somatic and motor components of action simulation. Curr. Biol. 17, 2129–2135. doi: 10.1016/j.cub.2007.11.045
Avenanti, A., Bueti, D., Galati, G., and Aglioti, S. M. (2005). Transcranial magnetic stimulation highlights the sensorimotor side of empathy for pain. Nat. Neurosci. 8, 955–960. doi: 10.1038/nn1481
Avenanti, A., Minio-Paluello, I., Bufalari, I., and Aglioti, S. M. (2009). The pain of a model in the personality of an onlooker: influence of state-reactivity and personality traits on embodied empathy for pain. Neuroimage 44, 275–283. doi: 10.1016/j.neuroimage.2008.08.001
Bartels, D. (2008). Principled moral sentiment and the flexibility of moral judgment and decision making. Cognition 108, 381–417. doi: 10.1016/j.cognition.2008.03.001
Bufalari, I., Aprile, T., Avenanti, A., Di Russo, F., and Aglioti, S. M. (2007). Empathy for pain and touch in the human somatosensory cortex. Cereb. Cortex 17, 2553–2561. doi: 10.1093/cercor/bhl161
Carr, L., Iacoboni, M., Dubeau, M., Mazziotta, J. C., and Lenzi, G. L. (2003). Neural mechanisms of empathy in humans: a relay from neural systems for imitation to limbic areas. Proc. Natl. Acad. Sci. U.S.A. 100, 5497–5502. doi: 10.1073/pnas.0935845100
Christov-Moore, L., and Iacoboni, M. (2016). Self-other resonance, its control and prosocial inclinations: brain-behavior relationships. Hum. Brain Mapp. 37, 1544–1558. doi: 10.1002/hbm.23119
Christov-Moore, L., Sugiyama, T., Grigaityte, K., and Iacoboni, M. (2017). Increasing generosity by disrupting prefrontal cortex increasing generosity by disrupting prefrontal cortex. Soc. Neurosci. 12, 174–181. doi: 10.1080/17470919.2016.1154105
Conway, P., and Gawronski, B. (2013). Deontological and utilitarian inclinations in moral decision-making: a process dissociation approach. J. Pers. Soc. Psychol. 104, 216–235. doi: 10.1037/a0031021
Davis, M. H. (1983). Measuring individual differences in empathy: evidence for a multidimensional approach. J. Pers. Soc. Psychol. 44, 113–126. doi: 10.1037/0022-35126.96.36.199
Epstein, S., Pacini, M., Denes-Raj, V., and Heier, H. (1996). Individual differences in intuitive-experiential and analytical-rational thinking styles. J. Pers. Soc. Psychol. 71, 390–405. doi: 10.1037/0022-35188.8.131.520
Fadiga, L., Fogassi, L., Pavesi, G., and Rizzolatti, G. (1995). Motor facilitation during action observation: a magnetic stimulation study. J. Neurophysiol. 73, 2608–2611.
Foot, P. (1967). The problem of abortion and the doctrine of double effect. Oxford Rev. 5, 5–15.
Friesdorf, R., Conway, P., and Gawronski, B. (2015). Gender differences in responses to moral dilemmas: a process dissociation analysis. Pers. Soc. Psychol. Rev. 42, 696–713. doi: 10.1177/0146167215575731
Gleichgerrcht, E., and Young, L. (2013). Low levels of empathic concern predict utilitarian moral judgment. PLOS ONE 8:e60418. doi: 10.1371/journal.pone.0060418
Greene, J. D., Morelli, S. A., Lowenberg, K., Nystrom, L. E., and Cohen, J. D. (2008). Cognitive load selectively interferes with utilitarian moral judgment. Cognition 107, 1144–1154. doi: 10.1016/j.cognition.2007.11.004
Greene, J. D., Nystrom, L. E., Engell, A. D., Darley, J. M., and Cohen, J. D. (2004). The neural bases of cognitive conflict and control in moral judgment. Neuron 44, 389–400. doi: 10.1016/j.neuron.2004.09.027
Greene, J. D., Sommerville, R. B., Nystrom, L. E., Darley, J. M., and Cohen, J. D. (2001). An fMRI investigation of emotional engagement in moral judgment. Science 293, 2105–2108. doi: 10.1126/science.1062872
Hein, G., Lamm, C., Brodbeck, C., and Singer, T. (2011). Skin conductance response to the pain of others predicts later costly helping. PLOS ONE 6:e22759. doi: 10.1371/journal.pone.0022759
Iacoboni, M. (2009). Imitation, empathy, and mirror neurons. Annu. Rev. Psychol. 60, 653–670. doi: 10.1146/annurev.psych.60.110707.163604
Jabbi, M., Swart, M., and Keysers, C. (2007). Empathy for positive and negative emotions in the gustatory cortex, Neuroimage 34, 1744–1753. doi: 10.1016/j.neuroimage.2006.10.032
Jacoby, L. L. (1991). A process dissociation framework: separating automatic from intentional uses of memory. J. Mem. Lang. 30, 513–541. doi: 10.1016/0749-596X(91)90025-F
Kahane, G. (2015). Sidetracked by trolleys: Why sacrificial moral dilemmas tell us little (or nothing) about utilitarian judgment. Soc. Neurosci. 10, 551–560. doi: 10.1080/17470919.2015.1023400
Kant, I. (1785/1959). Foundation of the Metaphysics of Morals, trans. L. W. Beck. Indianapolis, IN: Bobbs-Merrill.
Kelly, R. E., Alexopoulos, G. S., Wang, Z., Gunning, F. M., Murphy, C. F., Morimoto, S. S., et al. (2010). Visual inspection of independent components: defining a procedure for artifact removal from fMRI data. J. Neurosci. Methods 189, 233–245. doi: 10.1016/j.jneumeth.2010.03.028
Koenigs, M., Young, L., Adolphs, R., Tranel, D., Cushman, F., Hauser, M., et al. (2007). Damage to the prefrontal cortex increases utilitarian moral judgments. Nature 446, 908–911. doi: 10.1038/nature05631
Lamm, C., Decety, J., and Singer, T. (2011). Meta-analytic evidence for common and distinct neural networks associated with directly experienced pain and empathy for pain. Neuroimage 54, 2492–2502. doi: 10.1016/j.neuroimage.2010.10.014
Ma, Y., Wang, C., and Han, S. (2011). Neural responses to perceived pain in others predict real-life monetary donations in different socioeconomic contexts. Neuroimage 57, 1273–1280. doi: 10.1016/j.neuroimage.2011.05.003
Mill, J. S. (1861/1998). Utilitarianism, ed. R. Crisp. New York, NY: Oxford University Press.
Miller, R. M., Hannikainen, I. A., and Cushman, F. A. (2014). Bad actions or bad outcomes? Differentiating affective contributions to the moral condemnation of harm. Emotion 14, 573–587. doi: 10.1037/a0035361
Moore, A. B., Clark, B. A., and Kane, M. J. (2008). Who shalt not kill? Individual differences in working memory capacity, executive control, and moral judgment. Psychol. Sci. 19, 549–557. doi: 10.1111/j.1467-9280.2008.02122.x
Morelli, S. A., Rameson, L. T., and Lieberman, M. D. (2014). The neural components of empathy: predicting daily prosocial behavior. Soc. Cogn. Affect. Neurosci. 9, 39–47. doi: 10.1093/scan/nss088
Paracampo, R., Tidoni, E., Borgomaneri, S., Pellegrino, G., and Avenanti, A. (2017). Sensorimotor network crucial for inferring amusement from smiles. Cereb. Cortex 11, 5116–5129.
Payne, B. K., and Bishara, A. J. (2009). An integrative review of process dissociation and related models in social cognition. Eur. Rev. Soc. Psychol. 20, 272–314. doi: 10.1080/10463280903162177
Pfeifer, J. H., Iacoboni, M., Mazziotta, J. C., and Dapretto, M. (2008). Mirroring others’ emotions relates to empathy and interpersonal competence in children. Neuroimage 39, 2076–2085. doi: 10.1016/j.neuroimage.2007.10.032
Reynolds, C., and Conway, P. (in press). Affective concern for bad outcomes does contribute to the moral condemnation of harm after all. Emotion.
Singer, T., Seymour, B., O’Doherty, J., Kaube, H., Dolan, R. J., and Frith, C. D. (2004). Empathy for pain involves the affective but not sensory components of pain. Science 303, 1157–1162. doi: 10.1126/science.1093535
Urgesi, C., Candidi, M., and Avenanti, A. (2014). Neuroanatomical substrates of action perception and understanding: an anatomic likelihood estimation meta-analysis of lesion-symptom mapping studies in brain injured patients. Front. Hum. Neurosci. 8:344. doi: 10.3389/fnhum.2014.00344
Wicker, B., Keysers, C., Plailly, J., Royet, J. P., Gallese, V., and Rizzolatti, G. (2003). Both of us disgusted in my insula: the common neural basis of seeing and feeling disgust. Neuron 40, 655–664. doi: 10.1016/S0896-6273(03)00679-2
Zaki, J., and Ochsner, K. (2012). The neuroscience of empathy: progress, pitfalls and promise. Nat. Neurosci. 15, 675–680. doi: 10.1038/nn.3085
Keywords: embodiment, empathy, fMRI, moral dilemmas, moral judgment, process dissociation, neural resonance
Citation: Christov-Moore L, Conway P and Iacoboni M (2017) Deontological Dilemma Response Tendencies and Sensorimotor Representations of Harm to Others. Front. Integr. Neurosci. 11:34. doi: 10.3389/fnint.2017.00034
Received: 28 August 2017; Accepted: 28 November 2017;
Published: 12 December 2017.
Edited by:Christophe Lopez, Centre National de la Recherche Scientifique (CNRS), France
Reviewed by:Alessio Avenanti, Università di Bologna, Italy
Andrei C. Miu, Babeş-Bolyai University, Romania
Copyright © 2017 Christov-Moore, Conway and Iacoboni. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Marco Iacoboni, firstname.lastname@example.org