Impact Factor 2.323

The 1st most cited journal in Multidisciplinary Psychology

General Commentary ARTICLE

Front. Psychol., 13 October 2017 | https://doi.org/10.3389/fpsyg.2017.01789

Commentary: Can Ordinary People Detect Deception after All?

  • 1Department of Psychology, University of Huddersfield, Huddersfield, United Kingdom
  • 2International Research Centre for Investigative Psychology, University of Huddersfield, Huddersfield, United Kingdom
  • 3Departamento de Psicología Básica, Universidad Autónoma de Madrid, Madrid, Spain

A commentary on
Can Ordinary People Detect Deception after All?

by ten Brinke, L., Vohs, K. D., and Carney, D. R. (2016). Trends Cogn. Sci. 20, 579–588. doi: 10.1016/j.tics.2016.05.012

No one likes to call someone a liar. But the authors of the tipping point account (ten Brinke et al., 2016) claim that it is evolutionary prudent to spot lies that can harm us in order to determine who to trust. As such, they propose the reputational costs of confronting a liar might be overcome by detecting lies unconsciously. When confronted with information that creates a threat response, the unconscious can use the threat response to detect deceptive cues and to unconsciously infer deception, all the while keeping this information out of the conscious mind. The account suggests this is beneficial because conscious awareness of the deception “could impel the perceiver to confront the liar” (p. 580).

The account is controversial insofar as it claims that people can detect deception, in contrast to past work showing otherwise (47% detection rate of lies, and 61% of truths, resulting from bias to judge statements as true: Bond and DePaulo, 2006), and also makes novel claims about an unconscious ability. Although it is welcoming to see new theoretical approaches to lie detection, the account (a) makes claims that do not match the data and conclusions presented in the studies cited to build its case, (b) offers no testable definition of unconscious processes, and (c) contains internal contradictions.

The first issue is with the mismatch between what the studies found and what the tipping point authors are interpreting from them. For instance, ten Brinke et al. (2016) interpret work on nonhuman animals (primates and canines) as showing that nonconscious thinking can detect deception (Wheeler, 2010; Takaoka et al., 2015), and predict that, for example, “Canine behavior will reveal a preference for approaching truth-tellers and avoiding liars” (p. 582). However, the cited works do not explore deception or lie detection. Takaoka et al. (2015) trained dogs to go to a container that concealed food, identified by a person pointing at the correct container. After training, the dogs are shown which container is baited, and then a person points to the wrong container. The dogs correctly choose the baited container. Is the canine drawing on knowledge of deception, or is this evidence that dogs do not use unreliable information when they have more robust information available (i.e., having seen which container is baited)? We, and the original authors, would argue for the latter. Similarly, ten Brinke et al. cite Wheeler (2010) in support of the claim that “[n]onhuman primates can detect deception at higher rates than humans” (p. 582). But this study does not test deception or lie detection, let alone compare human and primate performance.

In the same vein, neuropsychological work is cited (Grèzes et al., 2004; specifically, Grèzes et al., 2006; Lissek et al., 2008) to argue that brain- or body-based physiological responses occur when observing deception. This may seem to suggest that people are unconsciously responding to deception. But in these studies participants were explicitly made aware of the possibility of deception and were asked to make lie-truth judgments, sometimes reaching 100% accuracy. It is not clear how one would show that the physiological activity is not indicative of the conscious judgment they were asked to make.

The largest body of evidence supporting unconscious lie detection stems from the indirect method. Participants are not consciously informed about the possibility of deception. Instead, they judge whether the speaker, for instance, appears to be thinking hard. These studies find that judgments of thinking hard (or some other indirect judgment of deception) distinguishes liars and truth-tellers more accurately than an explicit lie-truth judgment. ten Brinke et al. cite work showing that people feel less comfortable and more suspicious (two indirect judgments) when viewing their friends' deceptions compared to viewing their truths, but were at chance accuracy in making an explicit lie-truth judgment (Anderson et al., 2002). It would appear that the rater cannot explicitly distinguish lies from truths, but feels uncomfortable when listening to lies, which might suggest some form of unconscious knowledge. However, Anderson and colleagues demonstrated that their result was a methodological effect attributable to the fact that the scale used to collect explicit ratings was less sensitive than the one used for indirect ratings, an effect which has been found in a meta-analysis (Bond and DePaulo, 2006). In fact, indirect lie detection often performs worse than direct lie detection (Levine and Bond, 2014; Bond et al., 2015), and can be explained by entirely conscious processes (Street and Richardson, 2015; Street and Vadillo, 2016).

To the best of our knowledge, only two of the studies cited by ten Brinke et al. (Reinhard et al., 2013; ten Brinke et al., 2014) contain unambiguous evidence in favor of unconscious lie detection. But the reliability of these two findings has been called into question by failures to replicate the former (Moi and Shanks, 2015) and several oddities in the analysis of the latter (Levine and Bond, 2014; Franz and von Luxburg, 2015).

The second issue with the tipping point account is its falsifiability. The authors offer two and a half pages of predictions, but unfortunately, none of them test whether the effect is unconscious. For instance, it is predicted that, “[e]xperiencing social exclusion will enhance accuracy” (p. 583). If this prediction was supported, we cannot know whether it arises from unconscious thinking. The authors do not explain what the unconscious is or how it is possible to test whether the unconscious is involved. There is an active and ongoing debate around whether the unconscious exists (e.g., Newell and Shanks, 2014). Because of the lack of a definition of what the unconscious is, how it could be measured, or how it should work, the tipping point theory's claim to the unconscious is unfalsifiable.

The third issue with the account is that there are a number of inconsistencies. For example, it is predicted that increasing reputational and relationship costs of accusing others of deception should detriment accuracy: “When social norms shift and license people to catch liars, thus attenuating the social costs of declaring someone a liar, accuracy improves” (p. 586). But the account also attempts to harness findings showing that when the costs to the relationship are perceived to be particularly high, accuracy actually improves, citing Ein-Dor and Perry (2013). It seems difficult to reconcile these two contradictory positions.

While a threat to the self may engage the unconscious to help detect the lie, an overwhelming threat may lead people to be suspicious and judge whatever they hear to be a lie, even at the expense of accuracy. What is an overly potent threat? The authors cite work showing that police officers are biased to judge “lie” when rating footage of students committing mock theft and vandalism (Meissner and Kassin, 2002). If this is sufficiently threatening to overwhelm any accuracy effects, the level of threat that the unconscious has evolved to detect seems particularly benign. Yet deception gets people to offer up their financial details (Wright et al., 2010) and being person-trafficked (United Nations Office on Drugs and Crime, 2004; Hübschle, 2014). These are potent threats that should create a lie bias, but people seem to believe the persuader.

The tipping point account acknowledges that the threat response only allows higher accuracy “when cues to deception are present and perceptible” (ten Brinke et al., 2016, p. 580). That threat may make people judge statements as lies (a “lie bias”) is consistent with current theories that do not rely on unconscious processing. The effect of threats creating a lie bias is consistent with the context-general information use of ALIED theory (Street, 2015) and with the concept of triggers in truth-default theory (Levine, 2014), neither of which require a claim to the unconscious. The suggestion that the social repercussions of accusing others may cause a truth bias has been made by O'Sullivan et al. (1988) and O'Sullivan (2003) without claiming there is an unconscious element. While the accusatory reluctance position has been suggested in the literature, it has received little direct empirical testing. A useful contribution of the tipping point theory, then, is to make explicit a number of predictions that could test for the presence of accusatory reluctance (see Box 1). But this exploration can take place without reference to an undefined hidden process (Street and Vadillo, 2016).

Box 1. Developing the tipping point account.

If the account aims to make an unconscious claim, it would benefit from (a) defining what “unconscious” means and how it is supposed to increase accuracy, (b) outlining predictions that test its unconscious claims, and (c) couching the discussion of the unconscious in the unconscious cognition literature. However, given the lack of support from the cited research, the internal inconsistencies that may in part be remedied by removing the claim to the unconscious, and the lack of a testable definition, the account would likely benefit from making no claims to the unconscious.

Given that the predictions may just as easily be accommodated by conscious processes, why does the account argue for an unconscious process? The authors suggest that “if cues to deception enter into consciousness, they could impel the perceiver to confront the liar.” (p. 580). But, consciousness is not impelled to communicate. People are capable of holding conscious thoughts without making them public. In fact, one might even call this a definition of deception: To be aware that what one is saying does not match with what one believes to be true.

Author Contributions

CS and MV contributed approximately equally to the conception, research, and writing of this commentary.

Funding

MV was supported by Grant 2016-T1/SOC-1395 from madri+d Science Foundation.

Conflict of Interest Statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgments

We would like to thank Maria Ioannou for suggesting sources of literature we were previously unaware of. We would also like to thank the two reviewers of this manuscript for their time and suggestions.

References

Anderson, D. E., DePaulo, B. M., and Ansfield, M. E. (2002). The development of deception detection skill: a longitudinal study of same-sex friends. Person. Soc. Psychol. Bull. 28, 536–545. doi: 10.1177/0146167202287010

CrossRef Full Text | Google Scholar

Bond, C. F., and DePaulo, B. M. (2006). Accuracy of deception judgments. Person. Soc. Psychol. Rev. 10, 214–234. doi: 10.1207/s15327957pspr1003_2

PubMed Abstract | CrossRef Full Text | Google Scholar

Bond, C. F., Levine, T. R., and Hartwig, M. (2015). “New findings in nonverbal lie detection,” in Detecting Deception: Current Challenges and Cognitive Approaches, eds P. A. Granhag, A. Vrij, and B. Verschuere (Chichester: Wiley), 37–58.

Ein-Dor, T., and Perry, A. (2013). Full house of fears: evidence that people high in attachment anxiety are more accurate in detecting deceit. J. Pers. 82, 83–92. doi: 10.1111/jopy.12035

PubMed Abstract | CrossRef Full Text | Google Scholar

Franz, V. H., and von Luxburg, U. (2015). No evidence for unconscious lie detection: a significant difference does not imply accurate classification. Psychol. Sci. 26, 1646–1648. doi: 10.1177/0956797615597333

PubMed Abstract | CrossRef Full Text | Google Scholar

Grèzes, J., Berthoz, S., and Passingham, R. E. (2006). Amygdala activation when one is the target of deceit: did he lie to you or to someone else? NeuroImage 30, 601–608. doi: 10.1016/j.neuroimage.2005.09.038

PubMed Abstract | CrossRef Full Text | Google Scholar

Grèzes, J., Frith, C., and Passingham, R. E. (2004). Brain mechanisms for inferring deceit in the actions of others. J. Neurosci. 24, 5500–5505. doi: 10.1523/JNEUROSCI.0219-04.2004

PubMed Abstract | CrossRef Full Text | Google Scholar

Hübschle, A. (2014). Of bogus hunters, queenpins and mules: the varied roles of women in transnational organized crime in Southern Africa. Trends Organ. Crime 17, 31–51. doi: 10.1007/s12117-013-9202-8

CrossRef Full Text

Levine, T., and Bond, C. F. (2014). Direct and indirect measures of lie detection tell the same story: a reply to ten Brinke, Stimson, and Carney (2014). Psychol. Sci. 25, 1960–1961. doi: 10.1177/0956797614536740

PubMed Abstract | CrossRef Full Text | Google Scholar

Levine, T. R. (2014). Truth-default theory (TDT): a theory of human deception and deception detection. J. Lang. Soc. Psychol. 33, 378–392. doi: 10.1177/0261927X14535916

CrossRef Full Text | Google Scholar

Lissek, S., Peters, S., Fuchs, N., Witthaus, H., Nicolas, V., Tegenthoff, M., et al. (2008). Cooperation and deception recruit different subsets of the theory-of-mind network. PLoS ONE 3:e2023. doi: 10.1371/journal.pone.0002023

PubMed Abstract | CrossRef Full Text | Google Scholar

Meissner, C. A., and Kassin, S. M. (2002). “He's guilty!”: investigator bias in judgments of truth and deception. Law Hum. Behav. 26, 469–480. doi: 10.1023/A:1020278620751

PubMed Abstract | CrossRef Full Text | Google Scholar

Moi, W. Y., and Shanks, D. R. (2015). Can lies be detected unconsciously? Front. Psychol. 6:1221. doi: 10.3389/fpsyg.2015.01221

PubMed Abstract | CrossRef Full Text | Google Scholar

Newell, B. R., and Shanks, D. R. (2014). Unconscious influences on decision making: a critical review. Behav. Brain Sci. 37, 1–19. doi: 10.1017/S0140525X12003214

PubMed Abstract | CrossRef Full Text | Google Scholar

O'Sullivan, M. (2003). The fundamental attribution error in detecting deception: the boy-who-cried-wolf effect. Person. Soc. Psychol. Bull. 29, 1316–1327. doi: 10.1177/0146167203254610

PubMed Abstract | CrossRef Full Text | Google Scholar

O'Sullivan, M., Ekman, P., and Friesen, W. V. (1988). The effect of comparisons on detecting deceit. J. Nonverbal Behav. 12, 203–215. doi: 10.1007/BF00987488

CrossRef Full Text | Google Scholar

Reinhard, M. A., Greifeneder, R., and Scharmach, M. (2013). Unconscious processes improve lie detection. J. Pers. Soc. Psychol. 105, 721–739. doi: 10.1037/a0034352

PubMed Abstract | CrossRef Full Text | Google Scholar

Street, C. N. H. (2015). ALIED: humans as adaptive lie detectors. J. Appl. Res. Mem. Cogn. 4, 335–343. doi: 10.1016/j.jarmac.2015.06.002

CrossRef Full Text | Google Scholar

Street, C. N. H., and Richardson, D. C. (2015). The focal account: indirect lie detection need not access unconscious, implicit knowledge. J. Exp. Psychol. Appl. 21, 342–344. doi: 10.1037/xap0000058

PubMed Abstract | CrossRef Full Text | Google Scholar

Street, C. N. H., and Vadillo, M. A. (2016). Can the unconscious boost lie-detection accuracy? Curr. Dir. Psychol. Sci. 25, 246–250. doi: 10.1177/0963721416656348

CrossRef Full Text | Google Scholar

Takaoka, A., Maeda, T., Hori, Y., and Fujita, K. (2015). Do dogs follow behavioral cues from an unreliable human? Anim. Cogn. 18, 475–483. doi: 10.1007/s10071-014-0816-2

PubMed Abstract | CrossRef Full Text | Google Scholar

ten Brinke, L., Stimson, D., and Carney, D. R. (2014). Some evidence for unconscious lie detection. Psychol. Sci. 25, 1098–1105. doi: 10.1177/0956797614524421

PubMed Abstract | CrossRef Full Text | Google Scholar

ten Brinke, L., Vohs, K. D., and Carney, D. R. (2016). Can ordinary people detect deception after all? Trends Cogn. Sci. 20, 579–588. doi: 10.1016/j.tics.2016.05.012

PubMed Abstract | CrossRef Full Text | Google Scholar

United Nations Office on Drugs and Crime (2004). United Nations Convention against Transnational Organized Crime and the Protocols Thereto. New York, NY: United Nations.

Wheeler, B. C. (2010). Production and perception of situationally variable alarm calls in wild tufted capuchin monkeys (Cebus paella nigritus). Behav. Ecol. Sociobiol. 64, 989–1000. doi: 10.1007/s00265-010-0914-3

CrossRef Full Text | Google Scholar

Wright, R., Chakraborty, S., Basoglu, A., and Marett, K. (2010). Where did they go right? Understanding the deception in phishing communications. Group Dec. Negot. 19, 391–416. doi: 10.1007/s10726-009-9167-9

CrossRef Full Text | Google Scholar

Keywords: lie detection, unconscious, tipping point, indirect lie detection, unconscious lie detection, unconscious thought theory, deception detection

Citation: Street CNH and Vadillo MA (2017) Commentary: Can Ordinary People Detect Deception After All? Front. Psychol. 8:1789. doi: 10.3389/fpsyg.2017.01789

Received: 09 July 2017; Accepted: 26 September 2017;
Published: 13 October 2017.

Edited by:

Matthias Gamer, University of Würzburg, Germany

Reviewed by:

Bruno Verschuere, University of Amsterdam, Netherlands
Timothy R. Levine, University of Alabama School of Medicine, United States

Copyright © 2017 Street and Vadillo. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Chris N. H. Street, c.street@hud.ac.uk