- 1Department of Psychology and Sports Science, University of Giessen, Giessen, Germany
- 2Department of Psychology, Nicolaus Copernicus University, Toruń, Poland
Introduction
In “direct” lie detection, the receiver of a message is explicitly asked to judge its veracity. In “indirect” or “implicit1 ” lie detection, the receiver is asked to rate some global impression(s) of the sender (e.g., “appears friendly,” “thinks hard”), sometimes with, sometimes without knowing that the study is about deception. “Unconscious” vs. “conscious” lie detection refers to the cognitive processes assumed to be involved in detecting deception. Some studies have shown that indirect or unconscious methods may lead to more accurate veracity assessment than a direct or conscious approach. However, results appear inconsistent, and it is unclear which mechanisms could be responsible for this advantage (Granhag, 2006). Some researchers suggest the role of unconscious knowledge or intuition (Reinhard et al., 2013a; ten Brinke et al., 20142). Others propose that indirect questions do not activate stereotypical beliefs about deception cues (Vrij et al., 2001; Street and Richardson, 2015).
Thus, although some of the veracity assessment methods labeled today as “indirect” or “unconscious” have been utilized for at least 40 years (DePaulo and Morris, 2004), most questions concerning their theoretical background, their accuracy, or practical applications (Granhag, 2006) remain unsolved. Even the simple decision to classify a procedure as “indirect” or “unconscious” still generates debate. This problem alone may be one of the reasons why conclusions from different studies on “indirect” or “unconscious” lie detection approaches appear so contradictory, leading Levine (2019) to question the value of these methods categorically.
It seems that different approaches have pursued their work from specific vantage points, ignoring each other like ships passing each other in the night. Although the five reasons Levine (2019) postulates are essential and well worth considering, his conclusions may be based on limited evidence [see Bond et al. (2015)]. Street and Vadillo (2016) have also criticized these approaches claiming the involvement of unconscious processes in lie detection as a specific instance of the general replication crisis observed with other social psychological phenomena attributed to unconscious processing. We concur with their theoretical arguments in so far as we also like to see evidence that indeed unconscious processes, in the absence of conscious processing, are at work in “unconscious lie detection.” However, we disagree with their conclusions based on Bond et al. (2015) meta-analysis (see below). Thus, we propose a broader view on the vast realm of studies and suggest that before researchers draw firm conclusions about these methods' potential, they should consider related theoretical approaches and their findings. To demonstrate the diversity of the procedures of studies using indirect or unconscious lie detection, we provide a detailed summary of the methodological aspects of these studies in Table 1. Furthermore, we draw attention to boundary conditions and moderator variables that should allow more differentiated conclusions about indirect and unconscious lie detection if considered in future research.
Table 1. Review of experiments on direct vs. indirect lie detection and unconscious vs. conscious lie detection.
A Comparison of Paradigms
First, it is important to consider whether indirect and unconscious veracity assessment methods3 activate similar processes and may be treated as examples of the same lie detection paradigm. We think they are not. As summarized in Section I of Table 1, in studies on indirect lie detection, an observer is asked to assess some aspects of a sender's statement or behavior other than veracity (e.g., whether a sender is “thinking hard4”) or is asked to evaluate their cognitions or emotions evoked by a sender (e.g., whether the receiver felt comfortable watching a sender; Granhag, 2006). Although some argue that these indirect questions lead to more accurate intuitive judgments, others suggest that this method's postulated higher accuracy relies on redirecting the focus of a detector's attention away from veracity per se but instead on a small set of cues (or even on a single cue) considered more diagnostic (Vrij et al., 2001; Street and Richardson, 2015).
In studies on unconscious lie detection (Section II of Table 1), the controlled, conscious processing of the sender's statement or behavior is disrupted or impeded, usually by an additional task. Some of these studies applied methodology taken from the unconscious-thought theory (UTT; Dijksterhuis and Nordgren, 2006). The UTT assumes two modes of thought: conscious and unconscious, which have different characteristics and should be used under different circumstances (Dijksterhuis and Nordgren, 2006). According to the UTT, people often use these modes inappropriately; for example, they try to solve complex problems (here, assess the veracity) by engaging in conscious thinking rather than unconscious processing. Unconscious thought, which automatically processes available data while one is consciously engaged in a different task, allows for more processing capacity to deal with complex information (i.e., behavioral cues to deception) and prevents stereotyping (i.e., relying on inaccurate stereotypical deception cues). As a result, unconscious (automatic) lie detectors are more accurate than conscious (controlled) thinkers because they can analyze and consider a higher number of diagnostic cues (Reinhard et al., 2013a).
Although indirect lie detection and UTT-based veracity assessment are thought to result in decisions based on more diagnostic behavioral cues, both approaches seem to achieve it by engaging different psychological processes (see Granhag, 2006). However, what we miss in almost all UTT-based studies is that there is no explicit, independent evidence for “unconscious processes” and their relationships to the judgments made5.
In some studies on unconscious decision-making, participants are told beforehand that deception is an issue; thus, they are aware of their task (Table 1, column “Aware”). In others, they learn about it only after seeing a videotape. In the latter procedure, raters, who are not aware that the study is concerned with detecting deception, may find the task not very motivating. In contrast, in many studies on indirect deception detection, participants usually watch videotapes and rate their global impression of one or several of the senders' behaviors. In between-participants designs, another group of raters assesses the veracity of each account. In within-participants designs, raters judge the veracity either immediately before or after rating each stimulus, or, alternatively, before or after a block of stimuli.
Although within-participant designs (see Table 1, column “Design”) may have statistical advantages, we note several methodological issues: (1) If indirect and direct measures are used together, they may mutually influence each other. A possible solution would be a pretest–posttest with a control group mixed-model design as in studies on training to detect deception (see Hauch et al., 2016), or even a Solomon four-group design (Campbell and Stanley, 1963). (2) If there are many (similar or long) videos shown en bloc, judges may forget who said what. Therefore, some studies have used only a few short (or partial) video clips. (3) If the speaker is reintroduced via a photograph, veracity judgments may be affected by the sender's perceived attractiveness or likeableness (Zebrowitz et al., 1996). (4) Using short video segments (“thin slices”), perhaps without tone, favors automatic, rapid judgments over conscious, deliberate analysis of the content of the message. (5) Using only a few stimuli violates the principle of stimulus sampling (see below).
Focus on the Decision-Making Process vs. the Outcome
We do not deny that research has shown the superiority of unconscious over conscious or deliberate decisions in various fields other than lie detection. However, as demonstrated by the disproportion of studies in Section I and Section II of Table 1, the evidence for this claim in the lie detection field is rather scarce and equivocal and may largely depend on the methodology used. Concerning any type of decision-making, people seem to have little insight into the reasons why they make certain decisions (Nisbett and Wilson, 1977). Nonetheless, “fast and frugal” decisions are claimed to be often accurate (Gigerenzer et al., 1999). In the eyewitness literature, there is also evidence that fast decisions are a reliable cue for their accuracy and that having witnesses “think aloud” (Ericsson and Simon, 1993) while arriving at their decisions improve observers' evaluations of the accuracy of these decisions (Kaminski and Sporer, 2017). We suggest that think-aloud protocols and Brunswikian lens model analyses6 could also help us understand the process of indirect or unconscious deception detection (see Reinhard et al., 2013a).
Methodological Issues and Recommendations7
Specific Observable Behavior Ratings vs. Global Indirect Cues
Our summary in Table 1 shows that indirect questions concern a broad variety of sender behaviors, ranging from more specific to global nonverbal and paraverbal cues (e.g., hand/finger movements, response latencies, nervous), personality attributes (e.g., active, eloquent), emotions (e.g., cheerful, angry), a variety of verbal content cues (e.g., detailed), as well as inferred processes (e.g., thinking hard) that may influence both verbal and nonverbal behaviors. Verbal content cue ratings were also sometimes more general (e.g., consistency, plausibility) or referred to specified content qualities as used in research on criteria-based content analysis (e.g., reproduction of conversations) and within the reality monitoring approach (e.g., temporal details; see Sporer, 1997, 2004).
Although more thorough comparisons of the efficacy of different types of indirect questions are needed (see Ulatowska, 2014, 2020), we argue that indirect questions regarding verbal content cues should lead to higher accuracies because this approach enables the rater to find more diagnostic cues (Street and Richardson, 2015). We present evidence on the possible superiority of verbal content cues below.
Moreover, we urge researchers not only to ask for assessments of global impressions like “thinking hard” but also to collect additional data on specific, observable behaviors from independent raters and correlate these two sets of data. This would add more objective, operational definitions of indirect questions to the global impression ratings.
Importantly, significant differences in indirect measures do not imply that judges would actually use these differences to classify accounts correctly. An additional experimental group is necessary to investigate this, in which raters who rated specific behaviors or global impressions will subsequently assess each statement's veracity based on these ratings. This methodology is usually used in studies on training to detect deception (Hauch et al., 2016).
When comparisons between lies and truths on any dependent measures are analyzed, effect sizes (Cohen's d) with 95% CIs for all behaviors observed should be reported (not only for significant differences). Before conducting a study, power analyses should be performed, and non-significant differences should be accompanied by Bayes factor analyses (which are beyond this paper's scope).
Defining Accuracy
Usually, recipients are asked to either make binary judgments (0 vs. 1) or assess veracity on continuous credibility rating scales (see Table 1, last two columns). To make studies comparable, dependent variables need to be defined and labeled consistently. When rating scales are used (for veracity judgments or other measures), results should be referred to as a difference between lies and truths rather than lie/truth detection accuracy. When binary judgments were made, sometimes researchers wrote about “lie detection,” which makes it unclear whether accurate classifications of true statements are included in this term. To avoid such ambiguities, accuracy of classifications of lies (lie detection accuracy) and of truths (truth detection accuracy) and overall accuracy (averaged across both) should be reported.
In everyday life, we expect most statements to be truthful. When the number of lies and truths in a study is not equal (as is likely to be the case in the real world), the proportions should be specified because accuracy will depend on expected and actual base rates of lies in a set of stimuli. Moreover, reporting both lie and truth accuracy is particularly important in areas where there are likely to be very high or very low base rates of lies or truths. Consequently, a method with high false alarm rates (i.e., classifying a truth-teller as a liar) would lead to large numbers of false accusations. In certain situations (e.g., airport security settings), lying base rates may be very low, leading to high false alarm rates. Hence, these methods may only be used as screening devices to be followed up by more thorough investigations to obtain further evidence.
Both lie accuracy and truth accuracy can be combined via signal detection theory, yielding measures of performance (d', A') as well as a (presumably independent) measure of response bias (C, beta, beta”). As recently proposed by Smith and Leach (2019), signal detection models also allow us to use confidence in a lie–truth detection judgment to achieve even better discrimination8. When accuracies are reported for different confidence levels, the proportions of cases in each confidence bin should also be noted.
For all differences between methods groups, effect sizes with the direction of effect (Cohen's d, or odds ratios, not partial eta2) should be reported (see Sporer et al., 2021). Accuracies calculated from automatic classification methods (e.g., with artificial intelligence algorithms) and from multivariate classification methods (e.g., multiple discriminant or logistic regression analyses, in particular when they were obtained without cross-validation) must not be compared with the accuracies of human raters (Sporer et al., 2021).
Sample Sizes of Judges vs. Sample Size of Senders
Many studies have used relatively large samples of judges but only small samples of stimuli to be judged (often only 8 or 10; Table 1, Columns: N Ps and N Stimuli)9. Note that with 8 or 10 accounts, an individual judge will only be accurate above chance level if she or he classifies 7 or 8 of 8 (or 9 or 10 of 10) accounts correctly according to the binomial distribution. While some accuracy differences may become significant at the aggregate level with large numbers of judges, this does not imply that an individual judge will achieve accuracy above chance. Considering that there is much more variability across studies to be explained in sender detectability than in judge ability (Bond and DePaulo, 2008a,b), studies ought to use large, representative samples of stimuli. Therefore, we expect that differences in detection accuracy will be more likely detected if participants rate large numbers of stimuli (e.g., in several sessions) and different types of stimuli (e.g., different lie scenarios). Furthermore, if multiple rating scales are used, sample sizes of stimuli should be large enough to allow for factor analyses and the construction of theoretically meaningful subscales (reporting Cronbach's alphas and corrected item-total correlations for each subscale; Sporer et al., 2021).
Using small samples of stimuli violates the principle of stimulus sampling (Wells and Windschitl, 1999): Findings should hold not only across large numbers of participants but also for large numbers of stimuli from different situations and contexts. Applied to detecting deception, outcomes may vary for different types of lies. Studies on indirect methods have sometimes used lies about emotions (toward a person or attitude object), sometimes only “thin slices” of behavior (i.e., short video segments), but also reports about facts or events. Studies on unconscious detection were more likely about accounts of complex (autobiographical) events. These differences in stimuli used (see Table 1) may explain why some replication attempts may have failed and why results from indirect approaches and UTT studies cannot really be compared.
Effects of Presentation Medium
Bond and DePaulo's (2006) meta-analysis showed the lowest detection accuracy in the visual-only format and the highest when statements were presented in audio or audiovisual form. In the visual-only modality, receivers may rely on stereotypical nonverbal cues (The Global Deception Research Team, 2006), most of which are non-diagnostic (see DePaulo et al., 2003; Sporer and Schwandt, 2006, 2007). Furthermore, senders in auditory-only communication settings may have less opportunity to strategically adjust or manipulate their demeanor (Burgoon et al., 2005) or gain from their attractiveness or likeableness. Another explanation could be that the availability of auditory information allows an analysis of verbal content cues, which are more diagnostic than nonverbal cues (e.g., Reinhard et al., 2013b; see below). However, the deception medium's influence was rarely explicitly tested in indirect and, especially, unconscious lie detection studies (see Table 1, “Mode” column). If these approaches indeed focus detectors' attention on diagnostic cues, we would expect that narrowing their attention to one communication channel may lead to even higher accuracy.
Furthermore, audiovisual communication may contain both non-diagnostic stereotypical nonverbal cues as well as diagnostic content cues. Processing both the visual information and the verbal content information may lead to an overload of information that is too difficult to process and integrate. This argument would imply that transcripts that allow only the use of verbal content cues should lead to higher accuracy than studies using only visual cues. On the other hand, if the UTT-based lie detection is supposed to work better in complex decision conditions, its advantage over conscious lie detection may be more apparent in full audiovisual deception.
Diagnosticity of Verbal Content Cues vs. Nonverbal Cues
Criticisms of the indirect and UTT approaches have argued that there are no diagnostic cues, and therefore, it is not surprising that they do not yield better results than direct lie–truth judgments (either on binary or continuous credibility rating scales). In contrast, we argue that there are valid verbal content cues to deception that have not been considered in these criticisms (Levine and Bond, 2014; Bond et al., 2015; see also Luke, 2019). Verbal content cues have been primarily studied in accounts about autobiographical events, not about reported emotions, opinions, or attitudes.
These recent reanalyses of cues to deception were based on DePaulo et al.'s (2003) comprehensive meta-analysis of published and unpublished studies of 1,338 estimates of 158 cues and on Bond and DePaulo's (2006) meta-analysis of accuracy rates in over 200 studies. It is important to note that in DePaulo et al.'s (2003) ubiquitously cited meta-analysis, all but four studies were from the years 1970 to 1997 (two studies were from the 1920s, one from 1998 and 1999 each). This era was dominated by Zuckerman, Ekman, DePaulo, Burgoon, and colleagues' communication research on nonverbal and paraverbal cues to deception, mostly about emotions but less frequently about facts or events. Hence, by relying on outdated databases, the mentioned recent reanalyses barely took the dozens of studies on verbal content cues to deception on CBCA (Steller and Köhnken, 1989) and reality monitoring criteria (Sporer, 1997, 2004) into account. DePaulo et al.'s meta-analysis included only six studies reporting any CBCA criteria, only four studies reporting any RM criteria, and two other studies reporting one additional verbal content cue each. It is clear that, since then, many more studies on verbal content cues have become available10 that have not been considered in the debate on unconscious lie detection by previous commentators.
In the last 5 years, several (but still incomplete) meta-analyses on verbal content cues have been published, which reported substantially larger effect sizes than any previous meta-analyses and reanalyses, with positive d values indicating a stronger presence of cues in true accounts—Amado et al., 2015: 16 studies with children, mean d = 0.40; Amado et al., 2016: 46 studies with adults, mean d = 0.25; Hauch et al., 2015: 79 linguistic cues in 44 studies, mean d for the length of accounts = 0.25; Hauch et al., 2016: 30 studies, mean d for eight training studies with content cues, d = 0.73 vs. six studies with nonverbal cues, d = 0.24; Oberlader et al., 201611: 56 effect sizes; Vrij's, 2005: vote-counting of 37 studies). While DePaulo et al. (2003) had reported a Mdn d = 0.10 across all cues, these content-based meta-analyses yielded much larger mean effect sizes for verbal content cues. Thus, the conclusions drawn from the reanalyses mentioned above are in dire need to be updated with these newer data sets of effect sizes.
However, before becoming too optimistic about these larger effect sizes, we caution the reader to critically scrutinize all, both older and newer, meta-analyses because a large number of studies contained therein have been statistically underpowered, biased by selective reporting and other “questionable research practices” and publication bias (for a thorough discussion of these and other issues, see Luke, 2019).
The availability of valid verbal content cues also has implications for the unconscious–conscious debate. The proper use of valid verbal content cues requires intensive training to code them reliably (Köhnken, 2004; Hauch et al., 2016, 2017, recommends a 3-week training program). Reliability of coding is a prerequisite for the discriminative validity of these cues. Applying them to a text corpus (written or oral) requires a careful analysis of all details of a statement, either by reading a transcript (or listening to an audio/videotape) repeatedly or taking notes. Normally, this requires a conscious, analytic approach—but see the study by Vrij et al. (2004) who have shown that rapid judgments may also catch some of these cues, which might be useful as a preliminary screening.
Complexity of Accounts
Studies often vary widely in the length of the accounts evaluated (measured in seconds or number of words). If the assumptions of UTT are correct, complexity appears as a prerequisite. Hence, to test its effectiveness, stimuli ought to have sufficient length and complexity. With complex events and consequently longer accounts, the burden on working memory is higher. Hence, receivers may only remember and use gist memory when distracted and verbatim memory when deliberating (Abadie et al., 2016). Materials (video- or audiotapes) ought to be also of sufficient resolution and auditory quality.
Furthermore, because senders' statements depend on the questions asked, questions should also be included if an interview was used. As we know from eyewitness testimony research from the beginning of the 20th century (Stern, 1904; see Sporer, 2008), an answer can only be properly understood when the question is also known. Regarding deception, there is a voluminous body of research on false confessions that became apparent when the questions asked in the police interrogations became available (e.g., Garrett, 2010).
Response Latencies Regarding Lies About Story Elements
Deception is not all or none. We need to distinguish studies investigating answers to individual questions from those evaluating an account's overall truthfulness assessed by a final veracity judgment. In the psychophysiological literature (National Research Council, 2003), comparisons are made not between different accounts or stimulus persons or between liars and truth-tellers but between response options of individual items (like the alternatives in a multiple-forced-choice test). A frequently used measure is response latency as an indirect indicator of an association (for a historical review, see Antonelli and Sporer, in press). For example, in a recent meta-analysis of studies on the concealed information test and similar paradigms, participants responded considerably faster to true than false items (Suchotzki et al., 2017). However, the effects varied widely, and countermeasures reduced the effect.
The Case for Conscious vs. Unconscious Decision-Making in Criminal Courts
Consider for a moment an example of lie detection in a most serious context, the evaluation of the credibility of a witness or a defendant in a criminal court case. Depending on the judiciary system, the decision about the determination of guilt lies in the hands of professional judges or a jury.
Leaving the issue of group vs. individual decision-making aside, judges or jurors must decide whether a witness (or a defendant) is telling the truth. For a procedure to be fair, would you not expect, or even demand, that fact finders pay close attention to the witness's utterances, including both the content and the “demeanor” with which the statement is provided?12 Would you not expect that any statements be critically scrutinized by consciously deliberating any nuances and possibilities indicating the possibility of deception? Would you consider it fair if a judge or jurors were not paying attention before arriving at a decision in the case?
When deciding on a sentence, judges must provide “reasons” why a particular decision was made, including a rational evaluation of the witness's incriminating evidence. This requirement led the German Supreme Court to overturn lower-court decisions (Bungesgerichtshof in Strafsachen (BGHSt), 1999), in which “experts” had insufficiently assessed key witness statements. As a consequence, the Supreme Court demanded that, in future cases, experts should follow a series of guidelines for Statement Validity Analysis and Criteria-Based Content Analysis (see Steller and Köhnken, 1989) and explicitly refer to them in their expert testimony. In our view, this would leave no room for unconscious assessments. In another case in Canada, the question arose whether a witness ought to remove her niqab while testifying for her demeanor to become visible (see Snook et al., 2017; Denault and Jupe, 2018). This is an interesting empirical issue demonstrating the importance of our discussion in the real world.
Of course, there is a long literature on judicial decision-making demonstrating that “unconscious” factors may determine such decisions (e.g., Goodman-Delahunty and Sporer, 2010). Consequently, both experts and courts may be affected by a host of biases amply documented in the social psychological literature (Pohl, 2004). We also know from that research that reasons to justify a decision are often post-hoc rationalizations of a previous decision to make it appear rational and protect judges from appellate courts overturning their decisions. Nonetheless, we would hope that decisions about defendants' lives are made with the uttermost diligence and careful (“conscious”) consideration of all facts pertinent to the decision. But just how these facts are integrated into a final decision remains a puzzle.
Conclusion
Despite inconsistent results, we argue that indirect methods to veracity assessment deserve further attention for the following reasons: (1) Indirect measures and “unconscious” approaches should be pursued separately. (2) Critiques on indirect approaches were based on outdated meta-analytical databases summarizing mainly research on nonverbal and paraverbal cues and only a handful of verbal content cues. These databases need to be updated by newer studies and meta-analyses on verbal content cues with higher diagnostic value. (3) Research on detecting emotions and opinions has to be treated separately from assessing the veracity of complex reports about facts or events. (4) More emphasis should be placed on stimulus than on participant sampling. (5) The relationships and mutual dependencies of indirect cues and explicit veracity judgments need to be further explored. (6) Evidence for “unconscious” processes needs to be provided.
Author Contributions
All authors have made substantial contributions to planning, writing and commenting the content, and also approved the version to be submitted in the journal.
Funding
This paper was supported by the Excellence Initiative - Research University programme at Nicolaus Copernicus University.
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Footnotes
1. ^Some authors have used the term “indirect,” others “implicit” (e.g., Granhag, 2006) but referring to similar methods. We generally use the term “indirect” except when referring to specific studies where we use the respective authors' term. Otherwise, we consider them interchangeable but distinct from “unconscious lie detection.”
2. ^There is a severe criticism of the conclusions drawn from the statistics presented in this study by Franz and von Luxburg (2015) that is beyond the scope of this article.
3. ^Apart from indirect or unconscious lie detection, some studies used rapid judgments or other ratings to assess verbal content like plausibility, detailedness, etc. (Table 1, Section 1a). We argue that this type of global impression judgment may be valuable as a screening device when deciding whether a detailed content analysis proposed in the SVA/CBCA or the Reality Monitoring approach (see below) may be useful. However, the latter methods require a thorough conscious analysis of a text corpus that is only possible after intensive training, with high demands on interrater reliability. These types of content analysis are not an indirect or unconscious process approach.
4. ^As presented in Table 1, Section I, “thinking hard” is the most frequently used indirect measure (40% of studies), followed by a question about a sender's confidence (20% of studies).
5. ^As far as we can tell, only Reinhard et al. (2013a, Exp. 5) (Table 1, Section II) made such an attempt. However, the pattern of objectively diagnostic and non-diagnostic cues in that study seems to reflect stereotypical cues and is also at odds with research on verbal content cues to deception (Amado et al., 2016) and the study on rapid judgments by Vrij et al. (2004).
6. ^See Fiedler (1989), Köhnken (1990); Sporer and Küpper (1995), and Hartwig and Bond (2011) for detailed descriptions of the model in the lie detection domain.
7. ^Some of these recommendations are suitable for all types of lie detection studies.
8. ^Initially used in radar technology and further developed in medical and perception research, the SDT methodology here is borrowed from face recognition and eyewitness identification studies, where it has become standard (Macmillan and Creelman, 2005; Wixted and Wells, 2017).
9. ^In Table 1, the total number of senders varied between 1 and 73 (M = 17.23; SD = 18.05) in studies on indirect measures (Section I and Ia) and between 1 and 10 (M = 7.58; SD = 2.74) in studies on unconscious lie detection (except one study that had a total of 72 senders but only one per participant) (Section II).
10. ^The first author has over 100 studies on verbal content cues in his database.
11. ^Note that this meta-analysis integrated effect sizes of mean differences of cue summary scores with effect sizes from accuracy rates of raters and accuracy rates based on multivariate statistical analyses. Because the latter were usually obtained without cross-validation, their integration with effect sizes based on accuracy rates of human raters is problematic (see Sporer et al., 2021). The same criticism holds for Vrij (2005, Table 4) averaging of rater judgment accuracies and accuracies from multivariate statistical analyses.
12. ^In the communication literature, researchers have investigated for over half a century which aspects of “demeanor” (i.e., nonverbal and paraverbal cues) may reveal deception. In a widely received article by Denault and several dozen deception researchers (Denault et al., 2020), judicial, governmental, and security agencies are criticized for relying on pseudoscientific advisors (e.g., synergologists) instead of relying on proper, peer-reviewed science, costing the public millions of dollars. Police, attorneys, and judges often appear uninformed (or even unaware of its existence) about the evidence-based detection of deception literature (Jupe and Denault, 2019).
References
Abadie, M., Waroquier, L., and Terrier, P. (2016). Information presentation format moderates the unconscious-thought effect: The role of recollection. Memory 24, 1123–1133. doi: 10.1080/09658211.2015.1070179
Albrechtsen, J. S., Meissner, C. A., and Susa, K. J. (2009). Can intuition improve deception detection performance? J. Exp. Soc. Psychol. 45, 1052–1055. doi: 10.1016/j.jesp.2009.05.017
Amado, B. G., Arce, R., and Fariña, F. (2015). Undeutsch hypothesis and criteria-based content analysis: a meta-analytic review. Eur. J. Psychol. Appl. Legal Context 7, 3–12. doi: 10.1016/j.ejpal.2014.11.002
Amado, B. G., Arce, R., Fariña, F., and Vilariño, M. (2016). Criteria-Based Content Analysis (CBCA) reality criteria in adults: a meta-analytic review. Int. J. Clin. Health Psychol. 16, 201–210. doi: 10.1016/j.ijchp.2016.01.002
Anderson, D. E., DePaulo, B. M., Ansfield, M. E., Tickle, J. J., and Green, E. (1999). Beliefs about cues to deception: mindless stereotypes or untapped wisdom? J. Nonverbal Behav. 23, 67–89. doi: 10.1023/A:1021387326192
Antonelli, M., and Sporer, S. L. (in press). The history of eyewitness testimony the foundations of the “Lie Detector” in Austria Italy. Rivista Sperimentale di Psichiatria.
Bond, C. F., and DePaulo, B. M. (2006). Accuracy of deception judgements. Personal. Soc. Psychol. Rev. 10, 214–234. doi: 10.1207/s15327957pspr1003_2
Bond, C. F., Levine, T. R., and Hartwig, M. (2015). “New findings in nonverbal lie detection,” in Detecting Deception, eds P. A. Granhag, A. Vrij, and B. Verschuere (Chichester: Wiley Blackwell), 38–58. doi: 10.1002/9781118510001.ch2
Bond, C. F. Jr., and DePaulo, B. M. (2008a). Individual differences in judging deception: accuracy and bias. Psychol. Bull. 134, 477–492. doi: 10.1037/0033-2909.134.4.477
Bond, C. F. Jr., and DePaulo, B. M. (2008b). Individual differences in judging deception: reply to O'Sullivan (2008) and Pigott and Wu (2008). Psychol. Bull. 134, 501–503. doi: 10.1037/0033-2909.134.4.501
Bond, C. F. Jr., Omar, A. S., Mahmoud, A., and Bonser, N. (1990). Lie detection across cultures. J. Nonverbal Behav. 14, 189–204. doi: 10.1007/BF00996226
Bradford, D., Goodman-Delahunty, J., and Brooks, K. R. (2013). The impact of presentation modality on perceptions of truthful and deceptive confessions. J. Criminol. 164546, 1–10. doi: 10.1155/2013/164546
Burgoon, J. K., Blair, J. P., and Strom, R. E. (2005). “Heuristics and modalities in determining truth versus deception,” in Proceedings of the 38th Annual Hawaii International Conference on System Sciences (Big Island, HI). doi: 10.1109/HICSS.2005.294
Campbell, D. T., and Stanley, J. C. (1963). Experimental and Quasi-Experimental Designs. Chicago: Rand McNally.
Canter, D., Ioannou, M., Youngs, D., and Chungh, G. (2016). Person perception aspects of judgments of truthfulness in public appeals. Psychiatry Psychol. Law. 23, 547–562. doi: 10.1080/13218719.2015.1081315
Denault, V., and Jupe, L. M. (2018). Justice at risk! An evaluation of a pseudoscientific analysis of a witness' nonverbal behavior in the courtroom. J. Forensic Psychiatry Psychol. 126, 1–22. doi: 10.1080/14789949.2017.1358758
Denault, V., Plusquellec, P., Jupe, L. M., St- Yves, M., Dunbar, N. E., Hartwig, M., et al. (2020). The analysis of nonverbal communication: the dangers of pseudoscience in security and justice contexts. Anuario de Psicologia Jurídica 30, 1–12. doi: 10.5093/apj2019a9
DePaulo, B. M., Jordan, A., Irvine, A., and Laser, P. S. (1982a). Age changes in the detection of deception. Child Dev. 53, 701–709. doi: 10.2307/1129383
DePaulo, B. M., Lindsay, J. J., Malone, B. E., Muhlenbruck, L., Charlton, K., and Cooper, H. (2003). Cues to deception. Psychol. Bull. 129, 74–118. doi: 10.1037/0033-2909.129.1.74
DePaulo, B. M., and Morris, W. L. (2004). “Discerning lies from truths: behavioral cues to deception and the indirect pathway of intuition,” in The Detection of Deception in Forensic Contexts, eds P. A. Granhag, and L. Strömwall (Cambridge: Cambridge University Press), 15–41. doi: 10.1017/CBO9780511490071.002
DePaulo, B. M., and Rosenthal, R. (1979). Telling lies. J. Pers. Soc. Psychol. 37, 1713–1722. doi: 10.1037/0022-3514.37.10.1713
DePaulo, B. M., Rosenthal, R., Green, C. R., and Rosenkrantz, J. (1982b). Diagnosing deceptive and mixed messages from verbal and nonverbal cues. J. Exp. Soc. Psychol. 18, 433–446. doi: 10.1016/0022-1031(82)90064-6
Dijksterhuis, A., and Nordgren, L. F. (2006). A theory of unconscious thought. Perspect Psychol. Sci. 1, 95–109. doi: 10.1111/j.1745-6916.2006.00007.x
Dunbar, N. E., Jensen, M. L., Harvell-Bowman, L. A., Kelley, K. M., and Burgoon, J. K. (2017). The viability of using rapid judgments as a method of deception detection. Commun. Methods Meas. 11, 121–136. doi: 10.1080/19312458.2016.1276893
Ericsson, K. A., and Simon, H. A. (1993). Protocol Analysis: Verbal Reports as Data. Boston, MA: A Bradford Book. doi: 10.7551/mitpress/5657.001.0001
Evanoff, C., Porter, S., and Black, P. J. (2016). Video killed the radio star? The influence of presentation modality on detecting high-stakes, emotional lies. Legal Criminol. Psychol. 21, 332–343. doi: 10.1111/lcrp.12064
Feeley, T. H., and Young, M. J. (2000). Self-reported cues about deceptive and truthful communication: the effects of cognitive capacity and communicator veracity. Commun. Q. 48, 101–119. doi: 10.1080/01463370009385585
Fiedler, K. (1989). “Suggestion and credibility: lie detection based on content-related cues,” in Suggestion and Suggestibility, eds V. A. Gheorghiu, P. Netter, H. J. Eysenck, and R. Rosenthal (Berlin: Springer-Verlag), 323–335. doi: 10.1007/978-3-642-73875-3_25
Franz, V. H., and von Luxburg, U. (2015). No evidence for unconscious lie detection: a significant difference does not imply accurate classification. Psychol. Sci. 26, 1646–1648. doi: 10.1177/0956797615597333
Gigerenzer G. Todd P. M. The ABC Research Group (1999). Simple Heuristics that Make us Smart. New York, NY: Oxford University Press.
Goodman-Delahunty, J., and Sporer, S. L. (2010). Unconscious influences in sentencing decisions: a research review of psychological sources of disparity. Aust. J. Forensic Sci. 42, 19–36. doi: 10.1080/00450610903391440
Granhag, P. A. (2006). Rethinking implicit lie detection. J. Credibil. Assess. Witness Psychol. 7, 180–190.
Hart, C. L., Fillmore, D. G., and Griffith, J. D. (2009). Indirect detection of deception: looking for change. Curr. Res. Social Psychol. 14, 134–142.
Hartwig, M., and Bond, C. F. (2011). Why do lie-catchers fail? A lens model meta-analysis of human lie judgments. Psychol. Bull. 137, 643–659. doi: 10.1037/a0023589
Hauch, V., Blandón-Gitlin, I., Masip, J., and Sporer, S. L. (2015). Are computers effective lie detectors? A meta-analysis of linguistic cues to deception. Personal. Soc. Psychol. Rev. 19, 307–342. doi: 10.1177/1088868314556539
Hauch, V., Sporer, S. L., Masip, J., and Bland?n-Gitlin, I. (2017). Can credibility criteria be assessed reliably? A meta-analysis of criteria-based content analysis. Psychol. Assess. 29, 819–834. doi: 10.1037/pas0000426
Hauch, V., Sporer, S. L., Michael, S. W., and Meissner, C. A. (2016). Does training improve the detection of deception? A meta-analysis. Commun. Res. 43, 283–343. doi: 10.1177/0093650214534974
Hurd, K., and Noller, P. (1988). Decoding deception: a look at the process. J. Nonverbal Behav. 12, 217–233. doi: 10.1007/BF00987489
Jupe, L. M., and Denault, V. (2019). Science or pseudoscience? A distinction that matters for police officers, lawyers and judges. Psychiatry Psychol. Law 26, 753–765. doi: 10.1080/13218719.2019.1618755
Kaminski, K. S., and Sporer, S. L. (2017). Discriminating between correct and incorrect eyewitness identifications: the use of appropriate cues. J. Exp. Psychol. 23, 59–70. doi: 10.1037/xap0000110
Klaver, J. R., Lee, Z., Spidel, A., and Hart, S. D. (2009). Psychopathy and deception detection using indirect measures. Legal Criminol. Psychol. 14, 171–182. doi: 10.1348/135532508X289964
Köhnken, G. (2004). Statement validity analysis and the “detection of the truth,” in The Detection of Deception in Forensic Contexts, eds P. A. Granhag and L. A. Strömwall (Cambridge, UK: Cambridge University Press), 41–63. doi: 10.1017/CBO9780511490071.003
Landström, S., and Granhag, P. A. (2010). In-court versus out-of-court testimonies: children's experiences and adults' assessments. Appl. Cogn. Psychol. 24, 941–955. doi: 10.1002/acp.1606
Landström, S., Granhag, P. A., and Hartwig, M. (2005). Witnesses appearing live versus on video: effects on observers' perception, veracity assessments and memory. Appl. Cogn. Psychol. 19, 913–933. doi: 10.1002/acp.1131
Landström, S., Granhag, P. A., and Hartwig, M. (2007). Children's live and videotaped testimonies: how presentation mode affects observers' perception, assessment and memory. Legal Criminol. Psychol. 12, 333–348. doi: 10.1348/135532506X133607
Levine, T. R. (2019). Five reasons why i am skeptical that indirect or unconscious lie detection is superior to direct deception detection. Front. Psychol. 10:1354. doi: 10.3389/fpsyg.2019.01354
Levine, T. R., and Bond, C. F. (2014). Direct and indirect measures of lie detection tell the same story: a reply to ten Brinke, Stimson, and Carney (2014). Psychol. Sci. 25, 1960–1961. doi: 10.1177/0956797614536740
Luke, T. J. (2019). Lessons from Pinocchio: cues to deception may be highly exaggerated. Perspect. Psychol. Sci. 14, 646–671. doi: 10.1177/1745691619838258
Macmillan, N. A., and Creelman, C. D. (2005). Detection Theory: A User's Guide. Mahwah, NJ: Erlbaum. doi: 10.4324/9781410611147
Moi, W. Y., and Shanks, D. R. (2015). Can lies be detected unconsciously? Front. Psychol. 6:1221. doi: 10.3389/fpsyg.2015.01221
National Research Council (2003). The Polygraph and Lie Detection. Committee to Review the Scientific Evidence on the Polygraph. Division of Behavioral and Social Sciences and Education. Washington, DC: The National Academies Press. Available online at: https://www.nap.edu/read/10420/chapter/1#ii
Nisbett, R. E., and Wilson, T. D. (1977). Telling more than we can know: verbal reports on mental processes. Psychol. Rev. 84, 231–259. doi: 10.1037/0033-295X.84.3.231
Oberlader, V. A., Naefgen, C., Koppehele-Gossel, J., Quinten, L., Banse, R., and Schmidt, A. F. (2016). Validity of content-based techniques to distinguish true and fabricated statements: a meta-analysis. Law Hum. Behav. 40, 440–457. doi: 10.1037/lhb0000193
Pohl, R. F., (eds.). (2004). Cognitive Illusions: A Handbook on Fallacies and Biases in Thinking, Judgment and Memory. London: Psychology Press.
Reinhard, M. A., Greifeneder, R., and Scharmach, M. (2013a). Unconscious processes improve lie detection. J. Person. Soc. Psychol. 105, 721–739. doi: 10.1037/a0034352
Reinhard, M. A., Sporer, S. L., and Scharmach, M. (2013b). Perceived familiarity with a judgmental situation improves lie detection ability. Swiss J. Psychol. 72, 43–52. doi: 10.1024/1421-0185/a000098
Smith, A. M., and Leach, A.-M. (2019). Confidence can discriminate between accurate and inaccurate lie decisions. Perspect. Psychol. Sci. 14, 1062–1071. doi: 10.1177/1745691619863431
Snook, B., McCardle, M. I., Fahmy, W., and House, J. C. (2017). Assessing truthfulness on the witness stand: eradicating deeply rooted pseudoscientific beliefs about credibility assessment by triers of fact. Can. Crim. Law Rev. 22, 297–306.
Sporer, S. L. (1997). The less travelled road to truth: verbal cues in deception detection in accounts of fabricated and self-experiences events. Appl. Cogn. Psychol. 11, 373–397. doi: 10.1002/(SICI)1099-0720(199710)11:5 <373::AID-ACP461>3.0.CO
Sporer, S. L. (2004). “Reality monitoring and the detection of deception,” in Deception Detection in Forensic Contexts, eds P.-A. Granhag and L. Stromwall (Cambridge: Cambridge University Press), 64–102. doi: 10.1017/CBO9780511490071.004
Sporer, S. L. (2008). Lessons from the origins of eyewitness testimony research in Europe. Appl. Cogn. Psychol. 22, 737–757. doi: 10.1002/acp.1479
Sporer, S. L., and Küpper, B. (1995). Realitätsüberwachung und die Beurteilung des Wahrheitsgehaltes von Erzaehlungen: Eine experimentelle Studie [Reality monitoring and the judgment of credibility of stories: an experimental investigation]. Zeitschrift fur Sozialpsychologie 26, 173–193.
Sporer, S. L., Manzanero, A. L., and Masip, J. (2021). Optimizing CBCA and RM research: recommendations for analyzing and reporting data on content cues to deception. Psychol. Crime Law 27, 1–39. doi: 10.1080/1068316X.2020.1757097
Sporer, S. L., and Schwandt, B. (2006). Paraverbal indicators of deception: a meta-analytic synthesis. Appl. Cogn. Psychol. 20, 421–446. doi: 10.1002/acp.1190
Sporer, S. L., and Schwandt, B. (2007). Moderators of nonverbal indicators of deception: a meta-analytic synthesis. Psychol. Public Policy Law 13, 1–34. doi: 10.1037/1076-8971.13.1.1
Stel, M., and Van Dijk, E. (2018). When do we see that others misrepresent how they feel? detecting deception from emotional faces with direct and indirect measures. Social Influence 13, 137–149. doi: 10.1080/15534510.2018.1473290
Stel, M., Van Dijk, E., and Olivier, E. (2009). You want to know the truth? then don't mimic!. Psychol. Sci. 20, 693–699. doi: 10.1111/j.1467-9280.2009.02350.x
Steller, M., and Köhnken, G. (1989). “Criteria-based statement analysis,” in Psychological Methods in Criminal Investigation and Evidence, ed D. C. Raskin (Berlin: Springer Publishing Company), 217–245.
Stern, L. W. (1904). “Die Aussage als geistige Leistung und als Verhörsprodukt [Testimony as mental achievement and product of interrogation],” in Beiträge zur Psychologie der Aussage, Vol. 1, ed. L. W. Stern (Leipzig: Barth), 267–415. doi: 10.1515/9783111527567
Street, C. N. H., and Richardson, D. C. (2015). The focal account: Indirect lie detection need not access unconscious, implicit knowledge. J. Exp. Psychol. Appl. 21, 342–355. doi: 10.1037/xap0000058
Street, C. N. H., and Vadillo, M. A. (2016). Can the unconscious boost lie-detection accuracy? Curr. Direct. Psychol. Sci. 25, 246–250. doi: 10.1177/0963721416656348
Suchotzki, K., Verschuere, B., Van Bockstaele, B., Ben-Shakhar, G., and Crombez, G. (2017). Lying takes time: a meta-analysis on reaction time measures of deception. Psychol. Bull. 143, 428–453. doi: 10.1037/bul0000087
ten Brinke, L., Stimson, D., and Carney, D. R. (2014). Some evidence for unconscious lie detection. Psychol. Sci. 25, 1098–1105. doi: 10.1177/0956797614524421
The Global Deception Research Team (2006). A world of lies. J. Cross. Cult. Psychol. 37, 60–74. doi: 10.1177/0022022105282295
Ulatowska, J. (2010). The influence of providing the context of the assessment on the accuracy of the indirect method of deception detection. Prob. Forensic Sci. 84, 380–391.
Ulatowska, J. (2014). Different questions – different accuracy? The accuracy of various indirect question types in deception detection. Psychiatry Psychol. Law 21, 231–240. doi: 10.1080/13218719.2013.803278
Ulatowska, J. (2018). Accuracy of indirect method in detection of false intent. Legal Criminol. Psychol. 23, 24–36. doi: 10.1111/lcrp.12116
Ulatowska, J. (2020). “Metoda pośredniej oceny szczerości – szansa na wyzsza skuteczność wykrywania kłamstwa czy technika bez mozliwości praktycznego zastosowania? [Indirect method of veracity assessment: a chance for accurate lie detection or a technique without real-life application possibilities],” in Nowe technologie i metody w psychologii, eds M. Trojan and M. Gut (Warsaw: Liberi Libri), 395–410. doi: 10.47943/lib.9788363487430.rozdzial18
van't Veer, A. E., Gallucci, M., Stel, M., and van Beest, I. (2015). Unconscious deception detection measured by finger skin temperature and indirect veracity judgments - results of a registered report. Front. Psychol. 6:672. doi: 10.3389/fpsyg.2015.00672
Vrij, A. (2005). Criteria-based content analysis: a qualitative review of the first 37 studies. Psychol. Public Policy Law 11, 3–41. doi: 10.1037/1076-8971.11.1.3
Vrij, A., Edward, K., and Bull, R. (2001). Police officers' ability to detect deceit: the benefit of indirect deception detection measures. Legal Criminol. Psychol. 6, 185–196. doi: 10.1348/135532501168271
Vrij, A., Evans, H., Akehurst, L., and Mann, S. (2004). Rapid judgements in assessing verbal and nonverbal cues: their potential for deception researchers and lie detection. Appl. Cogn. Psychol. 18, 283–296. doi: 10.1002/acp.964
Wells, G. L., and Windschitl, P. D. (1999). Stimulus sampling and social psychological experimentation. Pers. Soc. Psychol. Bull. 25, 1115–1125. doi: 10.1177/01461672992512005
Wixted, J. T., and Wells, G. L. (2017). The relationship between eyewitness confidence and identification accuracy: a new synthesis. Psychol. Sci. Public Interest 18, 10–65. doi: 10.1177/1529100616686966
Wu, S., Mei, H., and Yan, J. (2019). Do not think carefully? Re-examining the effect of unconscious thought on deception detection. Front. Psychol. 10:893. doi: 10.3389/fpsyg.2019.00893
Keywords: deception detection, direct lie detection, indirect lie detection, unconscious lie detection, veracity assessment, stimulus sampling, methodological recommendations for indirect and unconscious deception detection
Citation: Sporer SL and Ulatowska J (2021) Indirect and Unconscious Deception Detection: Too Soon to Give Up? Front. Psychol. 12:601852. doi: 10.3389/fpsyg.2021.601852
Received: 01 September 2020; Accepted: 12 April 2021;
Published: 26 May 2021.
Edited by:
Emmanouil Konstantinidis, University of Warwick, United KingdomReviewed by:
Mircea Zloteanu, Kingston University, United KingdomHugues Delmas, Université Paris 8, France
Copyright © 2021 Sporer and Ulatowska. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Joanna Ulatowska, julatowska@umk.pl