- 1Department of Computer Science, University of Houston, Houston, TX, United States
- 2School of Computer Science and AI, Tel Aviv University, Tel Aviv, Israel
- 3Department of Computer Sciences, University of Houston, Houston, TX, United States
- 4InstaBase, San Francisco, CA, United States
- 5Department of Computer Sciences, University of California, Berkeley, Berkeley, CA, United States
Introduction: Internet-based economies and societies are drowning in deceptive attacks. These attacks take many forms, such as fake news, phishing, and job scams, which we call “domains of deception.” Machine learning and natural language processing researchers have been attempting to ameliorate this precarious situation by designing domain-specific detectors. Only a few recent works have considered domain-independent deception. We collect these disparate threads of research and investigate domain-independent deception.
Methods: First, we provide a new computational definition of deception and break down deception into a new taxonomy. Then, we briefly mention the debate on linguistic cues for deception. We build a new comprehensive real-world dataset for studying deception. We investigate common linguistic features for deception using both classical and deep learning models in a variety of situations including cross-domain experiments.
Results: We find common linguistic cues for deception and give significant evidence for knowledge transfer across different forms of deception.
Discussion: We list several directions for future work based on our results.
1 Introduction
History is replete with famous lies and deceptions. Examples include P. T. Barnum, Nicolo Machiavelli, Sun Tzu, Operation Mincemeat, and the Trojan Horse (Levine, 2014). A chronology of deception is included in Levine (2014). More recently, the proliferation of deceptive attacks such as fake news, phishing, and disinformation is rapidly eroding trust in Internet-dependent societies. The situation has deteriorated so much that 45% of the US population believes the 2020 US election was stolen.1
Social media platforms have come under severe scrutiny regarding how they police content. Facebook and Google are partnering with independent fact-checking organizations that typically employ manual fact-checkers.
Natural-language processing (NLP) and machine learning (ML) researchers have joined the fight by designing fake news, phishing, and other kinds of domain-specific detectors.
Building single-domain detectors may be sub-optimal. Composing them sequentially requires more time, and composing them in parallel requires more hardware. Moreover, building single-domain detectors means one can only react to new forms of deception after they emerge.
Our goal here is to spur research on domain-independent deception. Unfortunately, research in this area is currently hampered by the lack of computational definitions and taxonomy, high-quality datasets, and systematic approaches to domain-independent deception detection. Thus, the results are neither generalizable nor reliable, leading to much confusion.
Accordingly, we make the following contributions:
• We propose a new computational definition and a new comprehensive taxonomy of deception. (We use the unqualified term “deception” for the domain-independent case. When the goals of the deception are unclear, we refer to “lies”).
• We examine the debate on linguistic deception detection, identify works that demonstrate the challenges that must be overcome to develop domain-independent deception detectors, and examine them critically.
• We conduct linguistic analysis of several detection datasets for general cues and find several statistically significant ones.
• We conduct deep learning experiments of deception sets and study correlations in performance for pairs of datasets.
This article is organized as follows: Section 2 presents a new definition of deception. Section 3 introduces our new taxonomy. Section 4 summarizes related work. Sections 5 and 6 describe our experiments, results, and analysis of domain-independent markers for deception. Cross-domain detection results are in Section 7. Finally, Section 8 presents some conclusions and directions for the future. The appendices provide the list of features tested and some preliminary significance testing of cues on four public deception datasets.
2 Definition
We first examine a general definition of deception, taken from Galasinski (2000), intended to capture a wide variety of deceptive situations and attacks.
Definition 1 (Preliminary). Deception is an intentional act of manipulation to gain compliance. Thus, it has at least one source, one target, and one goal. The source is intentionally manipulating the target into beliefs, or actions, or both, intended to achieve the goals.
Since we are interested in automatic verifiability, we would like to modify this definition of deception and propose one that is computationally feasible. Because intentions are notoriously hard to establish, we will use the effect of exposing the manipulation/goals instead.
Our revised definition is the following:
Definition 2 (Deception). Deception is an act of manipulation designed to gain compliance such that, exposing the manipulation or the goal(s) of compliance significantly decreases the chance of compliance. Thus, it has at least one source, one target, and one goal. The source is manipulating the target into beliefs, or action, or both, intended to achieve the goals.
One might argue that the goals of deception should be harmful to an individual or organization. However, this would necessitate either a computational definition of harm or a comprehensive list of potential harms, which could be checked computationally and is, therefore, a less desirable alternative.
To formalize our definition, we borrow from the language of Markov decision processes. Let A be an action taken by an actor, and let C be a desired compliance state. We use K(A, T) to denote the action A plus the full and truthful explanation of the actor's relevant private information to target T. We formalize (computational) deception using conditional probabilities as follows:
Definition 3 (Computational Deception—Formalized). An action A deceives target T if
Moreover, we can quantify the degree to which A is deceptive by the amount θ, where 0 ≤ θ ≤ 1.
Definition 4 (Computational Deception—Quantified). An action A θ-deceives target T if
In practice, practitioners can apply this by exposing the manipulation and/or goals and measuring the change in compliance rates. For example, a Florida woman recently sued Kraft alleging that the “ready in min” on the label of their microwavable Velveeta Shells & Cheese is deceptive. To determine whether the claim is, in fact, deceptive, a researcher could present the product by itself to one group of random consumers and the product with an explanation that the min does not include the time to add water to another group. If there is a statistically significant decrease in purchases (which is the desired compliance) for the group with the explanation, then the claim is deceptive.
There is some work on finding out how good humans are at detecting certain kinds of deceptive attacks. For the detection capabilities of automatic detectors on specific domains of deception, one can look at surveys on fake news detection (Sharma et al., 2019; Zhou et al., 2019, 2020) and phishing detection (Das et al., 2020).
3 Taxonomy and examples
In this section, we give a new taxonomy for deception and some examples to illustrate it. Note that this taxonomy is intended to be comprehensive and capture all nuances of deception, which means also the legal aspects when the source of the deception is being charged with a crime for example. Hence, this taxonomy will take into account the intent of the source in contrast with the above section.
There have been a few attempts at constructing taxonomies for fake news, phishing, or other forms of deception. Molina et al. (2021) give a taxonomy of fake news with four dimensions: message and linguistic, sources and intentions, structural, and network. Kapantai et al. (2021) conducted a systematic search for papers proposing taxonomies for disinformation and synthesized a taxonomy with three dimensions: factuality, motivation, and verifiability.
No one, to our knowledge, has given a comprehensive taxonomy of real-world deception.
3.1 The new taxonomy
We put forward a multi-dimensional taxonomy (Figure 1). Under our definition, deception explicitly involves four elements: (1) agents: the sources, and the targets; (2) stratagems for manipulation; (3) goals; and (4) threat/mechanisms of exposure. These explicit elements can be further broken down as follows:
1) Agents. Rowe (2006) calls this category “participant,” and he further elaborates this into: (a) agent, who initiates the action, (b) beneficiary, who benefits, (c) object, what the action is done to, and (d) recipient, who receives the action. Rowe also includes experiencer (“who senses the action”) and instrument (“what helps accomplish the action”) components in this category, but we include them in the Channel category below.
1a) Sources. This includes human (individual or group), bot, etc., or mixed, in other words, combinations such as a human assisted by a bot. This Sources category includes initiators and beneficiaries.
1b) Targets. This includes humans (individual or group), automatic detectors, or both. For example, spam targets automatic detectors, and phishing targets humans, but needs to fool automatic detectors also. This Targets category includes the objects and the recipients.
2) Stratagems. The stratagem subtree in the taxonomy includes two sub-taxonomies for persuasion and action, which we discuss below. We believe that persuasion is fundamental to deception since its goal is to change the reasoning of the target(s), with the deception's end goal of compliance. The action taxonomy is adapted from Rowe (2006). It includes space, time, causality, quality, essence, and speech-act theory, which specifies the external and internal preconditions for the action. The persuasion taxonomy combines Cialdini (2006) and Da San Martino et al. (2023).
3) Goals.
3a) Harmless: satire, parody, satisfying participation, as in a laboratory experiment where participants may be asked to lie, etc.
3b) Harmful. This includes a wide range of objectives, such as stealing money or identity information, malware installation, manipulation of votes, planting fear, sowing confusion, initiating chaos, gaining an unfair edge in a competition (e.g., swaying opinions and preferences on products), persuading people to take harmful actions, winning competitions/games, etc. We avoid the terms defensive and offensive since they are dependent on the perspective of the participants/agents.
4) Exposure.
4a) Facticity. Can we establish whether it is factual or not? For example, currently, we are unable to establish the truth or falsity of utterances such as, “There are multiple universes in existence right now.”
4b) Verifiability. Assuming facticity, how easy or difficult it is to verify whether it is legitimate or deceptive? Here, we are interested in machine or automatic verification. If a simple machine-learning algorithm can detect it with high recall and precision, we will deem it easy.

Figure 1. Proposed deception taxonomy—the full manipulation (or stratagem) and motivation subtrees are not shown.
In addition, there are four implicit concepts in the taxonomy: (1) motivations behind the goals; (2) communication channels or media; (3) modality of deception; and (4) manner or timeliness of the exchange.
1) Motivation. This is the rationale for the goals. The agents involved and their characteristics reveal the underlying motivations, which could be political hegemony (nation-states), religious domination, revenge (disgruntled employee), ideological gains, money, control, power, etc.
2) Channel. This dimension includes two aspects:
2a) Breadth: Whether the targets are a few specific individuals or detector types or broad classes of people/categories of detectors.
2b) Media. How the deceptive capsule is conveyed to the target. Media also includes the experiencer and instrument components of Rowe (2006).
3) Modality. This dimension refers to the presentation of deceptive content. It includes:
3a) Unimodal. This includes only one type of modality such as (a) gestural (i.e., body language is used to deceive), (b) audio (a.k.a. verbal), (c) textual (e.g., SMS/email), and (d) visual (e.g., images or videos).
3b) Multimodal: combinations of different modalities. For example, audio-visual has both speech and visual components but lacks face-to-face communication in which gestures could facilitate deception.
4) Manner/Timeliness.
4a) Interactive/synchronous. A real-time interview or debate is an interactive scenario.
4b) Non-interactive/asynchronous. An Amazon Mechanical Turker typing a deceptive opinion or essay is a non-interactive one. An asynchronous interaction can have multiple stages or steps some (but not all) of which may be synchronous.
3.1.1 Stratagems
Rowe's (2006) approach is based on linguistics. He states, “Each action has associated concepts that help particularize it, and these are conveyed in language by modifiers, prepositional phrases, participial phrases, relative clauses, infinitives, and other constructs.” These associated concepts are called “semantic cases” Fillmore, (1968) in analogy to the syntactic cases that occur in some languages for nouns. Rowe claims that “every deception action can be categorized by an associated semantic case or set of cases.” However, there is no canonical list of semantic cases in linguistics. Rowe prefers the detailed list from Copeck et al. (1992), which he supplements with two important relationships from artificial intelligence, the upward type-supertype and upward part-whole links, and two speech-act conditions from Austin (1975), to get 32 cases altogether. However, since we include his “participant” category in the Agents and Channel categories, we have only 26 subcategories in the Stratagems category.
1. Space, which consists of: (a) direction, of the action, (b) location-at, where something occurred, (c) location-from, where something started, (d) location-to, where something finished, (e) location-through, where some action passed through, and (f) orientation, in some space.
2. Time, which is subdivided into: (a) frequency of occurrence of repeated action, (b) time-at, time at which something occurred, (c) time-from, the time at which something started, (d) time-to, the time at which something ended, and (e) time-through, the time through which something occurred.
3. Causality, which consists of: (a) cause, (b) contradiction, what this action opposes if anything, (c) effect, and (d) purpose.
4. Quality, which is sub-divided into: (a) accompaniment, an additional object associated with the action, (b) content, what is contained by the action object, (c) manner, the way in which the action is done, (d) material, the atomic units out of which the action is composed, (e) measure, the measurement associated with the action, (f) order, with respect to other actions, and (g) value, the data transmitted by the action (the software sense of the term).
5. Essence, which consists of: (a) supertype, a generalization of the action type, and (b) whole, of which the action is a part.
6. Speech-act theory, which is sub-divided into: (a) an external precondition on the action and (b) an internal precondition on the ability of the agent to perform the action.
3.1.2 Persuasion
We summarize the persuasion taxonomy in Table 1. For this taxonomy, we adapt the SemEval 2023 Persuasion Task's categories (Da San Martino et al., 2023), and Cialdini's (Cialdini, 2006) persuasion principles, which are essentially persuasion techniques or strategies. The persuasion strategies taxonomy of Guerini et al. (2007) is orthogonal to this taxonomy since their definition of persuasion is broader than ours, but we do include their specific strategies under techniques.

Table 1. Persuasion taxonomy, adapted from Da San Martino et al. (2023), is a sub-taxonomy in the deception taxonomy.
The techniques used for each category are as follows (30 in total):
• Justification: appeal to popularity, appeal to authority/expert, appeal to values [or commitment (Cialdini, 2006)], appeal to fear/prejudice, reciprocity (Cialdini, 2006) [or goal balance (Guerini et al., 2007)], scarcity (Cialdini, 2006), reward, appeal to relevant empirical evidence, relevant statistics, and relevant examples.
• Simplification: causal oversimplification, false dilemma or no choice, and consequential oversimplification.
• Distraction: straw man, red herring (includes irrelevant empirical evidence, statistics or examples), whataboutism, flag waving, and liking (Cialdini, 2006).
• Call: slogans, social proof (Cialdini, 2006), appeal to time, and conversation killer.
• Manipulative wording/images: loaded language/images, repetition, exaggeration or minimization, and obfuscation—vagueness or confusion.
• Attack on reputation: name calling or labeling, doubt, guilt by association, appeal to hypocrisy, questioning the reputation.
To the best of our knowledge, we are the first to give detailed taxonomies for persuasion and strategem in this context and we are the first to use the following dimensions in a taxonomy of deception: target, persuasion, goal, dissemination, and timeliness. We add these to give a comprehensive view of deception, to aid in domain-independent deception detection, and to clarify and classify deception in all its different manifestations. Such a comprehensive taxonomy will provide a solid foundation on which to build automatic and semi-automatic detection methods and training programs for the targets of deception.
3.2 Examples
To demonstrate the applicability of this taxonomy, we give three examples. More discussion of stratagems and examples of cyber deception can be found in Rowe (2006).
Phishing is when attackers pretend to be from reputable companies to trick victims into revealing personal information. The agents are the attackers as initiators and the targets are the Internet/email users. The harmful goals include information or malware installation. Establishing facticity is difficult if the attacker is determined. The medium is the Internet/email. The breadth is high for phishing and narrower for spear phishing. The modality is text for phishing and audio for vishing. Images may also be used in phishing emails. The manner is non-interactive for phishing and interactive for vishing. Deliberate falsification and persuasion techniques such as authority, social proof, and reward or loss claims are employed in the stratagem.
Fake news is manufactured and misleading information presented as news. Here, the harmful goals include swaying opinion, sowing unrest, and division. The sources could be individuals, organizations, or nation-states. The breadth could vary depending on how deep-pocketed and determined the source(s) is (are). The modality could be text, audio, images, or video. The manner is asynchronous. Fake news could employ a range of techniques in the action component of the stratagem: from deliberate falsification to evasion and the persuasion component could include authority, social proof, etc.
Fake reviews are reviews designed to give consumers a false impression of a product or business. The harmful goal is to convince consumers to buy their product or avoid a competitor. The sources could be humans, bots, or their combinations. The targets are potential customers as well as the platform's fake review detector. The breadth is thus a broad range of people. While most fake reviews use only texts, deliberate attacks could be multi-modal, adding visuals and/or audio. Falsification and social proof are the main stratagems. Facticity and verifiability could vary depending on the stratagems used. The manner is asynchronous.
4 Related work
Deception has a vast social science literature. Hence, we focus on the most closely related work on computational deception, which can be categorized into taxonomies, datasets, detection, and literature reviews. Of the latter, we focus here on reviews of linguistic deception detection. The DBLP2 query “domain decepti”3 gave 43 matches of which 21 were deemed relevant.
Remark 1. Unfortunately, previous researchers have generally left the term “domain” undefined. In Glenski et al. (2020), different social networks, such as Twitter and Reddit, are referred to as domains. Hence, terms such as “cross-domain deception” in previous work could mean that the topics of essays or reviews are varied whereas the goals could stay pretty much the same.
4.1 Taxonomies
Whaley and Aykroyd (2007) gave a taxonomy of perception in which deception was defined succinctly as “other-induced misperception.” The full definition given in Whaley and Aykroyd (2007) is: “Any attempt—by words or actions—intended to distort another person's or group's perception of reality.” In Bell and Whaley (2017), two groups were introduced as essential for deception: simulation (overt, showing the false) and dissimulation (covert, hiding what is real). They introduced three simulation techniques: mimicking, inventing, and decoying, and three dissimulation techniques: masking, repackaging, and dazzling.
Dunnigan and Nofi (2001) gave a taxonomy of deception in the military context. This included concealment, camouflage, disinformation, lies, displays, ruses, demonstrations, feints, and insight.
The most comprehensive previous taxonomy of deception, to our knowledge, is proposed in Rowe (2006). It is inspired by linguistic case theory and includes 32 cases which are grouped into seven categories: space (six cases), time (five cases), participant (six cases), causality (four cases), quality (seven cases), essence (two cases), speech-act theory (two cases). Analyzing this taxonomy, we find that, except for the participant category, all the other categories fit neatly into the stratagems class for deception in our taxonomy.
More recently, a few researchers have proposed more specialized taxonomies for what they call defensive deception (Oluoha et al., 2021; Pawlick et al., 2019; Pawlick and Zhu, 2021). Some folksy and psychological taxonomies are given in Druckman and Bjork (1992).
4.2 Datasets
Several datasets have been collected for studying lies. However, researchers have not carefully delineated the scope by considering the goals of the deception. There is also another potentially more serious issue: Some datasets are constructed by asking participants to lie in a laboratory setting, where there are no consequences and no incentive to lie. We will refer to them as Lab Datasets. Others are constructed by collecting samples of real attacks. We call them Real-World Datasets. Finally, there are some datasets in which data from laboratory settings are combined with real-world attack samples. We call them Mixed Datasets.
Lab Datasets include Zhou et al. (2004), wherein students were paired and one student in each pair was asked to deceive the other using messages. In Perez-Rosas (2014), researchers collected demographic data and 14 short essays (seven truthful and seven false) on open-ended topics from 512 Amazon Mechanical Turkers (AMT). They tried to predict demographic information and facticity. We refer to this as the Open-Lies dataset. In Pérez-Rosas and Mihalcea (2014), researchers collected short essays on three topics: abortion, best friend, and the death penalty by people from four different cultural backgrounds. In Capuozzo et al. (2020), truthful and deceptive opinions on five topics are collected in two languages (English and Italian). See Ludwig et al. (2016) for more such efforts.
Next, we consider real-world datasets, where the goals may be information, disruption, financial, or psychological. Here, we have several datasets for fake news detection (Raponi et al., 2022),4 opinion spam (a.k.a. fake reviews) detection (Ren and Ji, 2019), phishing (Verma et al., 2019), and a company's reward program (Ludwig et al., 2016).
Some researchers have mixed data obtained from laboratory settings with non-laboratory data, such as reviews obtained from forums. For example, in Hernández-Castañeda et al. (2017), researchers analyzed three datasets: a two-class, balanced-ratio dataset of 236 Amazon reviews, a hotel opinion spam dataset consisting of 400 fabricated opinions from AMT plus 400 reviews from TripAdvisor (likely to be truthful), and 200 essays from Pérez-Rosas and Mihalcea (2014). In Xarhoulacos et al. (2021), researchers studied a masking technique on two datasets: a hotel, restaurant, and doctor opinion spam dataset and the dataset from Pérez-Rosas and Mihalcea (2014). In Cagnina and Rosso (2017), in-domain experiments were done with a positive and negative hotel opinion spam dataset, and cross-domain experiments were conducted with the hotel, restaurant, and doctor opinion spam dataset.
A few works have developed domain-independent deception datasets in our sense, wherein the goals of deception can be quite different. In Rill-García et al. (2018), researchers used two datasets: the American English subset consisting of a balanced-ratio 600 essays and transcriptions of 121 trial videos (60 truthful and 61 deceptive), which we call Real-Life_Trial. In Vogler and Pearl (2020), three datasets were used: positive and negative hotel reviews, essays on emotionally-charged topics, and personal interview questions. In Xarhoulacos et al. (2021), multiple fake news datasets, a COVID-19 dataset, and some micro-blogging datasets were collected and analyzed. In Shahriar et al. (2021), researchers collected fake news, Twitter rumors, and spam datasets. (Spam is essentially advertising. Deception is employed to fool automatic detectors rather than the human recipient of the spam. We focus on human targets.) They applied their models trained on these datasets to a new COVID-19 dataset. In Yeh and Ku (2021), seven datasets were collected (Diplomacy, Mafiascum, Open-Domain, LIAR, Box of Lies, MU3D, and Real-Life_Trial) and analyzed using LIWC categories, without claiming domain independence or cross-domain analysis. However, their datasets do involve different goals. LIAR, for instance, includes political lies with the goal of winning elections, whereas the lies in Real-Life_Trial have other goals, and Diplomacy/Mafiascum are about winning online games. In Feng et al. (2012), four datasets were collected: trip-advisor gold, a balanced hotel reviews dataset of 800 reviews introduced in Ott et al. (2011), trip-advisor heuristic, another balanced reviews dataset of 800 reviews collected by the authors, a third 800 review Yelp dataset of uncertain ground-truth collected by the authors, and the 296 essays on three topics dataset of Mihalcea and Strapparava (2009). They show that features based on CFG parse trees along with unigrams performed the best on these datasets.
Thus, we still lack large, comprehensive datasets for deception that have a wide variety of deceptive goals.
4.3 Detection
Deception detection in general is a useful and challenging open problem. There have been many attempts at specific applications such as phishing and fake news. On phishing alone (query: phish), there are 2,200+ DBLP results, including over 70 surveys and reviews. Similarly, there are 1,100+ papers on scams (query: scam, not all of them are relevant, since many occurrences are part of acronyms such as SCAMP), 100+ on opinion spam, 200+ on fake reviews, and 2,600+ on fake news.5
A soft domain transfer method is proposed in Shahriar et al. (2022). They found that partial training on tweets helped in phishing and fake news detection. In Panda (2022) and Panda and Levitan (2023), the authors study deception detection across languages and modalities. Other works on domain-independent deception detection have been discussed above under datasets.
4.4 Reviews on linguistic markers
Recently, Gröndahl and Asokan (2019) conducted a survey of the literature on deception. They defined implicit and explicit deception, focused on automatic deception detection using input texts, and then proceeded to review 17 papers on linguistic deception detection techniques (explicit deception is when the deceiver explicitly mentions the false proposition in the deceptive communication). These papers covered two forms of deception: (a) dyadic pairs in the laboratory, where one person sends a short essay or message to another (some truthful and some lies), and (b) fake reviews (a.k.a. opinion spam). Based on their analysis of the literature on laboratory deception experiments and the literature on opinion spam, they concluded that there is no linguistic or stylistic trace that works for deception in general. Similarly, the authors of Vogler and Pearl (2020) assert that extensive psychology research shows that “a generalized linguistic cue to deception is unlikely to exist.” We collectively refer to Gröndahl and Asokan (2019), Fitzpatrick et al. (2015), and Vogler and Pearl (2020); Vrij (2008) as the Critiques.
As opposed to the critiques, the meta-analyses by DePaulo et al. (2003) and Hauch (2016) did find small markers of deception in the studies they examined despite analyzing studies of specific forms or situations of deception, not general domain-independent datasets. Similarly, the following papers all point to evidence for cross-domain deception detection: Rill-García et al. (2018), Shahriar et al. (2021), Vogler and Pearl (2020), Xarhoulacos et al. (2021), and Yeh and Ku (2021). These researchers created so-called “domain-independent datasets,” which consist of two or three kinds of attacks and developed features and techniques for deception detection across the collected domains.
We believe that a deeper investigation/analysis of the linguistic cues for deception debate is needed, for the simple reason that none of the above works created a comprehensive dataset of different forms of deceptive attacks and analyzed it.
5 Linguistic cues/analysis
Because of the problems enumerated above, we collect and analyze datasets for domain-independent linguistic cues to tackle: (1) the ground truth problem for deception detection and (2) evidence of linguistic cues for deception across domains.
A ground truth is something that is known to be correct, but this information is difficult to obtain, so we need models that do not rely on having too much ground truth data. Our approach is to focus on using linguistic information from the text. For the second challenge, we try to find universal linguistic markers for deception by looking for features that behave similarly across domains. We hope that an ML model built with these features could generalize across domains (Gokhman et al., 2012).
5.1 Datasets
We summarize our deception domains and scenarios below. We focus on real-world datasets.
In the product review domain, we use the Amazon reviews dataset mentioned above (Garcia, 2019).
In the job scam domain, we identify fraudulent job listings. Our dataset contains the bodies of 13,735 legitimate and 608 fraudulent job listings.
In the phishing domain, we distinguish between legitimate emails and phishing emails. Our dataset contains the bodies of 9,202 legitimate and 6,134 phishing samples. The IWSPA-AP dataset analyzed above is a subset of this dataset.
In the political statement domain, we determine the truthfulness of claims made by US political speakers. Our dataset contains 7,167 truthful and 5,669 deceptive statements evaluated by PolitiFact.
In the fake news scenario, we distinguish between legitimate and fake news. Here, we use the WELFake dataset (Verma et al., 2021).
We analyzed each dataset for any artifacts of data collection and cleaned them to remove such artifacts. The cleaning procedures include two parts: text removal and text cleaning. We then sanitize the texts using the methods discussed in Zeng et al. (2022). We remove meta-data in emails and source leaks in news and replace HTML break tags with new lines. In addition, the authors of Zeng et al. (2022) found that the provided labels in WELFake Verma et al., (2021) are flipped, so we flip its labels as a final cleaning step. We are making the combined, cleaned dataset available on Zenodo.6
5.2 Sources for linguistic cues
Function words (FW) are words that express a grammatical relationship between words in a sentence. Unlike content words, function words such as “when,” “at,” and “the” are independent of specific domains. Function words and n-grams are useful for many text classification tasks, including author gender classification, authorship attribution (Argamon and Levitan, 2005), and deception detection (Siagian and Aritsugi, 2020). To gain an insight into the transfer of knowledge between domains, we utilized three types of explainable features: function words, part-of-speech (POS) tags of function words, and engineered linguistic features. POS tags were used to determine whether a word was a function or a content word; the content words were then removed. The last experiment utilized 151 engineered linguistic features (13+55+86 − 3 duplicates removed by the colinearity check below).
The engineered features are drawn from three sources. Linguistic Inquiry and Word Count (Boyd et al., 2022), a popular source of features in the NLP literature, was the source of 86 features. The authorship attribution paper (Fabien et al., 2020) was the source of 55 features. Thirteen features were collected from two papers, one on deception (Zhou et al., 2004) and the other on fake news (Verma et al., 2021), after significance testing using t-tests with and without the Bonferroni-Holm correction of p-values.
The initial significance testing of 27 linguistic features from the two papers (Zhou et al., 2004; Verma et al., 2021) on four public datasets is described in Appendix A. Appendix B describes an analysis of function word n-grams on the same datasets as in Appendix A. A complete source-wise list of the 55 features from Fabien et al. (2020) and 86 features from Boyd et al. (2022) is in Appendix C. Function words as features for deception have been studied before, in Siagian and Aritsugi (2020), for example. We also experimented with the part-of-speech tags of function words.
5.3 Results of feature analysis
We used the Stanza (Qi et al., 2020) POS tagger and OntoNotes Release 5.0/Penn Treebank (Marcus et al., 1993) tagset in all experiments involving POS tags. This tagset builds on top of the original Penn Treebank and adds seven new tags:
ADD—Email, AFX—Affix, HYPH—Hyphen, NFP—Superfluous punctuation, UH—Interjection, SP—Space, and XX—Unknown.
Due to the parser's limitations, several samples of text that had a length more than one million characters had to be discarded. We did not remove stop words or further alter the data in any manner. Function words and their respective POS tags were separately vectorized as word unigrams using the tf-idf scheme. The raw texts were processed and vectorized identically and used as a baseline. The motivation behind it was to (i) understand whether it is possible to achieve similar results while using only a few non-domain-specific features that are highly indicative of deception and (ii) investigate the impact of content words on deception through the contrast between the baseline and function words.
For each dataset, and for each set of features, we applied three techniques to select the most relevant features. First, a random forest algorithm (Breiman, 2001) was used, which allowed us to rank features by their importance. The least important ones were removed under the condition that the out-of-bag accuracy on the validation set either increased or remained the same after removing the features. Next, we applied scipy's (Virtanen et al., 2020) single linkage hierarchical clustering (Gower and Ross, 1969) with Spearman's correlation (Spearman, 1904) as the measure of feature colinearity. Features exhibiting a high degree of colinearity were removed with their redundancy validated in the same manner as with the first technique. Finally, taking the remaining features, we applied Hyperopt's (Bergstra et al., 2013) feature selection and the eXtreme Gradient Boosting algorithm (Chen and Guestrin, 2016) with SHAP (Lundberg and Lee, 2017) as a metric of each feature's contribution to the overall model performance. Ultimately, the aforementioned approach produced a subset of the features for each of the five datasets. A total of 81 linguistic, 28 function word POS, and 61 function word features were selected; 50/81, 22/28, and 29/61 were shared with at least one other dataset.
For our analysis of the potential for knowledge transfer, any feature unique to a dataset was removed, leaving only those significant for at least two datasets and therefore being of interest for understanding of transfer between domains. The relationships of function words, function words' POS tags, and engineered linguistic features across datasets are depicted in Table 2. Several trends can be noticed from this table. For example, all five datasets share 6+10+1 = 17 common features, and the fake news, job scams, and phishing datasets have a total of 31 features in common. In addition, the subset {F, J} has 35 common features, and {J, P, Pr, Ps} has 26 common features. Job scams and phishing together have 43 common features. Similarly, we see that deceptive attacks can be differentiated using features such as “to,” personal pronouns, singular present verb forms, modals, and adverbs (compare with the quote from Rowe, 2006 in Section 3.1.1). The richness, possessive ending, and interjection features are significant for fake news, job scams, and phishing. Fake news and product reviews have many significant LIWC features.
5.3.1 Linguistic overlap across deceptive domains
The observed overlap in common features across domains suggests a shared linguistic substrate of deceptive or persuasive communication. Many of the common function words (e.g., you, this, are, and that) are tied to reader-directed or modal constructions, which have been found to correlate with manipulative intent (Zhou et al., 2004; DePaulo et al., 2003).
Part-of-speech tags such as PRP, VB, TO, and IN reflect structural scaffolding typical in persuasive or fraudulent writing. These tags often co-occur in imperative or passive constructions that are used to command attention or obscure agency (Ott et al., 2011). For instance, phishing emails and fake job offers rely on templates such as “to confirm your account...” or “you are selected...”, which map to these tags.
Engineered features such as avg_len, sen_len, and focuspast point to reduced syntactic complexity and temporal distancing, both recognized as cues in deceptive text (Hancock et al., 2007). Shorter messages and generic phrasing enable broad applicability and reduce the chance of contradiction.
Clusters such as {F, J, P} tend to involve transactional deception (e.g., scams), while overlaps in {Pr, Ps, P} suggest persuasive manipulation.
Datasets that share a significant number of features are good candidates for domain adaptation; however, the performance of a model using a potentially limited set of features shared across tasks must remain robust. To this end, we combined previously selected linguistic, function words, and function word POS features that were shared by two or more datasets. This resulted in a final set of 91 features. Upon further applying feature selection, the number of significant features of all three types shared among datasets has been reduced to 45.
To evaluate the features' performance, we used a random forest classifier with five-fold cross-validation. The model hyperparameters were set to 50 trees with the leaf nodes of five samples, and 50% of the features were considered on each split. Gini impurity was used as a criterion of split quality.
The accuracy and F1-scores of the model using each of the feature sets across the five datasets are shown in Figures 2, 3, respectively. It is important to note that Job Scams' data appear to be heavily imbalanced and the models' performance on it is not an ideal indicator of feature quality. Generally, the combined set of shared features is nearly on par with the baseline, with linguistic, function word, and function word POS following in the order given. Notable exceptions are Product Reviews where linguistic and combined features beat the others, including the baseline, and Fake News with linguistic features outperforming the rest by a significant margin. We hypothesize that the relative length and richness of news articles may be in part responsible for this phenomenon.

Figure 2. Random Forest F1 scores for the five feature types: linguistic (ling), function words (fw), pos tags of function words (fw-pos), combination of the three (all), and unigram tf-idf (baseline); F, fake news; J, job scams; P, phishing; Pr, product reviews; Ps, political statements.

Figure 3. Random Forest accuracies for the five feature types: linguistic (ling), function words (fw), pos tags of function words (fw-pos), combination of the three (all), and unigram tf-idf (baseline); F, fake news; J, job scams; P, phishing; Pr, product reviews; Ps, political statements.
6 Deep-learning based experiments
To investigate the possible existence of other deception signals, we turn to deep learning. If universal deception signals exist, then a deep-learning model can learn to detect them. To determine whether this happens, we perform two experiments on the same five cleaned datasets of the previous section. First, we evaluate the performance of models trained on multiple domains. Then, we train models on one domain and evaluate their performance on other domains.
6.1 Model
Our model architecture consists of a base pre-trained transformer model, a dropout layer, and a linear layer. As standard in NLP, we prepend a [CLS] token to the text, pass the text through the base model, and perform classification on the last-layer embedding of the [CLS] token.
6.2 Multi-domain experiment
If deep-learning models trained on multiple domains pick up on universal deception signals, then we should expect performance on individual domains to be positively correlated among each other. Conversely, if they only learn domain-specific signals, then we should expect performance on individual domains to be negatively correlated with one another.
We train 100 models on the union of our datasets. We use a random 80/10/10 train/validate/test split for each dataset with uniformly drawn hyperparameters. We use BERT-base and RoBERTa-base for our base models, dropout percentages between 0.1 and 0.5, and the AdamW optimizer with learning rates between 0.00001 and 0.0001.
We then evaluate each model on the individual test sets. We exclude models that failed to converge and models that have an outlier F1 score using the IQR test and perform pairwise linear regression on the remaining F1 scores.
We present our results without outliers in Figure 4. All pairs of tasks except for product reviews and phishing are positively correlated, with five of them significant at the 0.05 level.
6.3 Cross-domain generalization experiment
If a deep-learning model primarily learns a universal deception signal, then it should generalize to deception domains that it has not yet seen. In particular, they should be able to achieve a higher F1 score than a coin flip classifier, which we can calculate using the formula CF F1 = q/(0.5+q), where q is the portion of the dataset that is deceptive.
On each dataset, we train 100 models with hyperparameters drawn from uniform distributions. We use BERT-base and RoBERTa-base for our base models and values between 0.1 and 0.5 for dropout percentage. For our learning rate, we use a different range for each task to minimize divergence: [0.00001, 0.00006] for product reviews, [0.00001, 0.000025] for job scams, [0.00001, 0.00010] for phishing, [0.00001, 0.00004] for political statements, and [0.00001, 0.00010] for fake news.
Each model is evaluated on each dataset, ignoring models that fail to converge. We perform a 1-sample t-test with the alternative hypothesis “the mean F1 in domain Y of models trained on X is less than or equal to the coin flip F1 of Y.” We report the resulting p-values in Table 3. In ten cases, models trained on one domain manage to beat the coin-flip baseline at a 0.01 significance level, with nine cases beating the coin-flip baseline at the 10σ (p < 7.62 × 10−24) level. However, we also find that eight pairs have a p-value of 1.00, meaning that they performed worse than the coin-flip baseline.7
Interestingly, we also find that the fake news models manage to beat the coin flip on all domains. We suspect that this is due in part to its larger size but leave this as a direction for future research.
6.4 Discussion
The multi-domain experiment provides strong support for the existence of universal deception signals. All but one pair are positively correlated. Five are statistically significant, and the one negative correlation is not statistically significant. In contrast, the results of our cross-domain generalization experiment are mixed. While some pairs beat the coin-flip baseline, others performed worse than the baseline.
Taken together, these results suggest that both universal and domain-specific deception signals exist. Models trained on a single task will learn both universal and task-specific signals, potentially resulting in poor generalization to other deception domains. Therefore, training a domain-independent deception detector will likely require a diverse domain-independent dataset.
7 Cross-domain detection
We also built detectors for deception and conducted the following experiments (Zeng et al., 2022).
7.1 Single-task baselines
To evaluate the diversity and difficulty of our collected tasks, we fine-tuned BERT-base (Devlin et al., 2018) classification models on each of our datasets. As standard in NLP, our model adds a linear layer that generates the prediction from the final classification token embedding. We trained using the AdamW optimizer with a learning rate of 2 × 10−5, a batch size of 16, dropout of 0.1, and a random 80/10/10 train-val-test split. We trained for five epochs with early stopping on the validation set on NVIDIA V100 GPUs with automatic mixed precision. We then evaluated our models against all datasets using a TPU from Google Colab.8 If a dataset was used to train the model, we used the held-out test set. Otherwise, we evaluated against the full dataset.
We report the accuracies and F1-scores measured in Tables 4, 5. BERT performed well on the phishing, job scams, and fake news tasks, with F1 scores greater than 0.98. However, it performed poorly on product reviews and political statements, with F1 scores of 0.594 and 0.708, respectively. We suspect that this is due in part to the lengths of the texts; many product reviews and almost all political statements are only one or two sentences long. Interestingly, we find that product reviews and fake news transfer well to the Job Scams task, achieving F1 scores greater than 0.950, but not vice versa.
7.2 Models for the combined dataset
We fine-tuned four models on the union of all our datasets using the same method, with an 80/10/10 train-val-test split for each individual dataset, and hyperparameters as our individual models. For our base models, we used BERT-base and RoBERTa (Liu et al., 2019) (110 million parameters), and the larger BERT-large and RoBERTa-large (340 million parameters) pre-trained models.
Surprisingly, the small base models performed better than the large models, with RoBERTa performing slightly better. BERT-base and RoBERTa-base achieved 0.904 and 0.904 F1 scores (the slight gap disappears on rounding), respectively, while their large counterparts achieved F1 scores of 0.882 and 0.900. However, when we break down performance by task (Tables 6, 7), we find that BERT-base performed better, achieving the highest F1 score in 3/5 tasks and was still close in performance for the other two tasks (to the winners in the individual task experiment and to the winners on combined).
7.3 Discussion
Our results show that a single model can recognize multiple forms of detection. For example, a BERT model has high accuracy/F1 scores on 3 out of 5 tasks and is still close to individual models on the other two. Another interesting result is that BERT and BERT(L) trained on the combined dataset beats in accuracy and F1 the individual BERT trained and tested on job scams dataset. RoBERTa could do it only for F1 on the Politics dataset.
8 Conclusion and future work
We have provided new definitions for deception based on explanations and probability theory. We gave a new taxonomy of deception that clarifies the explicit and implicit elements of deception.
We have argued against hasty conclusions regarding linguistic cues for deception detection and especially their generalizability. The critiques contained in Fitzpatrick et al. (2015), Vogler and Pearl (2020), and Vrij (2008) may present a valid point, namely that some linguistic cues might not generalize across the broad class of attacks. However, over-generalizations should be made with caution as they discourage future domain-independent deception research. Moreover, we have presented evidence showing that there do exist common linguistic cues in deceptive attacks with widely varying goals and topical content.
Our linguistic analysis of four datasets and cross-dataset analysis of five different deception datasets shows that there are linguistic features, some at the surface level and some deeper, that can be used to build classifiers for more general deception datasets. With all the new developments in machine learning and NLP, we believe that research on linguistic deception detection is poised to take off and could result in significant advances.
We propose three concrete directions for future work: (a) investigation of the domain pairs that underperform in cross-domain detection, (b) comparing our BERT/RoBERTa results with the latest BERT variants, e.g., the nBERT model (Rasool et al., 2025), and (c) exploring multimodal datasets that integrate different modalities, e.g., text, images, audio and video, and different domains of deception.
Author's note
This is a thoroughly revised version of a 2023 arXiv draft containing substantial new material. Cross-domain detection results in this paper were presented in a poster at ACM CODASPY 2022.
Data availability statement
Publicly available datasets were analyzed in this study. This data can be found here: https://zenodo.org/records/8371762.
Author contributions
RV: Conceptualization, Formal analysis, Funding acquisition, Investigation, Methodology, Project administration, Supervision, Writing – original draft, Writing – review & editing. ND: Conceptualization, Supervision, Writing – original draft, Writing – review & editing. VZ: Data curation, Investigation, Methodology, Software, Validation, Visualization, Writing – review & editing. DB: Data curation, Investigation, Methodology, Software, Validation, Writing – review & editing. XL: Data curation, Investigation, Methodology, Software, Validation, Writing – review & editing.
Funding
The author(s) declare that financial support was received for the research and/or publication of this article. Rakesh M. Verma's research was partially supported by NSF grants 1433817, 1950297, 2210198, and 2244279, ARO grants W911NF-20-1-0254, W911NF-23-1-0191, and ONR award N00014-19-S-F009. Dainis Boumber's research was partly supported by ARO award W911NF-20-1-0254. Victor Zeng's research was supported by ONR award N00014-19-S-F009. Xuting Liu's research was supported by NSF award 1950297.
Acknowledgments
The authors are grateful to the reviewers for their constructive suggestions. We thank all those who supplied datasets for this research and Vu Minh Hoang Dang for his comments on a previous draft of this article.
Conflict of interest
RV is the founder of Everest Cyber Security and Analytics, Inc.
The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
The author(s) declared that they were an editorial board member of Frontiers, at the time of submission. This had no impact on the peer review process and the final decision.
Generative AI statement
The author(s) declare that no Gen AI was used in the creation of this manuscript.
Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.
Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary material
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fdata.2025.1581734/full#supplementary-material
Footnotes
1. ^https://www.surveymonkey.com/curiosity/axios-january-6-revisited
3. ^Searched on 17 February 2025.
4. ^Note that the topics can vary in a heterogeneous application, such as fake news detection, since some items could be on sport and some on politics or religion. Moreover, the goals may or may not be different. Hence, we avoid the term “domain” to refer to applications such as fake news.
5. ^All these DBLP search results are as of 17 February 2025.
6. ^At https://zenodo.org/record/6512468#.ZBVRUhTMLQM.
7. ^This negative transfer plus domain shift can be mitigated by a few modifications that include: (a) masking domain-specific tokens, (b) adapter-based DANN training (Ganin et al., 2016), (c) calibrated thresholds, and (d) reducing source dataset size for transfer (Triplett et al., 2025).
8. ^In our experiments, we noticed slight differences in outputted predictions evaluating on TPUs vs. on GPUs due to differences in floating-point representations, but the differences were not statistically significant.
References
Argamon, S., and Levitan, S. (2005). “Measuring the usefulness of function words for authorship attribution,” in Proceedings of the 17th Joint International Conference on Humanities Computing and Digital Scholarship (Bryan: Association for Computer Humanities), 4–6.
Bansal, S., and Aggarwal, C. (2019). TextSTAT. Available onine at: https://pypi.org/project/textstat (Accessed July 31, 2023).
Bell, J. B., and Whaley, B. (2017). Cheating and Deception. London: Routledge. doi: 10.4324/9781315081496
Bergstra, J., Yamins, D., and Cox, D. D. (2013). “Hyperopt: a Python library for optimizing the hyperparameters of machine learning algorithms,” in Proceedings of the 12th Python in Science Conference (SciPy 2013), Volume 13 (San Francisco, CA: Curvenote), 13–19. doi: 10.25080/Majora-8b375195-003
Bogert, J. (1985). In defense of the Fog index. Bull. Assoc. Bus. Commun. 48, 9–12. doi: 10.1177/108056998504800203
Boyd, R. L., Ashokkumar, A., Seraj, S., and Pennebaker, J. W. (2022). The Development and Psychometric Properties of LIWC-22. Technical report. Austin, TX: University of Texas at Austin.
Cagnina, L. C., and Rosso, P. (2017). Detecting deceptive opinions: Intra and cross-domain classification using an efficient representation. Int. J. Uncertain. Fuzziness Knowl-Based Syst. 25(Suppl. 2), 151–174. doi: 10.1142/S0218488517400165
Capuozzo, P., Lauriola, I., Strapparava, C., Aiolli, F., and Sartori, G. (2020). “DecOp: a multilingual and multi-domain corpus for detecting deception in typed text,” in Proceedings of the 12th Language Resources and Evaluation Conference (Marseille: European Language Resources Association), 1423–1430.
Chen, T., and Guestrin, C. (2016). “XGBoost: a scalable tree boosting system,” in Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD '16 (New York, NY: ACM), 785–794. doi: 10.1145/2939672.2939785
Cialdini, R. B. (2006). Influence: The Psychology of Persuasion, Revised Edition. New York, NY: William Morrow.
Da San Martino, G., Nakov, P., Piskorski, J., and Stefanovitch, N. (2023). SemEval 2023 task 3: Detecting the Genre, the Framing, and the Persuasion Techniques in Online News in a Multi-lingual Setup. Available online at: https://propaganda.math.unipd.it/ (Accessed November 22, 2023).
Das, A., Baki, S., Aassal, A. E., Verma, R. M., and Dunbar, A. (2020). SoK: a comprehensive reexamination of phishing research from the security perspective. IEEE Commun. Surv. Tutor. 22, 671–708. doi: 10.1109/COMST.2019.2957750
DePaulo, B. M., Lindsay, J. J., Malone, B. E., Muhlenbruck, L., Charlton, K., Cooper, H., et al. (2003). Cues to deception. Psychol. Bull. 129, 74–118. doi: 10.1037/0033-2909.129.1.74
Devlin, J., Chang, M., Lee, K., and Toutanova, K. (2018). BERT: pre-training of deep bidirectional transformers for language understanding. arXiv [prprint]. arXiv:1810.04805. doi: 10.48550/arXiv.1810.04805
Druckman, D., and Bjork, R. A. (Eds.) (1992). In the Mind's Eye: Enhancing Human Performance. Washington, DC: National Academy Press. National Research Council (U.S.). Committee on Techniques for the Enhancement of Human Performance.
Dunnigan, J. F., and Nofi, A. A. (2001). Victory and Deceit: Deception and Trickery at War. New York, NY: Writers Club Press.
Fabien, M., Villatoro-Tello, E., Motlicek, P., and Parida, S. (2020). “Bertaa: Bert fine-tuning for authorship attribution,” in Proceedings of the 17th International Conference on Natural Language Processing (ICON) (NLP Association of India), 127–137.
Feng, S., Banerjee, R., and Choi, Y. (2012). “Syntactic stylometry for deception detection,” in Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers) (Jeju Island: ACL), 171–175.
Fillmore, C. J. (1968). “The case for case,” in Universals in Linguistic Theory. New York, NY: Holt, Rinehart & Winston.
Fitzpatrick, E., Bachenko, J., and Fornaciari, T. (2015). Automatic Detection of Verbal Deception. Synthesis Lectures on Human Language Technologies. San Rafael, CA: Morgan & Claypool Publishers. doi: 10.1007/978-3-031-02158-9
Galasinski, D. (2000). The Language of Deception: A Discourse Analytical Study. London: Sage Publications. doi: 10.4135/9781452220345
Ganin, Y., Ustinova, E., Ajakan, H., Germain, P., Larochelle, H., Laviolette, F., et al. (2016). Domain-adversarial training of neural networks. JMLR, 17, 1–35.
Garcia, L. (2019). Amazon-Reviews-Dataset. Available online at: https://www.kaggle.com/lievgarcia/amazon-reviews (Accessed August 2, 2023).
Gelman, A., Hill, J., and Yajima, M. (2012). Why we (usually) don't have to worry about multiple comparisons. J. Res. Educ. Eff. 5, 189–211. doi: 10.1080/19345747.2011.618213
Glenski, M., Ayton, E., Cosbey, R., Arendt, D., and Volkova, S. (2020). “Towards trustworthy deception detection: benchmarking model robustness across domains, modalities, and languages,” in Proceedings of the 3rd International Workshop on Rumours and Deception in Social Media (RDSM) (Barcelona: ACL), 1–13.
Gokhman, S., Hancock, J., Prabhu, P., Ott, M., and Cardie, C. (2012). “In search of a gold standard in studies of deception,” in Proceedings of the Workshop on Computational Approaches to Deception Detection (Avignon: Association for Computational Linguistics), 23–30.
Gower, J. C., and Ross, G. J. S. (1969). Minimum spanning trees and single linkage cluster analysis. J. R. Stat. Soc. Ser. C 18, 54–64. doi: 10.2307/2346439
Gröndahl, T., and Asokan, N. (2019). Text analysis in adversarial settings: does deception leave a stylistic trace? ACM Comput. Surv. 52, 1–36. doi: 10.1145/3310331
Guerini, M., Stock, O., and Zancanaro, M. (2007). A taxonomy of strategies for multimodal persuasive message generation. Appl. Artif. Intell. 21, 99–136. doi: 10.1080/08839510601117169
Hancock, J. T., Curry, L. E., Goorha, S., and Woodworth, M. (2007). On lying and being lied to: a linguistic analysis of deception in computer-mediated communication. Discourse Processes 45, 1–23. doi: 10.1080/01638530701739181
Hauch, V. (2016). Meta-Analyses on the Detection of Deception with Linguistic and Verbal Content Cues [PhD thesis]. Giessen: Justus-Liebig-Universität Gießen.
Hernández-Casta neda, Á., Calvo, H., Gelbukh, A. F., and Flores, J. J. G. (2017). Cross-domain deception detection using support vector networks. Soft Comput. 21, 585–595. doi: 10.1007/s00500-016-2409-2
Honnibal, M. (2015). Introducing spaCy. Available online at: https://explosion.ai/blog/introducing-spacy (Accessed June 17, 2022).
Kapantai, E., Christopoulou, A., Berberidis, C., and Peristeras, V. (2021). A systematic literature review on disinformation: toward a unified taxonomical framework. New Media Soc. 23, 1301–1326. doi: 10.1177/1461444820959296
Levine, T. R. (2014). Encyclopedia of Deception, Volume 2. London: Sage Publications. doi: 10.4135/9781483306902
Liu, Y., Ott, M., Goyal, N., Du, J., Joshi, M., Chen, D., et al. (2019). Roberta: a robustly optimized BERT pretraining approach. arXiv [preprint]. arXiv:1907.11692. doi: 10.48550/arXiv.1907.11692
Loper, E., and Bird, S. (2002). NLTK: the natural language toolkit. arXiv [preprint]. arXiv:cs/0205028. doi: 10.48550/arXiv.cs/0205028
Ludwig, S., Van Laer, T., De Ruyter, K., and Friedman, M. (2016). Untangling a web of lies: Exploring automated detection of deception in computer-mediated communication. J. Manag. Inf. Syst. 33, 511–541. doi: 10.1080/07421222.2016.1205927
Lundberg, S. M., and Lee, S.-I. (2017). “A unified approach to interpreting model predictions,” in Advances in Advances in Neural Information Processing Systems 30, eds. I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, et al. (Red Hook, NY: Curran Associates, Inc), 4765–4774.
Marcus, M. P., Marcinkiewicz, M. A., and Santorini, B. (1993). Building a large annotated corpus of english: the penn treebank. Comput. Linguist. 19, 313–330. doi: 10.21236/ADA273556
Mihalcea, R., and Strapparava, C. (2009). “The lie detector: explorations in the automatic recognition of deceptive language,” in Proceedings of the ACL-IJCNLP 2009 Conference: Short Papers (ACL: Singapore), 309–312. doi: 10.3115/1667583.1667679
Molina, M. D., Sundar, S. S., Le, T., and Lee, D. (2021). “Fake news” is not simply false information: a concept explication and taxonomy of online content. Am. Behav. Sci. 65, 180–212. doi: 10.1177/0002764219878224
Oluoha, O. U., Yange, T. S., Okereke, G. E., and Bakpo, F. S. (2021). Cutting edge trends in deception based intrusion detection systems-a survey. J. Inf. Secur. 12, 250–269. doi: 10.4236/jis.2021.124014
Ott, M., Choi, Y., Cardie, C., and Hancock, J. T. (2011). Finding deceptive opinion spam by any stretch of the imagination. arXiv [preprint]. doi: 10.48550/arXiv.1107.4557
Panda, S. (2022). Deception Detection Across Domains, Languages and Modalities [PhD thesis]. City University of New York, New York, NY.
Panda, S., and Levitan, S. I. (2023). Deception detection within and across domains: identifying and understanding the performance gap. ACM J. Data Inf. Qual. 15, 7:1–7:27 doi: 10.1145/3561413
Pawlick, J., Colbert, E., and Zhu, Q. (2019). A game-theoretic taxonomy and survey of defensive deception for cybersecurity and privacy. ACM Comput. Surv. 52, 1–28. doi: 10.1145/3337772
Pawlick, J., and Zhu, Q. (2021). Game Theory for Cyber Deception. Cham: Springer. doi: 10.1007/978-3-030-66065-9
Perez-Rosas, V. (2014). Exploration of Visual, Acoustic, and Physiological Modalities to Complement Linguistic Representations for Sentiment Analysis [PhD thesis]. University of North Texas, Denton, TX.
Pérez-Rosas, V., and Mihalcea, R. (2014). “Cross-cultural deception detection,” in Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers) (Baltimore, MD: ACL), 440–445. doi: 10.3115/v1/P14-2072
Qi, P., Zhang, Y., Zhang, Y., Bolton, J., and Manning, C. D. (2020). “Stanza: a Python natural language processing toolkit for many human languages,” in Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics: System Demonstrations (ACL), 101–108. doi: 10.18653/v1/2020.acl-demos.14
Raponi, S., Khalifa, Z., Oligeri, G., and Di Pietro, R. (2022). Fake news propagation: a review of epidemic models, datasets, and insights. ACM Trans. Web 16, 1–34. doi: 10.1145/3522756
Rasool, A., Aslam, S., Hussain, N., Imtiaz, S., and Riaz, W. (2025). nBERT: harnessing NLP for emotion recognition in psychotherapy to transform mental health care. Information 16:301. doi: 10.3390/info16040301
Ren, Y., and Ji, D. (2019). Learning to detect deceptive opinion spam: a survey. IEEE Access 7, 42934–42945. doi: 10.1109/ACCESS.2019.2908495
Rill-García, R., Pineda, L. V., Reyes-Meza, V., and Escalante, H. J. (2018). “From text to speech: a multimodal cross-domain approach for deception detection,” in Pattern Recognition and Information Forensics - ICPR 2018 International Workshops, CVAUI, IWCF, and MIPPSNA, Beijing, China, August 20-24, 2018, Revised Selected Papers (Cham: Springer), 164–177. doi: 10.1007/978-3-030-05792-3_16
Rowe, N. (2006). “A taxonomy of deception in cyberspace,” in International Conference on Information Warfare and Security (Princess Anne, MD), 173–181.
Shahriar, S., Mukherjee, A., and Gnawali, O. (2021). “A domain-independent holistic approach to deception detection,” in Proceedings of Recent Advances in Natural Language Processing (RANLP) (INCOMA Ltd.), 1308–1317. doi: 10.26615/978-954-452-072-4_147
Shahriar, S., Mukherjee, A., and Gnawali, O. (2022). “Deception detection with?feature-augmentation by?soft domain transfer,” in Social Informatics, eds. F. Hopfgartner, K. Jaidka, P. Mayr, J. Jose, and J. Breitsohl (Cham: Springer International Publishing), 373–380. doi: 10.1007/978-3-031-19097-1_23
Sharma, K., Qian, F., Jiang, H., Ruchansky, N., Zhang, M., Liu, Y., et al. (2019). Combating fake news: a survey on identification and mitigation techniques. ACM Trans. Intell. Syst. Technol. 10, 1–42. doi: 10.1145/3305260
Siagian, A. H. A. M., and Aritsugi, M. (2020). Robustness of word and character n-gram combinations in detecting deceptive and truthful opinions. J. Data Inf. Qual. 12, 1–24. doi: 10.1145/3349536
Spearman, C. (1904). The proof and measurement of association between two things. Am. J. Psychol. 15, 72–101. doi: 10.2307/1412159
Tausczik, Y. R., and Pennebaker, J. W. (2010). The psychological meaning of words: LIWC and computerized text analysis methods. J. Lang. Soc. Psychol. 29, 24–54. doi: 10.1177/0261927X09351676
Triplett, S., Minami, S., and Verma, R. M. (2025). “Effects of soft-domain transfer and named entity information on deception detection,” in International Conference on Information Systems Security (Cham: Springer), 146–155. doi: 10.1007/978-3-031-80020-7_8
Verma, P. K., Agrawal, P., Amorim, I., and Prodan, R. (2021). WELFake: word embedding over linguistic features for fake news detection. IEEE Trans. Comput. Soc. Syst. 8, 881–893. doi: 10.1109/TCSS.2021.3068519
Verma, R. M., Zeng, V., and Faridi, H. (2019). “Data quality for security challenges: case studies of phishing, malware and intrusion detection datasets,” in Proceedings of the 2019 ACM SIGSAC Conference on Computer and Communications Security, CCS 2019, London, UK, November 11-15, 2019, eds. L. Cavallaro, J. Kinder, X. Wang, and J. Katz (New York, NY: ACM), 2605–2607. doi: 10.1145/3319535.3363267
Virtanen, P., Gommers, R., Oliphant, T. E., Haberland, M., Reddy, T., Cournapeau, D., et al. (2020). SciPy 1.0: fundamental algorithms for scientific computing in Python. Nat. Methods 17, 261–272. doi: 10.1038/s41592-020-0772-5
Vogler, N., and Pearl, L. (2020). Using linguistically defined specific details to detect deception across domains. Nat. Lang. Eng. 26, 349–373. doi: 10.1017/S1351324919000408
Vrij, A. (2008). Detecting Lies and Deceit: Pitfalls and Opportunities. Hoboken, NJ: John Wiley & Sons.
Whaley, B., and Aykroyd, S. S. (2007). Textbook of Political-Military Counterdeception: Basic Principles & Methods. Washington, DC: National Defense Intelligence College.
Xarhoulacos, C., Anagnostopoulou, A., Stergiopoulos, G., and Gritzalis, D. (2021). Misinformation vs. situational awareness: the art of deception and the need for cross-domain detection. Sensors 21:5496. doi: 10.3390/s21165496
Yeh, M.-H., and Ku, L.-W. (2021). “Lying through one's teeth: a study on verbal leakage cues,” in Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing (Punta Cana: ACL), 4504–4510. doi: 10.18653/v1/2021.emnlp-main.370
Zeng, V., Liu, X., and Verma, R. M. (2022). “Does deception leave a content independent stylistic trace?” in Proceedings of the Twelfth ACM Conference on Data and Application Security and Privacy (CODASPY '22) (New York, NY: Association for Computing Machinery), 349–351. doi: 10.1145/3508398.3519358
Zeng, V., Zhou, X., Baki, S., and Verma, R. M. (2020). “PhishBench 2.0: a versatile and extendable benchmarking framework for phishin,” in Proceedings of the 2020 ACM SIGSAC Conference on Computer and Communications Security, CCS '20 (New York, NY: Association for Computing Machinery), 2077–2079. doi: 10.1145/3372297.3420017
Zhou, L., Burgoon, J., Nunamaker, J., and Twitchell, D. (2004). Automating linguistics-based cues for detecting deception in text-based asynchronous computer-mediated communication. Group Decis. Negot. 13, 81–106. doi: 10.1023/B:GRUP.0000011944.62889.6f
Zhou, L., and Zhang, D. (2007). An ontology-supported misinformation model: toward a digital misinformation library. IEEE Trans. Syst. Man Cybern- A: Syst. Hum. 37, 804–813. doi: 10.1109/TSMCA.2007.902648
Zhou, X., Jain, A., Phoha, V. V., and Zafarani, R. (2020). Fake news early detection: a theory-driven model. Digital Threats: Res. Pract. 1, 1–25. doi: 10.1145/3377478
Keywords: automatic/computational deception detection, cross-domain, domain-independent, email/message scams, fake news, meta-analysis, opinion spam, phishing
Citation: Verma RM, Dershowitz N, Zeng V, Boumber D and Liu X (2025) Domain-independent deception: a new taxonomy and linguistic analysis. Front. Big Data 8:1581734. doi: 10.3389/fdata.2025.1581734
Received: 23 February 2025; Accepted: 03 September 2025;
Published: 30 September 2025.
Edited by:
Murat Kantarcioglu, Virginia Tech, United StatesReviewed by:
Smita Bharne, Padmashree Dr. D.Y. Patil University, IndiaAbdur Rasool, University of Hawaii at Manoa, United States
Copyright © 2025 Verma, Dershowitz, Zeng, Boumber and Liu. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Rakesh M. Verma, cm12ZXJtYTJAY2VudHJhbC51aC5lZHU=; Nachum Dershowitz, bmFjaHVtZEB0YXUuYWMuaWw=
†Present address: Xuting Liu, University of Houston, Houston, TX, United States