Sec. Educational Psychology
Volume 6 - 2015 | https://doi.org/10.3389/fpsyg.2015.01100
Fifty psychological and psychiatric terms to avoid: a list of inaccurate, misleading, misused, ambiguous, and logically confused words and phrases
- 1Department of Psychology, Emory University, Atlanta, GA, USA
- 2Department of Psychology, Georgia State University, Atlanta, GA, USA
- 3Binghamton University – State University of New York, Binghamton, NY, USA
- 4Department of Psychology, Sacred Heart College, Fairfield, CT, USA
The goal of this article is to promote clear thinking and clear writing among students and teachers of psychological science by curbing terminological misinformation and confusion. To this end, we present a provisional list of 50 commonly used terms in psychology, psychiatry, and allied fields that should be avoided, or at most used sparingly and with explicit caveats. We provide corrective information for students, instructors, and researchers regarding these terms, which we organize for expository purposes into five categories: inaccurate or misleading terms, frequently misused terms, ambiguous terms, oxymorons, and pleonasms. For each term, we (a) explain why it is problematic, (b) delineate one or more examples of its misuse, and (c) when pertinent, offer recommendations for preferable terms. By being more judicious in their use of terminology, psychologists and psychiatrists can foster clearer thinking in their students and the field at large regarding mental phenomena.
“If names be not correct, language is not in accordance with the truth of things.”
(Confucius, The Analects)
Scientific thinking necessitates clarity, including clarity in writing (Pinker, 2014). In turn, clarity hinges on accuracy in the use of specialized terminology. Clarity is especially critical in such disciplines as psychology and psychiatry, where most phenomena, such as emotions, personality traits, and mental disorders, are “open concepts.” Open concepts are characterized by fuzzy boundaries, an indefinitely extendable indicator list, and an unclear inner essence (Pap, 1958; Meehl, 1986).
Many writers, including students, may take the inherent murkiness of many psychological and psychiatric constructs as an implicit license for looseness in language. After all, if the core concepts within a field are themselves ambiguous, the reasoning goes, precision in language may not be essential. In fact, the opposite is true; the inherent openness of many psychological concepts renders it all the more imperative that we insist on rigor in our writing and thinking to avoid misunderstandings (Guze, 1970). Researchers, teachers, and students in psychology and allied fields should therefore be as explicit as possible about what are they are saying and are not saying, as terms in these disciplines readily lend themselves to confusion and misinterpretation.
For at least two reasons, issues of terminology bear crucial implications for the education of forthcoming generations of students in psychology, psychiatry, and related domains. First, many instructors may inadvertently disseminate misinformation or foster unclear thinking by using specialized terms in inaccurate, vague, or idiosyncratic ways. Six decades ago, two prominent psychiatrists bemoaned the tendency of writers to use “jargon to blur implausible concepts and to convey the impression that something real is being disclosed” (Cleckley and Thigpen, 1955, p. 335). We hope that our article offers a friendly, albeit greatly belated, corrective in this regard. Second, if students are allowed, or worse, encouraged, to be imprecise in their language concerning psychological concepts, their thinking about these concepts is likely to follow suit. An insistence on clarity in language forces students to think more deeply and carefully about psychological phenomena, and serves as a potent antidote against intellectual laziness, which can substitute for the meticulous analysis of concepts. The accurate use of terminology is therefore a prerequisite to clear thinking within psychology and related disciplines.
Psychology has long struggled with problems of terminology (Stanovich, 2012). For example, numerous scholars have warned of the jingle and jangle fallacies, the former being the error of referring to different constructs by the same name and the latter the error of referring to the same construct by different names (Kelley, 1927; Block, 1995; Markon, 2009). As an example of the jingle fallacy, many authors use the term “anxiety” to refer interchangeably to trait anxiety and trait fear. Nevertheless, research consistently shows that fear and anxiety are etiologically separable dispositions and that measures of these constructs are only modestly correlated (Sylvers et al., 2011). As an example of the jangle fallacy, dozens of studies in the 1960s focused on the correlates of the ostensibly distinct personality dimension of repression-sensitization (e.g., Byrne, 1964). Nevertheless, research eventually demonstrated that this dimension was essentially identical to trait anxiety (Watson and Clark, 1984). In the field of social psychology, Hagger (2014) similarly referred to the “deja variable” problem, the ahistorical tendency of researchers to concoct new labels for phenomena that have long been described using other terminology (e.g., the use of 15 different terms to describe the false consensus effect; see Miller and Pedersen, 1999).
In this article, we present a provisional list of 50 commonly used terms in psychology, psychiatry, and allied fields that should be avoided, or at most used sparingly and with explicit caveats. For each term, we (a) explain why it is problematic, (b) delineate one or more examples of its misuse, and (c) when pertinent, offer recommendations for preferable terms. These terms span numerous topical areas within psychology and psychiatry, including neuroscience, genetics, statistics, and clinical, social, cognitive, and forensic psychology. Still, in proposing these 50 terms, we make no pretense at comprehensiveness. We are certain that many readers will have candidates for their own “least favorite” psychological and psychiatric terms, and we encourage them to contact us with their nominees. In addition, we do not include commonly confused terms (e.g., “asocial” with “antisocial,” “external validity” with “ecological validity,” “negative reinforcement” with “punishment,” “mass murderer” with ‘serial killer’), as we intend to present a list of these term pairs in a forthcoming publication. We also do not address problematic terms that are restricted primarily to popular (“pop”) psychology, such as “codependency,” “dysfunctional,” “toxic,” “inner child,” and “boundaries,” as our principal focus is on questionable terminology in the academic literature. Nevertheless, we touch on a handful of pop psychology terms (e.g., closure, splitting) that have migrated into at least some academic domains.
Our “eyeball cluster analysis” of these 50 terms has led us to group them into five overarching and partly overlapping categories for expository purposes: inaccurate or misleading terms, frequently misused terms, ambiguous terms, oxymorons, and pleonasms. Terms in all five categories, we contend, have frequently sown the seeds of confusion in psychology, psychiatry, and related fields, and in so doing have potentially impeded (a) their scientific progress and (b) clear thinking among students.
First, some psychological terms are inaccurate or misleading. For example, the term “hard-wired” as applied to human traits implies that genes rigidly prescribe complex psychological behaviors (e.g., physical aggression) and traits (e.g., extraversion), which is almost never the case. Second, some psychological terms are not incorrect per se, but are frequently misused. For example, although “splitting” carries a specific meaning as a defensive reaction in psychodynamic theory, it is commonly misused to refer to the propensity of people with borderline personality disorder (BPD) and related conditions to pit staff members against each other. Third, some psychological terms are ambiguous, because they can mean several things. For example, the term “medical model” can refer to any one (or more) of at least seven conceptual models of mental illness and its treatment. Fourth, some psychological terms are oxymorons. An oxymoron is a term, such as open secret, precise estimate, or final draft, which consists of two conjoined terms that are contradictory. For example, the term “stepwise hierarchical regression” is an oxymoron because stepwise and hierarchical multiple regression are incompatible statistical procedures. Fifth, some psychological terms are pleonasms. A pleonasm is a term, such as PIN number, Xerox copy, or advance warning, which consists of two or more conjoined terms that are redundant. For example, the term “latent construct” is a pleonasm because all psychological constructs are hypothetical and therefore unobservable.
Our list of 50 terms, grouped into the five aforementioned categories and presented in alphabetical order within each category, follows.
Inaccurate or Misleading Terms
(1) A gene for. The news media is awash in reports of identifying “genes for” a myriad of phenotypes, including personality traits, mental illnesses, homosexuality, and political attitudes (Sapolsky, 1997). For example, in 2010, The Telegraph (2010) trumpeted the headline, “‘Liberal gene’ discovered by scientists.” Nevertheless, because genes code for proteins, there are no “genes for” phenotypes per se, including behavioral phenotypes (Falk, 2014). Moreover, genome-wide association studies of major psychiatric disorders, such as schizophrenia and bipolar disorder, suggest that there are probably few or no genes of major effect (Kendler, 2005). In this respect, these disorders are unlike single-gene medical disorders, such as Huntington’s disease or cystic fibrosis. The same conclusion probably holds for all personality traits (De Moor et al., 2012).
Not surprisingly, early claims that the monoamine oxidase-A (MAO-A) gene is a “warrior gene” (McDermott et al., 2009) have not withstood scrutiny. This polymorphism appears to be only modestly associated with risk for aggression, and it has been reported to be associated with conditions that are not tied to a markedly heightened risk of aggression, such as major depression, panic disorder, and autism spectrum disorder (Buckholtz and Meyer-Lindenberg, 2013; Ficks and Waldman, 2014). The evidence for a “God gene,” which supposedly predisposes people to mystical or spiritual experiences, is arguably even less impressive (Shermer, 2015) and no more compelling than that for a “God spot” in the brain (see “God spot”). Incidentally, the term “gene” should not be confused with the term “allele”; genes are stretches of DNA that code for a given morphological or behavioral characteristic, whereas alleles are differing versions of a specific polymorphism in a gene (Pashley, 1994).
(2) Antidepressant medication. Medications such as tricyclics, selective serotonin reuptake inhibitors, and selective serotonin and norepinephrine reuptake inhibitors, are routinely called “antidepressants.” Yet there is little evidence that these medications are more efficacious for treating (or preventing relapse for) mood disorders than for several other conditions, such as anxiety-related disorders (e.g., panic disorder, obsessive-compulsive disorder; Donovan et al., 2010) or bulimia nervosa (Tortorella et al., 2014). Hence, their specificity to depression is doubtful, and their name derives more from historical precedence—the initial evidence for their efficacy stemmed from research on depression (France et al., 2007)—than from scientific evidence. Moreover, some authors argue that these medications are considerably less efficacious than commonly claimed, and are beneficial for only severe, but not mild or moderate, depression, rendering the label of “antidepressant” potentially misleading (Antonuccio and Healy, 2012; but see Kramer, 2011, for an alternative view).
(3) Autism epidemic. Enormous effort has been expended to uncover the sources of the “autism epidemic” (e.g., King, 2011), the supposed massive increase in the incidence and prevalence of autism, now termed autism spectrum disorder, over the past 25 years. The causal factors posited to be implicated in this “epidemic” have included vaccines, television viewing, dietary allergies, antibiotics, and viruses.
Nevertheless, there is meager evidence that this purported epidemic reflects a genuine increase in the rates of autism per se as opposed to an increase in autism diagnoses stemming from several biases and artifacts, including heightened societal awareness of the features of autism (“detection bias”), growing incentives for school districts to report autism diagnoses, and a lowering of the diagnostic thresholds for autism across successive editions of the Diagnostic and Statistical Manual of Mental Disorders (Gernsbacher et al., 2005; Lilienfeld and Arkowitz, 2007). Indeed, data indicate when the diagnostic criteria for autism were held constant, the rates of this disorder remained essentially constant between 1990 and 2010 (Baxter et al., 2015). If the rates of autism are increasing, the increase would appear to be slight at best, hardly justifying the widespread claim of an “epidemic.”
(4) Brain region X lights up. Many authors in the popular and academic literatures use such phrases as “brain area X lit up following manipulation Y” (e.g., Morin, 2011). This phrase is unfortunate for several reasons. First, the bright red and orange colors seen on functional brain imaging scans are superimposed by researchers to reflect regions of higher brain activation. Nevertheless, they may engender a perception of “illumination” in viewers. Second, the activations represented by these colors do not reflect neural activity per se; they reflect oxygen uptake by neurons and are at best indirect proxies of brain activity. Even then, this linkage may sometimes be unclear or perhaps absent (Ekstrom, 2010). Third, in almost all cases, the activations observed on brain scans are the products of subtraction of one experimental condition from another. Hence, they typically do not reflect the raw levels of neural activation in response to an experimental manipulation. For this reason, referring to a brain region that displays little or no activation in response to an experimental manipulation as a “dead zone” (e.g., Lamont, 2008) is similarly misleading. Fourth, depending on the neurotransmitters released and the brain areas in which they are released, the regions that are “activated” in a brain scan may actually be being inhibited rather than excited (Satel and Lilienfeld, 2013). Hence, from a functional perspective, these areas may be being “lit down” rather than “lit up.”
(5) Brainwashing. This term, which originated during the Korean War (Hunter, 1951) but which is still invoked uncritically from time to time in the academic literature (e.g., Ventegodt et al., 2009; Kluft, 2011), implies that powerful individuals wishing to persuade others can capitalize on a unique armamentarium of coercive procedures to change their long-term attitudes. Nevertheless, the attitude-change techniques used by so-called “brainwashers” are no different than standard persuasive methods identified by social psychologists, such as encouraging commitment to goals, manufacturing source credibility, forging an illusion of group consensus, and vivid testimonials (Zimbardo, 1997). Furthermore, there are ample reasons to doubt whether “brainwashing” permanently alters beliefs (Melton, 1999). For example, during the Korean War, only a small minority of the 3500 American political prisoners subjected to intense indoctrination techniques by Chinese captors generated false confessions. Moreover, an even smaller number (probably under 1%) displayed any signs of adherence to Communist ideologies following their return to the US, and even these were individuals who returned to Communist subcultures (Spanos, 1996).
(6) Bystander apathy. The classic work of (e.g., Darley and Latane, 1968; Latane and Rodin, 1969) underscored the counterintuitive point that when it comes to emergencies, there is rarely “safety in numbers.” As this and subsequent research demonstrated, the more people present at an emergency, the lower the likelihood of receiving help. In early research, this phenomenon was called “bystander apathy” (Latane and Darley, 1969) a term that endures in many academic articles (e.g., Abbate et al., 2013). Nevertheless, research demonstrates that most bystanders are far from apathetic in emergencies (Glassman and Hadad, 2008). To the contrary, they are typically quite concerned about the victim, but are psychologically “frozen” by well-established psychological processes, such as pluralistic ignorance, diffusion of responsibility, and sheer fears of appearing foolish.
(7) Chemical imbalance. Thanks in part to the success of direct-to-consumer marketing campaigns by drug companies, the notion that major depression and allied disorders are caused by a “chemical imbalance” of neurotransmitters, such as serotonin and norepinephrine, has become a virtual truism in the eyes of the public (France et al., 2007; Deacon and Baird, 2009). This phrase even crops up in some academic sources; for example, one author wrote that one overarching framework for conceptualizing mental illness is a “biophysical model that posits a chemical imbalance” (Wheeler, 2011, p. 151). Nevertheless, the evidence for the chemical imbalance model is at best slim (Lacasse and Leo, 2005; Leo and Lacasse, 2008). One prominent psychiatrist even dubbed it an urban legend (Pies, 2011). There is no known “optimal” level of neurotransmitters in the brain, so it is unclear what would constitute an “imbalance.” Nor is there evidence for an optimal ratio among different neurotransmitter levels. Moreover, although serotonin reuptake inhibitors, such as fluoxetine (Prozac) and sertraline (Zoloft), appear to alleviate the symptoms of severe depression, there is evidence that at least one serotonin reuptake enhancer, namely tianepine (Stablon), is also efficacious for depression (Akiki, 2014). The fact that two efficacious classes of medications exert opposing effects on serotonin levels raises questions concerning a simplistic chemical imbalance model.
(8) Family genetic studies. The phrase “family genetic studies” is commonly used in psychiatry to refer to designs in which investigators examine the familial aggregation of one or more disorders, such as panic disorder or major depression, within intact (i.e., non-adoptive) families (e.g., Weissman, 1993). Given that the familial aggregation of one or more disorders within intact families could be due to shared environment rather than—or in addition to—shared genes (Smoller and Finn, 2003), the phrase “family genetic study” is misleading. This term implies erroneously that familial clustering of a disorder is necessarily more likely to be genetic than environmental. It may also imply incorrectly (Kendler and Neale, 2009) that studies of intact families permit investigators to disentangle the effects of shared genes from shared environment. Twin or adoption studies are necessary to accomplish this goal.
(9) Genetically determined. Few if any psychological capacities are genetically “determined”; at most, they are genetically influenced. Even schizophrenia, which is among the most heritable of all mental disorders, appears to have a heritability of between 70 and 90% as estimated by twin designs (Mulle, 2012), leaving room for still undetermined environmental influences. Moreover, data strongly suggest that schizophrenia and most other major mental disorders are highly polygenic. In addition, the heritability of most adult personality traits, such as neuroticism and extraversion, appears to be between 30 and 60% (Kandler, 2012). This finding again points to a potent role for environmental influences.
(10) God spot. Seizing on functional imaging findings that religious ideation is associated with activations in specific brain regions, such as circumscribed areas of the temporal lobe, some media and academic sources have referred to the discovery of a “God spot” in the human brain (Connor, 1997). Such language is scientifically dubious given that complex psychological capacities, including religious experiences, are almost surely distributed across several sprawling networks that themselves encompass multiple brain regions. Not surprisingly, studies of people undergoing mystical experiences have reported activation in many brain areas, including the temporal lobe, caudate, inferior parietal lobe, and insula (Beauregard and Paquette, 2006; Jarrett, 2014). As one researcher (Mario Beauregard) observed, “There is no single God spot localized uniquely in the temporal lobe of the human brain” (Biello, 2007, p. 43). The same absence of localizational specificity holds for claims regarding the identification of other purported brain regions, such as an “irony spot” or “humor spot” (Jarrett, 2014).
(11) Gold standard. In the domains of psychological and psychiatric assessment, there are precious few, if any, genuine “gold standards.” Essentially all measures, even those with high levels of validity for their intended purposes, are necessarily fallible indicators of their respective constructs (Cronbach and Meehl, 1955; Faraone and Tsuang, 1994). As a consequence, the widespread practice referring to even well-validated measures of personality or psychopathology, such as Hare’s (1991/2003) Psychopathy Checklist-Revised, as “gold standards” for their respective constructs (Ermer et al., 2012) is misleading (see Skeem and Cooke, 2010). If authors intend to refer to measures as “extensively validated,” they should simply do so.
(12) Hard-wired. The term “hard-wired” has become enormously popular in press accounts and academic writings in reference to human psychological capacities that are presumed by some scholars to be partially innate, such as religion, cognitive biases, prejudice, or aggression. For example, one author team reported that males are more sensitive than females to negative news stories and conjectured that males may be “hard wired for negative news” (Grabe and Kamhawi, 2006, p. 346). Nevertheless, growing data on neural plasticity suggest that, with the possible exception of inborn reflexes, remarkably few psychological capacities in humans are genuinely hard-wired, that is, inflexible in their behavioral expression (Huttenlocher, 2009; Shermer, 2015). Moreover, virtually all psychological capacities, including emotions and language, are modifiable by environmental experiences (Merzenich, 2013).
(13) Hypnotic trance. The notion that hypnosis is characterized by a distinct “trance state” remains one of the enduring myths of popular psychology (Lilienfeld et al., 2009). In a sample of 276 undergraduates, Green (2003; see also Green et al., 2006) found that participants gave high ratings (between 5 and 5.5 on 1–7 scale in two experimental conditions) to the item, “Hypnosis is an altered state of consciousness, quite different from normal waking consciousness” (p. 373). Perhaps not surprisingly, the phrase “hypnotic trance” continues to appear in numerous articles written for the general public (Brody, 2008) as well as in academic sources (Raz, 2011). Nevertheless, the evidence that hypnosis is a distinct “trance” state that differs qualitatively from waking consciousness is scant. There is no consistent evidence for distinctive physiological (e.g., functional brain imaging) markers of hypnosis (Lynn et al., 2007). Nor is there persuasive, or even especially suggestive, evidence that hypnosis is associated with unique behavioral features. For example, suggested responses, including hallucinations, amnesia, and pain reduction, can be achieved in the absence of a “hypnotic induction” and even when participants report being awake and alert (Lynn et al., 2015).
(14) Influence of gender (or social class, education, ethnicity, depression, extraversion, intelligence, etc.) on X. “Influence” and cognate terms, such as effect, are inherently causal in nature. Hence, they should be used extremely judiciously in reference to individual differences, such as personality traits (e.g., extraversion), or group differences (e.g., gender), which cannot be experimentally manipulated. This is not to say that individual or group differences cannot exert a causal influence on behavior (Funder, 1991), only that research designs that examine these differences are virtually always (with the rare exception of “experiments of nature,” in which individual differences are altered by unusual events) correlation or quasi-experimental. Hence, researchers should be explicit that when using such phrases as “the influence of gender,” they are almost always proposing a hypothesis from the data, not drawing a logically justified conclusion from them. This inferential limitation notwithstanding, the phrase “the influence of gender” alone appears in over 45,000 manuscripts in the Google Scholar database (e.g., Bertakis et al., 1995).
(15) Lie detector test. Surely one of the most pernicious misnomers in psychology, the term “lie detector test” is often used synonymously with the storied polygraph test. This test is misnamed: it is an arousal detector, not a lie detector (Saxe et al., 1985). Because it measures non-specific psychophysiological arousal rather than the fear of detection per se, it is associated with high false-positive rates, meaning that it frequently misidentifies honest individuals as dishonest (Lykken, 1998). In addition, the polygraph test is susceptible to false-negatives stemming from the use of physical (e.g., biting the tongue) and mental (e.g., performing complex mental arithmetic) countermeasures (Honts et al., 1994). This evidence notwithstanding, the mythical allure of the polygraph test persists. In one survey, 45% of undergraduates agreed that this test is an accurate detector of falsehoods (Taylor and Kowalski, 2010).
(16) Love molecule. Over 6000 websites have dubbed the hormone oxytocin the “love molecule” (e.g., Morse, 2011). Others have named it the “trust molecule” (Dvorsky, 2012), “cuddle hormone” (Griffiths, 2014), or “moral molecule” (Zak, 2013). Nevertheless, data derived from controlled studies imply that all of these appellations are woefully simplistic (Wong, 2012; Jarrett, 2015; Shen, 2015). Most evidence suggests that oxytocin renders individuals more sensitive to social information (Stix, 2014), both positive and negative. For example, although intranasal oxytocin seems to increase within-group trust, it may also increase out-group mistrust (Bethlehem et al., 2014). In addition, among individuals with high levels of trait aggressiveness, oxytocin boosts propensities toward intimate partner violence following provocation (DeWall et al., 2014). Comparable phrases applied to other neural messengers, such as the term “pleasure molecule” as a moniker for dopamine, are equally misleading (see Landau et al., 2008; Kringelbach and Berridge, 2010, for discussions).
(17) Multiple personality disorder. Although the term “multiple personality disorder” was expunged from the American Psychiatric Association’s (1994) diagnostic manual over two decades ago and has since been replaced by “dissociative identity disorder” (DID), it persists in many academic sources (e.g., Hayes, 2014). Nevertheless, even ardent proponents of the view that DID is a naturally occurring condition that stems largely from childhood trauma (e.g., Ross, 1994) acknowledge that “multiple personality disorder” is a misnomer (Lilienfeld and Lynn, 2015), because individuals with DID do not genuinely harbor two or more fully developed personalities. Moreover, laboratory studies of the memories of individuals with DID demonstrate that the “alter” personalities or personality states of individuals with DID are not insulated by impenetrable amnestic barriers (Merckelbach et al., 2002).
(18) Neural signature. One group of authors, after observing that compliance with social norms was associated with activations in certain brain regions (lateral orbitofrontal cortex and right dorsolateral cortex), referred to the “neural signature” of social norm compliance (Spitzer et al., 2007, p. 185). Others have referred to neural signatures or “brain signatures” of psychiatric disorders, such as anorexia nervosa (Fladung et al., 2009) and autism spectrum disorder (Pelphrey and McPartland, 2012). Nevertheless, identifying a genuine neural signature would necessitate the discovery of a specific pattern of brain responses that possesses nearly perfect sensitivity and specificity for a given condition or other phenotype. At the present time, neuroscientists are not remotely close to pinpointing such a signature for any psychological disorder or trait (Gillihan and Parens, 2011).
(19) No difference between groups. Many researchers, after reporting a group difference that does not attain conventional levels of statistical significance, will go on to state that “there was no difference between groups.” Similarly, many authors will report that a non-significant correlation between two variables means that “there was no association between the variables.” But a failure to reject the null hypothesis does not mean that the null hypothesis, strictly speaking, has been confirmed. Indeed, if an investigator finds a correlation of r = 0.11 in a sample of 20 participants (which is not statistically significant), the best estimate for the true value of the correlation in the population, presuming that the sample has been randomly ascertained, is 0.11, not 0. Authors are instead advised to write “no significant difference between groups” or “no significant correlation between variables.”
(20) Objective personality test. Many authors refer to paper-and-pencil personality instruments that employ a standard (e.g., True–False) item response format, such as the Minnesota Multiphasic Personality Inventory-2 (MMPI-2), as “objective tests” (Proyer and Häusler, 2007), ostensibly to contrast them with more “subjective” measures, such as unstructured interviews or projective techniques (e.g., the Rorschach Inkblot Test). Nevertheless, although the former measures can be scored objectively, that is, with little or no error (but see Allard and Faust, 2000, for evidence of non-trivial error rates in the hand-scoring of the MMPI and other purported “objective” personality tests), they often require considerable subjective judgment on the part of respondents. For example, an item such as “I have many headaches” can be interpreted in numerous ways arising from ambiguity in the meanings of “many” and “headache’ (Meehl, 1945). So-called “objective” personality tests are also often subjective with respect to interpretation (Rogers, 2003). For example, even different computerized MMPI-2 interpretive programs display only moderate levels of inter-rater agreement regarding proposed diagnoses (Pant et al., 2014). Not surprisingly, clinicians routinely disagree in their interpretations of profiles on the MMPI-2 and other “objective” tests (Garb, 1998). We therefore recommend that these measures be called “structured” tests (Kaplan and Saccuzzo, 2012), a term that refers only to their response format and that carries no implication that they are interpreted objectively by either examinee or examiner.
(21) Operational definition. The credo that all psychological investigators must develop “operational definitions” of constructs before conducting studies has become something of a truism in many psychology methods textbooks and other research sources (e.g., Burnette, 2007). Operational definitions are strict definitions of concepts in terms of their measurement operations. As a consequence, they are presumed to be exact and exhaustive definitions of these concepts. Perhaps the best known example in psychology is Boring’s (1923) definition of intelligence as whatever intelligence tests measure.
Many psychologists appear unaware that the notion of operational definitions was roundly rejected by philosophers of science decades ago (Leahey, 1980; Green, 1992; Gravetter and Forzano, 2012). Operational definitions are unrealistic in virtually all domains of psychology, because constructs are not equivalent to their measurement operations (Meehl, 1986). For example, an “operational definition” of aggression as the amount of hot sauce a participant places in an experimental confederate’s drink is not an operational definition at all, because no researcher seriously believes that the amount of hot sauce placed in a drink is a perfect or precise definition of aggression that exhausts all of its potential manifestations. Operational definitions also fell out of favor because they led to logically absurd conclusions. For example, an operational definition of length would imply that length as measured by a wooden ruler cannot be compared with length as measured by a metal ruler, because these rulers are associated with different measurement operations. Hence, the fact that both rulers yield a length for a table of say, 27 inches, could not be taken as converging evidence that the table is in fact 27 inches long (Green, 1992).
Psychological researchers and teachers should therefore almost always steer clear of the term “operational definition.” The term “operationalization” is superior, as it avoids the implication of an ironclad definition and is largely free of the problematic logical baggage associated with its sister term.
(22) p = 0.000. Even though this statistical expression, used in over 97,000 manuscripts according to Google Scholar, makes regular cameo appearances in our computer printouts, we should assiduously avoid inserting it in our Results sections. This expression implies erroneously that there is a zero probability that the investigators have committed a Type I error, that is, a false rejection of a true null hypothesis (Streiner, 2007). That conclusion is logically absurd, because unless one has examined essentially the entire population, there is always some chance of a Type I error, no matter how meager. Needless to say, the expression “p < 0.000” is even worse, as the probability of committing a Type I error cannot be less than zero. Authors whose computer printouts yield significance levels of p = 0.000 should instead express these levels out to a large number of decimal places, or at least indicate that the probability level is below a given value, such as p < 0.01 or p < 0.001.
(23) Psychiatric control group. This phrase and similar phrases (e.g., “normal control group,” “psychopathological control group”) connote erroneously that (a) groups of ostensibly normal individuals or mixed psychiatric patients who are being compared with (b) groups of individuals with a disorder of interest (e.g., schizophrenia, major depression) are true “control” groups. They are not. They are “comparison groups” and should be referred to accordingly. The phrase “control group” in this context may leave readers with the unwarranted impression that the design of the study is experimental when it is actually quasi-experimental. Just as important, this term may imply that the only difference between the two groups (e.g., a group of patients with anxiety disorder and a group of ostensibly normal individuals) is the presence or absence of the disorder of interest. In fact, these two groups almost surely differ on any number of “nuisance” variables, such as personality traits, co-occurring disorders, and family background, rendering the interpretation of most group differences open to multiple interpretations (Meehl, 1969).
(24) Reliable and valid. If one earned a dollar for every time an author used the sentence “This test is reliable and valid” in a Method section, one would be a rich person indeed, as the phrase “reliable and valid” appears in more than 190,000 manuscripts in Google Scholar. There are at least three problems with this ubiquitous phrase. First, it implies that a psychological test is either valid or not valid. Much like the testing of scientific theories, the construct validation process is never complete, in essence reflecting a “work in progress.” As a consequence, a test cannot be said to be have been conclusively validated or invalidated (Cronbach and Meehl, 1955; Loevinger, 1957; Peter, 1981). Hence, authors should similarly refrain from using the term “validated’ with respect to psychological measures. At best, these measures are “empirically supported” or have “accrued substantial evidence for construct validity.” The same caveat applies to psychological treatments. When Division 12 (Society of Clinical Psychology) of the American Psychological Association put forth its criteria for, and lists of, psychotherapies found to work in controlled trials for specific mental disorders, it initially termed them “empirically validated therapies” (Chambless et al., 1998). Nevertheless, in recognition of the fact that “validation” implies certainty or finality (Garfield, 1996; Chambless and Hollon, 1998), the committee wisely changed the name to “empirically supported therapies,” which is now the term presently in use (Lilienfeld et al., 2013).
Second, the phrase “reliable and valid” implies that reliability and validity are unitary concepts. They are not. There are three major forms of reliability: test–retest, internal consistency, and inter-rater. Contrary to common belief, these forms of reliability often diverge, sometimes markedly (Schmidt and Hunter, 1996). For example, scores derived from the Thematic Apperception Test, a widely used projective technique, frequently display high levels of test–retest reliability but low levels of internal consistency (Entwistle, 1972). There are also multiple forms of validity (e.g., content, criterion-related, incremental), which similarly do not necessarily coincide. For example, a measure may possess high levels of criterion-related validity in multiple samples but little or no incremental validity above and beyond extant information (Garb, 2003).
Third, reliability and validity are conditional on the specific samples examined, and should not be considered inherent properties of a test. Hence, the notion that a test is “reliable and valid” independent of the nature of the sample runs counter to contemporary thinking in psychometrics (American Psychological Association and American Educational Research Association, 2014).
(25) Statistically reliable. This phrase appears in over 62,000 manuscripts according to Google Scholar. It is typically invoked when referring to statistical significance, e.g., “Although small in absolute terms, this difference was statistically reliable, t(157) = 2.86, p = 0.005” (Zurbriggen et al., 2011, p. 453). Nevertheless, despite what many psychologists believe (Tversky and Kahneman, 1971; Krueger, 2001), statistical significance bears at best a modest conceptual and empirical association with a result’s “reliability,” that is, its replicability or consistency over time (Carver, 1978). Indeed, given the low statistical power of most studies in psychology, a reasonable argument could be advanced that most statistically significant results are unlikely to be reliable. The statistical significance of a result should therefore not be confused with its likelihood of replication (Miller, 2009).
(26) Steep learning curve. Scores of authors use the phrase “steep learning curve” or “sharp learning curve” in reference to a skill that is difficult to master. For example, when referring to the difficulty of learning a complex surgical procedure (endoscopic pituitary surgery), one author team contended that it “requires a steep learning curve” (Koc et al., 2006, p. 299). Nevertheless, from the standpoint of learning theory, these and other authors have it backward, because a steep learning curve, i.e., a curve with a large positive slope, is associated with a skill that is acquired easily and rapidly (Hopper et al., 2007).
(27) The scientific method. Many science textbooks, including those in psychology, present science as a monolithic “method.” Most often, they describe this method as a hypothetical-deductive recipe, in which scientists begin with an overarching theory, deduce hypotheses (predictions) from that theory, test these hypotheses, and examine the fit between data and theory. If the data are inconsistent with the theory, the theory is modified or abandoned. It’s a nice story, but it rarely works this way (McComas, 1996). Although science sometimes operates by straightforward deduction, serendipity and inductive observations offered in the service of the “context of discovery” also play crucial roles in science. For this reason, the eminent philosopher of science Popper (1983) quipped that, “As a rule, I begin my lectures on Scientific Method by telling my students that the scientific method does not exist…” (p. 5).
Contrary to what most scientists themselves appear to believe, science is not a method; it is an approach to knowledge (Stanovich, 2012). Specifically, it is an approach that strives to better approximate the state of nature by reducing errors in inferences. Alternatively, one can conceptualize science as a toolbox of finely honed tools designed to minimize mistakes, especially confirmation bias – the ubiquitous propensity to seek out and selectively interpret evidence consistent with our hypotheses and to deny, dismiss, and distort evidence that does not (Tavris and Aronson, 2007; Lilienfeld, 2010). Not surprisingly, the specific research methods used by psychologists bear scant surface resemblance to those used by chemists, astrophysicists, or molecular biologists. Nevertheless, all of these methods share an overarching commitment to reducing errors in inference and thereby arriving at a more accurate understanding of reality.
(28) Truth serum. “Truth serum” is a supposed substance that, when administered intravenously, leads individuals to disclose accurate information that they have withheld. Most so-called truth serums are actually barbiturates, such as sodium amytal or sodium pentothal (Keller, 2005). Even today, some prominent psychiatrists still refer to these substances as truth serums (e.g., Lieberman, 2015), and they are still frequently administered for legal purposes in certain countries, such as India (Pathak and Srivastava, 2011). Nevertheless, there is no evidence that so-called truth serums reveal veridical information regarding past events, such as childhood sexual abuse (Bimmerle, 1993). To the contrary, like other suggestive memory procedures, they are associated with a heightened risk of false memories and false confessions (Macdonald, 1955), probably because they lower the response threshold for reporting all information, accurate and inaccurate alike. Furthermore, individuals can and do readily lie under the influence of truth serum (Piper, 1993).
(29) Underlying biological dysfunction. In this era of the increasing biologization of psychology and psychiatry (Miller, 2010; Satel and Lilienfeld, 2013), authors may be tempted to assume that biological variables, such as parameters of brain functioning, “underlie” psychological phenomena. For example, one set of authors wrote that “cognitive impairments are central to schizophrenia and may mark underlying biological dysfunction” (Bilder et al., 2011, p. 426). Nevertheless, conceptualizing biological functioning as inherently more “fundamental” than (that is, causally prior to) psychological functioning, such as cognitive and emotional functioning, is misleading (Miller, 1996). The relation between biological variables and other variables is virtually always bidirectional. For example, although the magnitude of the P300 event-related potential tends to be diminished among individuals with antisocial personality disorder (ASPD) compared with other individuals (Costa et al., 2000), this finding does not necessarily mean that the P300 deficit precedes, let alone plays a causal role in, ASPD. It is at least equally plausible that the personality dispositions associated with ASPD, such as inattention, low motivation, and poor impulse control, contribute to smaller P300 magnitudes (Lilienfeld, 2014). The same inferential limitation applies to many similar phrases, such as “biological bases of behavior,” “brain substrates of mental disorder,” and “neural underpinnings of personality” (Miller, 1996).
Frequently Misused Terms
(30) Acting out. Numerous articles use this term as a synonym for any kind of externalizing or antisocial behavior, including delinquency (e.g., Weinberger and Gomes, 1995). In fact, the term “acting out” carries a specific psychoanalytic meaning that refers to the behavioral enactment of unconscious drives that are ostensibly forbidden by the superego (Fenichel, 1945). Hence, this term should not be used interchangeably with disruptive behavior of all kinds and attributable to all causes.
(31) Closure. The term “closure” was introduced by Gestalt psychologists (Koffka, 1922) to refer to the tendency to perceive incomplete figures as wholes. This term has since been misappropriated by popular psychologists (Howard, 2011) and social scientists of various stripes (e.g., Skitka et al., 2004) to describe the purported experience of emotional resolution experienced by victims of trauma following an event of symbolic importance. For example, many advocates of the “closure movement” contend that the execution of a murderer assists the loved ones of victims to put an end to their grieving process. Nevertheless, this use of the term “closure” is hopelessly vague, as it is rarely if ever clear when trauma victims have achieved the desired emotional end-state (Radford, 2003; Weinstein, 2011). Nor is there research support for the proposition that many or most victims experience this end-state after events of symbolic significance, such as executions or funerals (Berns, 2011).
(32) Denial. Denial, a psychodynamic defense mechanism popularized by Freud (1937), is an ostensibly unconscious refusal to acknowledge obvious facts of reality, such as the death of a loved one in an automobile accident (Vaillant, 1977). Nevertheless, thanks largely to the popular psychology industry, this term has been widely misappropriated to refer to the tendency of individuals with a psychological condition, such as alcohol use disorder (formerly called alcoholism), to minimize the extent of their pathology (e.g., Wing, 1995).
(33) Fetish. A fetish, formally referred to as “Fetishistic Disorder” in the current version of the Diagnostic and Statistical Manual of Mental Disorders (DSM-5; American Psychiatric Association, 2013, p. 700), is a psychiatric condition marked by persistent, intense, and psychologically impairing sexual arousal derived from inanimate objects (e.g., shoes) or non-genital body parts (e.g., legs). This term, which is technically a paraphilia, should not be used to refer to generic preferences for specific objects, ideas, or people. One writer, for example, described the national fascination of the Japanese with smartphones as a “feature phone fetish” (Smith, 2015).
(34) Splitting. “Splitting” similarly refers to a psychodynamic defense mechanism, ostensibly ubiquitous in BPD, that forces individuals to see people as all good or all bad rather than in shades of gray, warts and all (Muller, 1992). By engaging in splitting, people with BPD and similar conditions are hypothesized to avoid the anxiety of perceiving those they love as the hopelessly flawed creatures that they are. Nevertheless, this term is consistently misused to refer to the propensity of people with BPD to “pit” staff members on a psychiatric unit (or other caregivers) against one another. This disruptive behavior, sometimes termed “staff splitting” (Linehan, 1989), should not be confused with the formal meaning of splitting.
(35) Comorbidity. This term, which has become ubiquitous in publications on the relations between two or more mental disorders (appearing in approximately 444,000 citations in Google Scholar), refers to the overlap between two diagnoses, such as major depression and generalized anxiety disorder. A similar term, “dual diagnosis,” which has acquired considerable currency in the substance abuse literature in particular, refers to the simultaneous presence of a mental disorder, such as schizophrenia, and a substance abuse disorder, such as alcoholism (Dixon, 1999). Some authors have taken the comorbidity concept further, extending it to “trimorbidity” (Cornelius et al., 2001) or “quatromorbidity” (Newman et al., 1998).
Nevertheless, “comorbidity” can mean two quite different things. It can refer to either the (a) covariation (or correlation) between two diagnoses within a sample or the population or (b) co-occurrence between two diagnoses within an individual (Lilienfeld et al., 1994; Krueger and Markon, 2006). The first meaning refers to the extent to which Condition A and B are statistically associated across individuals; for example, there is substantial covariation between ASPD and BPD (Becker et al., 2014). The second meaning is a conditional probability referring to the proportion of individuals with Condition A who meet diagnostic criteria for Condition B. For example, in the case of the latter meaning, researchers might note that 45% of patients with ASPD also meet diagnostic criteria for BPD. The difference between these two meanings is hardly trivial, because they tend to be differentially influenced by base rates (prevalences). If the base rates of one or more conditions change, the covariation between them will not necessarily be affected but the level of co-occurrence almost always will be (Lilienfeld et al., 1994). Moreover, depending on the base rates of the diagnoses in a sample, two conditions may display little or no covariation but substantial co-occurrence. For example, although ASPD and major depression typically display only modest covariation (Goodwin and Hamilton, 2003), the rates of co-occurrence between ASPD and major depression in an analysis conditioned on major depression (that is, the rates of ASPD among people with major depression) would be extremely high in a prison sample, because most prison inmates meet criteria for ASPD (Flint-Stevens, 1993). Hence, the levels of comorbidity would probably be negligible in the first case but high in the second. If authors elect to use the term “comorbidity,” they should therefore be explicit about which meaning (covariation or co-occurrence) they intend.
Some authors (Lilienfeld et al., 1994) have further questioned the routine use of the term comorbidity in psychopathology research given that this term, much like “dual diagnosis,” presupposes that the conditions in question are etiologically and pathologically separable entities (but see Rutter, 1994; Spitzer, 1994, for demurrals). For example, although the high level of “comorbidity” between ASPD and BPD may reflect covariation or co-occurrence between two distinct conditions, it may instead reflect the fact that the current diagnostic system is attaching different names to slightly different manifestations of a shared diathesis, thereby falling prey to a jangle fallacy. To take an admittedly extreme example, how likely is it that a participant in a published study who simultaneously met diagnostic criteria for all 10 DSM personality disorders (see Lilienfeld et al., 2013) genuinely possessed 10 distinct disorders at the same time? Critics of the expansive application of the term comorbidity to descriptive psychopathology contend that these diagnostic conundrums are best explained by a flawed diagnostic system that is attaching different names to highly overlapping constructs.
(36) Interaction. As Olweus (1977) observed in the context of the person-situation debate, the term “interaction” has multiple meanings, some of them logically incompatible. For example, the familiar phrase “genes and environment interact for Disorder X” can mean any one of four things: (a) genes and environment are both involved in the causes of Disorder X; (b) the relation between genes and environments are bidirectional, because genes influence the environments to which people are exposed (by means of gene-environment correlations), and environments influence which genes are activated or inactivated (by means of epigenetic processes); (c) the influences of genes and environment are inseparable because of continuous transaction within individuals; or (d) the statistical effects of genes depend on people’s environments, and the statistical effects of environments depend on people’s genes. Only meaning (d) refers to a statistical interaction in the standard multiple regression or analysis of variance sense.
Two points are worth noting here. First, psychologists routinely confuse meanings (a) and (d). For example, when researchers write that “All reasonable scholars today agree that genes and environment interact to determine complex cognitive outcomes” (Bates et al., 1998, p. 195), some readers may assume that they are referring to the standard statistical meaning of the term “interaction,” (McClelland and Judd, 1993), i.e., a multiplicative rather than additive relation between variables, such as that between genetic and environmental influences. Instead, in this case the authors appear to be saying only that both genes and environment play a role in cognitive outcomes, a scenario that does not require a multiplicative relation between genes and environment. Second, meanings (c) and (d) are logically incompatible, because if the effects of genes and environment are not separable, then clearly they cannot be distinguished in statistical designs. The bottom line: when authors use the term “interaction,” they should be explicit about which of the four meanings they intend.
(37) Medical model. Although many authors who invoke the term “medical model” presume that it refers to a single conceptualization (e.g., Mann and Himelein, 2008), it does not. Some authors insist that the term is so vague and unhelpful that we are better off without it (Meehl, 1995). Among other things, it has been wielded by various authors to mean (a) the assumption of a categorical rather than dimensional model of psychopathology; (b) an emphasis on underlying “disease” processes rather than on presenting signs and symptoms; (c) an emphasis on the biological etiology of psychopathology; (d) an emphasis on pathology rather than on health; (e) the assumption that mental disorders are better treated by medications and other somatic therapies than by psychotherapy; (f) the assumption that mental disorders are better treated by physicians than by psychologists; or (g) the belief that mentally ill individuals who engage in irresponsible behavior are not fully responsible for such behavior (see Blaney, 1975, 2015, for discussions). Similar semantic and conceptual ambiguities bedevil the term “disease model” when applied to addictions and most other psychological conditions (e.g., Graham, 2013).
(38) Reductionism. There may be no greater insult in psychological circles than to brand a colleague a “reductionist.” Indeed, merely accusing a fellow faculty member of “being reductionistic” is often an effective conversation-stopper at cocktail parties. The negative connotation attached to this term neglects the point, overlooked by many authors (e.g., Harris, 2015), that “reductionism” is not one approach. Robinson (1995) delineated multiple forms of reductionism, including (a) nominalistic reduction, i.e., reduction at the level of names (“A brain structure called the amygdala plays a key role in fear processing”); (b) nomological reduction, i.e., reduction at the level of scientific explanation (“The perception of edges is mediated in part by feature detection cells in the visual cortex”); and (c) ontological reduction, i.e., reduction by eliminating immaterial entities (“Neuroscientific data strongly suggest that there is no immaterial soul”).
More broadly, we can differentiate between two quite different brands of reductionism: constitutive and eliminative, the latter termed “greedy reductionism” by Dennett (1995). The constitutive reductionist believes merely that everything that is “mental” is ultimately material at some level, and that the “mind” is what the brain and rest of the central nervous system do. Constitutive reductionists (like nomological reductionists; Robinson, 1995), who appear to comprise an overwhelming majority of psychologists and neuroscientists, reject mind-body dualism, the claim that the mind is entirely separable from the brain. In contrast, eliminative reductionists go a large step further (Lilienfeld, 2007). They contend that the “mind” will eventually be explained away entirely by lower-level concepts derived from neuroscience, and that mentalist concepts, such as thoughts, motives, and emotions, will ultimately be rendered superfluous by neuroscientific explanations. For eliminative reductionists, the field of psychology will eventually be “gobbled up” by neuroscience. Although we do not attempt to adjudicate the dispute between constitutive and eliminative reductionists here, suffice it to say that “reductionism” does not carry a single meaning in psychology. As a result, psychologists who use “reductionist” as a handy term of opprobrium against their colleagues must be explicit about which form of reductionism they are invoking.
(39) Hierarchical stepwise regression. Hierarchical and stepwise multiple regression are entirely separate – and incompatible - procedures. Still, they are readily confused, because in hierarchical regression, variables are entered in sequential steps. Specifically, in hierarchical multiple regression the investigator specifies an a priori order of entry of the variables, ideally on theoretical grounds. In contrast, in stepwise multiple regression, the investigators allows the computer to select the order of entry of the variables (and the final variables in the equation) on empirical grounds, namely, by choosing each successive predictor based on the highest incremental contribution to variability in the outcome variable (Wampold and Freund, 1987; Petrocelli, 2003). Many authors have wisely warned against the routine use of stepwise regression procedures on the grounds that they typically capitalize heavily on chance fluctuations in datasets and rarely yield replicable results (Thompson, 1989).
(40) Mind-body therapies. The term “mind-body therapy” (e.g., Naliboff et al., 2008) refers to a panoply of treatments, such as relaxation, meditation, Reiki, yoga, and biofeedback, that purportedly harness mental functioning to enhance physical health (Wolsko et al., 2004). This term implies erroneously that the “mind” is materially separate from the “body” and thereby endorses a simplistic version of mind-body dualism. Rather than conceptualizing such interventions as making use of the mind to influence the body, we should conceptualize them as making use of one part of the body to influence another.
(41) Observable symptom. This term, which appears in nearly 700 manuscripts according to Google Scholar, conflates signs with symptoms. Signs are observable features of a disorder; symptoms are unobservable features of a disorder that can only be reported by patients (Lilienfeld et al., 2013; Kraft and Keeley, 2015). Symptoms are by definition unobservable.
(42) Personality type. Although typologies have a lengthy history in personality psychology harkening back to the writings of the Roman physician Galen and later, Swiss psychiatrist Carl Jung, the assertion that personality traits fall into distinct categories (e.g., introvert vs. extravert) has received minimal scientific support. Taxometric studies consistently suggest that normal-range personality traits, such as extraversion and impulsivity, are underpinned by dimensions rather than taxa, that is, categories in nature (Haslam et al., 2012). With the possible exception of schizotypal personality disorder (but see Ahmed et al., 2013), the same conclusion holds for personality disorders (Haslam et al., 2012). Hence, if authors elect to use the phrase “personality type,” they should qualify it by noting that the evidence for a genuine typology (i.e., a qualitative difference from normality) is in almost all cases negligible within the personality domain.
(43) Prevalence of trait X. Authors in the psychological and psychiatric literatures frequently refer to “the prevalence” or “base rate” of attributes that are dimensionally distributed in the population, such as personality traits and intelligence. For example, one author team referred to the “greater prevalence of extraversion in American students” (p. 1153) compared with Korean students (Song and Kwon, 2012). Nevertheless, such terms as “prevalence,” “incidence,” “base rate,” “false positive,” and “false negative” are premised on a taxonic model: they presume that the phenomena in question are inherently categorical, that is, either present or absent in nature. For psychological features that are continuously distributed, such terms should be avoided. In the aforementioned phrase, referring to “higher levels of extraversion in American students” would have been more accurate.
(44) Principal components factor analysis. According to Google Scholar, this phrase appears in thousands of articles, including one co-authored by the first author of this manuscript (Reynolds et al., 1988). Nevertheless, this phrase is incoherent, because principal components analysis (which is commonly misspelled as “principle components analysis”) and factor analysis are incompatible approaches to data analysis. Principal components analysis is a data reduction technique that relies on the total variance of the variables in a dataset; its principal goal is to create a smaller set of weighted variables (variates) that approximate the variance of the original variables (Weiss, 1970). In contrast, factor analysis relies only on the shared variance of the variables in a dataset, and it is designed to identify underlying dimensions that best explain the covariation among these variables (Bryant and Yarnold, 1995). In contrast to principal components analysis, whose primary aim is to simplify a dataset by yielding fewer observed variables, the primary aim of exploratory factor analysis is to identify dimensions that ostensibly account for the covariation among the observed variables.
(45) Scientific proof. The concepts of “proof” and “confirmation” are incompatible with science, which by its very nature is provisional and self-correcting (McComas, 1996). Hence, it is understandable why Popper (1959) preferred the term “corroboration” to “confirmation,” as all theories can in principle be overturned by new evidence. Nor is the evidence for scientific theories dichotomous; theories virtually always vary in their degree of corroboration. As a consequence, no theory in science, including psychological science, should be regarded as strictly proven. Proofs should be confined to the pages of mathematics textbooks and journals (Kanazawa, 2008).
(46) Biological and environmental influences. This phrase implies that biological influences are necessarily genetic, and cannot be environmental. Nevertheless, “environmental influences” encompass everything external to the organism that affects its behavior following its fertilization as a zygote. As a consequence, the environment comprises not only psychosocial influences, but also non-genetic biological influences, such as nutrition, viruses, and exposure to lead and other toxins (e.g., Nisbett et al., 2012). The phrase “biological and environmental influences” is therefore a partial pleonasm.
(47) Empirical data. “Empirical” means based on observation or experience. As a consequence, with the possible exception of information derived from archival sources, all psychological data are empirical (what would “non-empirical” psychological data look like?). Some of the confusion probably stems from the erroneous equation of “empirical” with “experimental” or “quantitative.” Data derived from informal observations, such as non-quantified impressions collected during a psychotherapy session, are also empirical. If writers wish to distinguish numerical data from other sources of data, they should simply call them “quantified data.”
(48) Latent construct. A “construct” in psychology is a hypothesized attribute of individuals that cannot be directly observed, such as general intelligence, extraversion, or schizophrenia (Cronbach and Meehl, 1955; Messick, 1987). Therefore, all constructs are latent. The same terminological consideration applies to the phrase “hypothetical construct.” Authors would be better advised to instead use “construct” or “latent variable.”
(49) Mental telepathy. Telepathy, one of the three ostensible types of extrasensory perception (along with clairvoyance and precognition), is the purported ability to read other’s minds by means of psychic powers (Hyman, 1995). Hence, all telepathy is necessarily mental. The term “mental telepathy,” which appears to be in common currency in the academic literature (e.g., Lüthi, 2013; Sagi-Schwartz et al., 2014), implies erroneously that there are “non-mental” forms of telepathy.
(50) Neurocognition. Many authors have invoked the term “neurocognition” to refer to cognition, especially when conceptualized within a biological framework (e.g., Mesholam-Gately et al., 2009). Nevertheless, because all cognition is necessarily neural at some level of analysis, the simpler term “cognition” will do. In fairness, “neurocognition” is merely one among dozens of terms preceded by the prefix “neuro” that have recently become popular, including neuroeducation, neuroaesthetics, neuropolitics, neuropsychoanalysis, and neurosexology (Satel and Lilienfeld, 2013). In the words of one psychologist, “Unable to persuade others about your viewpoint? Take a Neuro-Prefix – influence grows or your money back” (Laws, 2012).
We modestly hope that our admittedly selective list of 50 terms to avoid will become recommended, if not required, reading for students, instructors, and researchers in psychology, psychiatry, and similar disciplines. Although jargon has a crucial place in these fields, it must be used with care, as the imprecise use of terminology can engender conceptual confusion. At the very least, we hope that our article encourages further discussion regarding the vital importance of clear writing and clear thinking in science, and underscores the point that clarity in writing and thinking are intimately linked. Clear writing fosters clear thinking, and confused writing fosters confused thinking. In the words of author McCullough (2002), “Writing is thinking. To write well is to think clearly. That’s why it’s so hard.”
Conflict of Interest Statement
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Abbate, C. S., Ruggieri, S., and Boca, S. (2013). The effect of prosocial priming in the presence of bystanders. J. Soc. Psychol. 153, 619–622. doi: 10.1080/00224545.2013.791658
Ahmed, A. O., Green, B. A., Goodrum, N. M., Doane, N. J., Birgenheir, D., and Buckley, P. F. (2013). Does a latent class underlie schizotypal personality disorder? Implications for schizophrenia. J. Abnorm. Psychol. 122, 475–491. doi: 10.1037/a0032713
Akiki, T. (2014). The etiology of depression and the therapeutic implications. Glob. J. Med. Res. 13. Available at: http://medicalresearchjournal.org/index.php/GJMR/article/viewFile/465/383
Allard, G., and Faust, D. (2000). Errors in scoring objective personality tests. Assessment 7, 119–129. doi: 10.1177/107319110000700203
American Psychiatric Association. (1994). Diagnostic and Statistical Manual of Mental Disorders, 4th Edn. Washington, DC: American Psychiatric Association.
American Psychiatric Association. (2013). Diagnostic and Statistical Manual of Mental Disorders, 5th Edn. Washington, DC: American Psychiatric Association.
American Psychological Association and American Educational Research Association. (2014). The Standards for Educational and Psychological Testing. Washington, DC: APA and AERA.
Antonuccio, D., and Healy, D. (2012). Relabeling the medications we call antidepressants. Scientifica 2012:965908. doi: 10.6064/2012/965908
Bates, E., Elman, J., Johnson, M., Karmiloff-Smith, A., Parisi, D., and Plunkett, K. (1998). “Innateness and emergentism,” in A Companion to Cognitive Science, eds W. Bechtel and G. Graham (Oxford: Basil Blackwell), 590–601.
Baxter, A. J., Brugha, T. S., Erskine, H. E., Scheurer, R. W., Vos, T., and Scott, J. G. (2015). The epidemiology and global burden of autism spectrum disorders. Psychol. Med. 45, 601–613. doi: 10.1017/S003329171400172X
Beauregard, M., and Paquette, V. (2006). Neural correlates of a mystical experience in Carmelite nuns. Neurosci. Lett. 405, 186–190. doi: 10.1016/j.neulet.2006.06.060
Becker, D. F., Grilo, C. M., Edell, W. S., and McGlashan, T. H. (2014). Comorbidity of borderline personality disorder with other personality disorders in hospitalized adolescents and adults. Am. J. Psychiatry 157, 2011–2016. doi: 10.1176/appi.ajp.157.12.2011
Berns, N. (2011). Closure: The Rush to End Grief and What it Costs Us. Philadelphia, PA: Temple University Press.
Bertakis, K. D., Helms, L. J., Callahan, E. J., Azari, R., and Robbins, J. A. (1995). The influence of gender on physician practice style. Med. Care 33, 407–416. doi: 10.1097/00005650-199504000-00007
Bethlehem, R. A., Baron-Cohen, S., van Honk, J., Auyeung, B., and Bos, P. A. (2014). The oxytocin paradox. Front. Behav. Neurosci. 8:48. doi: 10.3389/fnbeh.2014.00048
Biello, D. (2007). Searching for God in the brain. Sci. Am. Mind 18, 38–45. doi: 10.1038/scientificamericanmind1007-38
Bilder, R. M., Howe, A., Novak, N., Sabb, F. W., and Parker, D. S. (2011). The genetics of cognitive impairment in schizophrenia: a phenomic perspective. Trends Cogn. Sci. 15, 428–435. doi: 10.1016/j.tics.2011.07.002
Bimmerle, G. (1993). Truth Drugs in Interrogation. Washington, DC: Central Intelligence Agency Library.
Blaney, P. H. (1975). Implications of the medical model and its alternatives. Am. J. Psychiatry 132, 911–914. doi: 10.1176/ajp.132.9.911
Blaney, P. H. (2015). “Medical model of mental disorders,” in Encyclopedia of Clinical Psychology, eds R. L. Cautin and S. O. Lilienfeld (New York: Wiley), 1767–1772.
Block, J. (1995). A contrarian view of the five-factor approach to personality description. Psychol. Bull. 117, 187–215. doi: 10.1037/0033-2909.117.2.187
Boring, E. G. (1923). Intelligence as the tests test it. New Republic 35, 35–37.
Brody, J. (2008). The possibilities in hypnosis, where the patient has the power. New York Times. http://www.nytimes.com/2008/11/04/health/04brody.html? (access November 3, 2008).
Bryant, F. B., and Yarnold, P. R. (1995). “Principal-components analysis and exploratory and confirmatory factor analysis,” in Reading and Understanding Multivariate Statistics, eds F. B. Bryant, P. R. Yarnold, and L. Grimm (Washington, DC: American Psychological Association), 96–136.
Buckholtz, J. W., and Meyer-Lindenberg, A. (2013). “MAOA and the bioprediction of antisocial behavior: science fact and science fiction,” in Bioprediction Biomarkers, and Bad Behavior: Scientific, Legal, and Ethical Challenges, eds I. Singh, W. P. Sinnott-Armstrong, and J. Savulescu (Oxford: Oxford University Press), 131.
Burnette, J. L. (2007). “Operationalization,” in Encyclopedia of Social Psychology, eds R. F. Baumeister and K. D. Vohs (Thousand Oaks, CA: Sage), 636–637. doi: 10.4135/9781412956253.n379
Byrne, D. (1964). Repression-sensitization as a dimension of personality. Prog. Exp. Personal. Res. 72, 169–220.
Carver, R. P. (1978). The case against statistical significance testing. Harv. Educ. Rev. 48, 378–399. doi: 10.17763/haer.48.3.t490261645281841
Chambless, D. L., Baker, M. J., Baucom, D. H., Beutler, L. E., Calhoun, K. S., Crits-Christoph, P., et al. (1998). Update on empirically validated therapies, II. Clin. Psychol. 51, 3–16.
Chambless, D. L., and Hollon, S. D. (1998). Defining empirically supported therapies. J. Consult. Clin. Psychol. 66, 7–18. doi: 10.1037/0022-006X.66.1.7
Cleckley, H. M., and Thigpen, C. H. (1955). The dynamics of illusion. Am. J. Psychiatry 112, 334–342. doi: 10.1176/ajp.112.5.334
Connor, S. (1997). “God Spot” is Found in Brain. Los Angeles Times. Available at: http://members.shaw.ca/tfrisen/Science/God%20Module%20off%20internet.htm (accessed October 29, 1997).
Cornelius, J. R., Salloum, I. M., Lynch, K., Clark, D. B., and Mann, J. J. (2001). Treating the Substance-Abusing Suicidal Patient. Ann. N. Y. Acad. Sci. 932, 78–93. doi: 10.1111/j.1749-6632.2001.tb05799.x
Costa, L., Bauer, L., Kuperman, S., Porjesz, B., O’Connor, S., Hesselbrock, V., et al. (2000). Frontal P300 decrements, alcohol dependence, and antisocial personality disorder. Biol. Psychiatry 47, 1064–1071. doi: 10.1016/S0006-3223(99)00317-0
Cronbach, L. J., and Meehl, P. E. (1955). Construct validity in psychological tests. Psychol. Bull. 52, 281–302. doi: 10.1037/h0040957
Darley, J. M., and Latane, B. (1968). Bystander intervention in emergencies: diffusion of responsibility. J. Personal. Soc. Psychol. 8, 377–383. doi: 10.1037/h0025589
Deacon, B. J., and Baird, G. L. (2009). The chemical imbalance explanation of depression: reducing blame at what cost? J. Soc. Clin. Psychol. 28, 415–435. doi: 10.1521/jscp.2009.28.4.415
De Moor, M. H., Costa, P. T., Terracciano, A., Krueger, R. F., De Geus, E. J., Toshiko, T., et al. (2012). Meta-analysis of genome-wide association studies for personality. Mol. Psychiatry 17, 337–349. doi: 10.1038/mp.2010.128
Dennett, D. C. (1995). Darwin’s dangerous idea. Sciences 35, 34–40. doi: 10.1002/j.2326-1951.1995.tb03633.x
DeWall, C. N., Gillath, O., Pressman, S. D., Black, L. L., Bartz, J. A., Moskovitz, J., et al. (2014). When the love hormone leads to violence: oxytocin increases intimate partner violence inclinations among high trait aggressive people. Soc. Psychol. Personal. Sci. 5, 691–697. doi: 10.1177/1948550613516876
Dixon, L. (1999). Dual diagnosis of substance abuse in schizophrenia: prevalence and impact on outcomes. Schizophr. Res. 35, S93–S100. doi: 10.1016/S0920-9964(98)00161-3
Donovan, M. R., Glue, P., Kolluri, S., and Emir, B. (2010). Comparative efficacy of antidepressants in preventing relapse in anxiety disorders—a meta-analysis. J. Affect. Disord. 123, 9–16. doi: 10.1016/j.jad.2009.06.021
Dvorsky, G. (2012). 10 Reasons Why Oxytocin is the Most Amazing Molecule in the World. Daily 10. Available at: http://io9.com/5925206/10-reasons-why-oxytocin-is-the-most-amazing-molecule-in-the-world (accessed June 21, 2015).
Ekstrom, A. (2010). How and when the fMRI BOLD signal relates to underlying neural activity: the danger in dissociation. Brain Res. Rev. 62, 233–244. doi: 10.1016/j.brainresrev.2009.12.004
Entwistle, D. R. (1972). To dispel fantasies about fantasy-based measures of achievement motivation. Psychol. Bull. 77, 377–391. doi: 10.1037/h0020021
Ermer, E., Kahn, R. E., Salovey, P., and Kiehl, K. A. (2012). Emotional intelligence in incarcerated men with psychopathic traits. J. Pers. Soc. Psychol. 103, 194–204. doi: 10.1037/a0027328
Falk, R. (2014). The allusion of the gene: misunderstandings of the concepts heredity and gene. Sci. Educ. 23, 273–284. doi: 10.1007/s11191-012-9510-4
Faraone, S., and Tsuang, M. (1994). Measuring diagnostic accuracy in the absence of a gold standard. Am. J. Psychiatry 151, 650–657. doi: 10.1176/ajp.151.5.650
Fenichel, O. (1945). Neurotic acting out. Psychoanal. Rev. 32, 197–206.
Ficks, C. A., and Waldman, I. D. (2014). Candidate genes for aggression and antisocial behavior: a meta-analysis of association studies of the 5HTTLPR and MAOA-uVNTR. Behav. Genet. 44, 427–444. doi: 10.1007/s10519-014-9661-y
Fladung, A. K., Grön, G., Grammer, K., Herrnberger, B., Schilly, E., Grasteit, S., et al. (2009). A neural signature of anorexia nervosa in the ventral striatal reward system. Am. J. Psychiatry 167, 206–212. doi: 10.1176/appi.ajp.2009.09010071
Flint-Stevens, G. (1993). Applying the diagnosis of antisocial personality to imprisoned offenders: looking for hay in in a haystack. J. Offender Rehabil. 19, 1–26. doi: 10.1300/J076v19n01_01
France, C. M., Lysaker, P. H., and Robinson, R. P. (2007). The “chemical imbalance” explanation for depression: origins, lay endorsement, and clinical implications. Prof. Psychol. Res. Practice 38, 411–420. doi: 10.1037/0735-7028.38.4.411
Freud, A. (1937). The Ego and the Mechanisms of Defense, trans. C. Baines. London: Hogarth Press.
Funder, D. C. (1991). Global traits: a neo-Allportian approach to personality. Psychol. Sci. 2, 31–39. doi: 10.1111/j.1467-9280.1991.tb00093.x
Garb, H. N. (1998). Studying the Clinician: Judgment Research and Psychological Assessment. Washington, DC: American Psychological Association. doi: 10.1037/10299-000
Garb, H. N. (2003). Incremental validity and the assessment of psychopathology in adults. Psychol. Assess. 15, 508–520. doi: 10.1037/1040-35184.108.40.2068
Garfield, S. L. (1996). Some problems associated with “validated” forms of psychotherapy. Clin. Psychol. 3, 218–229. doi: 10.1111/j.1468-2850.1996.tb00073.x
Gernsbacher, M. A., Dawson, M., and Goldsmith, H. H. (2005). Three reasons not to believe in an autism epidemic. Curr. Dir. Psychol. Sci. 14, 55–58. doi: 10.1111/j.0963-7214.2005.00334.x
Gillihan, S. J., and Parens, E. (2011). Should we expect ‘neural signatures’ for DSM diagnoses? J. Clin. Psychiatry 72, 1383–1389. doi: 10.4088/JCP.10r06332gre
Glassman, W. E., and Hadad, M. (2008). Approaches to Psychology. London: Open University Press.
Goodwin, R. D., and Hamilton, S. P. (2003). Lifetime comorbidity of antisocial personality disorder and anxiety disorders among adults in the community. Psychiatry Res. 117, 159–166. doi: 10.1016/S0165-1781(02)00320-7
Grabe, M. E., and Kamhawi, R. (2006). Hard wired for negative news? Gender differences in processing broadcast news. Commun. Res. 33, 346–369. doi: 10.1177/0093650206291479
Graham, G. (2013). The Disordered Mind: An Introduction to Philosophy of Mind and Mental Illness. New York: Routledge.
Gravetter, F. J., and Forzano, L. B. (2012). Research Methods for the Behavioral Sciences, 4th Edn. Belmont, CA: Wadsworth.
Green, C. D. (1992). Of immortal mythological beasts: operationism in psychology. Theory Psychol. 2, 291–320. doi: 10.1177/0959354392023003
Green, J. P. (2003). Beliefs about hypnosis: popular beliefs, misconceptions, and the importance of experience. Int. J. Clin. Exp. Hypn. 51, 369–381. doi: 10.1076/iceh.51.4.369.16408
Green, J. P., Page, R. A., Rasekhy, R., Johnson, L. K., and Bernhardt, S. E. (2006). Cultural views and attitudes about hypnosis: a survey of college students across four countries. Int. J. Clin. Exp. Hypn. 54, 263–280. doi: 10.1080/00207140600689439
Griffiths, S. (2014). Hugs Can Make You Feel younger: ‘Cuddle Hormone’ Could Improve Bone Health, and Combat. Muscle Wasting. Available at: http://www.dailymail.co.uk/sciencetech/article-2654224/Hugs-make-feel-younger-Cuddle-hormone-improve-bone-health-combat-muscle-wasting.html#ixzz3UP2WNT2J
Guze, S. B. (1970). The need for toughmindedness in psychiatric thinking. South. Med. J. 63, 662–671. doi: 10.1097/00007611-197006000-00012
Hagger, M. S. (2014). Avoiding the “déjà-variable” phenomenon: social psychology needs more guides to constructs. Front. Psychol. 5:52. doi: 10.3389/fpsyg.2014.00052
Hare, R. D. (1991/2003). Manual for the Hare Psychopathy Checklist-Revised. Toronto, CA: Multihealth Systems.
Harris, S. (2015). “Our narrow definition of science,” in This Idea Must Die, ed. J. Brockman (New York: Harper), 136–138.
Haslam, N., Holland, E., and Kuppens, P. (2012). Categories versus dimensions in personality and psychopathology: a quantitative review of taxometric research. Psychol. Med. 42, 903–920. doi: 10.1017/S0033291711001966
Hayes, C. (2014). Multiple personality disorder: an introduction for HCAs. Br. J. Healthcare Assist. 8, 29–33. doi: 10.12968/bjha.2014.8.1.29
Honts, C. R., Raskin, D. C., and Kircher, J. C. (1994). Mental and physical countermeasures reduce the accuracy of polygraph tests. J. Appl. Psychol. 79, 252–259. doi: 10.1037/0021-9010.79.2.252
Hopper, A. N., Jamison, M. H., and Lewis, W. G. (2007). Learning curves in surgical practice. Postgrad. Med. J. 83, 777–779. doi: 10.1136/pgmj.2007.057190
Howard, B. S. (2011). Overcoming Trauma: Find Closure to the Abuse, Tragedies, and Suffering of Life. St. Louis, MO: Lifetime Media, LLC.
Hunter, E. (1951). Brain-washing in Red China: The Calculated Destruction of Men’s Minds. New York: Vanguard Press.
Huttenlocher, P. R. (2009). Neural Plasticity. Cambridge, MA: Harvard University Press.
Hyman, R. (1995). Evaluation of the program on anomalous mental phenomena. J. Parapsychol. 59, 321–352.
Jarrett, C. (2014). Great Myths of the Brain. New York: John Wiley & Sons.
Jarrett, C. (2015). Great Myths of the Brain. New York: John Wiley & Sons.
Kanazawa, S. (2008). Temperature and evolutionary novelty as forces behind the evolution of general intelligence. Intelligence 36, 99–108. doi: 10.1016/j.intell.2007.04.001
Kandler, C. (2012). Nature and nurture in personality development: the case of neuroticism and extraversion. Curr. Dir. Psychol. Sci. 21, 290–296. doi: 10.1177/0963721412452557
Kaplan, R., and Saccuzzo, D. (2012). Psychological Testing: Principles, Applications, and Issues. Belmont, CA: Cengage Learning.
Keller, L. M. (2005). Is truth serum torture? Am. Univ. Int. Law Rev. 20, 521–621.
Kelley, E. L. (1927). Interpretation of Educational Measurements. Yonkers, NY: World.
Kendler, K. S. (2005). “A gene for”: the nature of gene action in psychiatric disorders. Am. J. Psychiatry 162, 1245–1252. doi: 10.1176/appi.ajp.162.7.1243
Kendler, K. S., and Neale, M. C. (2009). “Familiality” or heritability? Arch. Gen. Psychiatry 66, 452–453. doi: 10.1001/archgenpsychiatry.2009.14
King, C. R. (2011). A novel embryological theory of autism causation involving endogenous biochemicals capable of initiating cellular gene transcription: a possible link between twelve autism risk factors and the autism ‘epidemic’. Med. Hypotheses 76, 653–660. doi: 10.1016/j.mehy.2011.01.024
Kluft, R. P. (2011). Ramifications of incest. Psychiatr. Times 27, 1–11.
Koc, K., Anik, I., Ozdamar, D., Cabuk, B., Keskin, G., and Ceylan, S. (2006). The learning curve in endoscopic pituitary surgery and our experience. Neurosurg. Rev. 29, 298–305. doi: 10.1007/s10143-006-0033-9
Koffka, K. (1922). Perception: an introduction to the Gestalt-theorie. Psychol. Bull. 19, 531–585. doi: 10.1037/h0072422
Kraft, N. H., and Keeley, J. W. (2015). “Sign versus symptom,” in Encyclopedia of Clinical Psychology, eds R. L. Cautin and S. O. Lilienfeld (New York: John Wiley & Sons), 2635–2638.
Kramer, P. D. (2011). In defense of antidepressants. New York Times. Available at: http://www.nytimes.com/2011/07/10/opinion/sunday/10antidepressants.html (access July 9, 2011).
Kringelbach, M. L., and Berridge, K. C. (2010). The functional neuroanatomy of pleasure and happiness. Discov. Med. 9, 579–587.
Krueger, J. (2001). Null hypothesis significance testing: on the survival of a flawed method. Am. Psychol. 56, 16–26. doi: 10.1037/0003-066X.56.1.16
Krueger, R. F., and Markon, K. E. (2006). Reinterpreting comorbidity: a model-based approach to understanding and classifying psychopathology. Ann. Rev. Clin. Psychol. 2, 111–133. doi: 10.1146/annurev.clinpsy.2.022305.095213
Lacasse, J. R., and Leo, J. (2005). Serotonin and depression: a disconnect between the advertisements and the scientific literature. PLoS Med. 2:e392. doi: 10.1371/journal.pmed.0020392
Lamont, L. (2008). Commonsense Part of Brain is a “Dead Zone” When Texting. Sydney Morning Herald. Available at: http://www.smh.com.au/news/specials/science/commonsense-part-of-brain-is-a-dead-zone-when-texting/2008/09/17/1221330929936.html?page
Landau, J., Garrett, J., and Webb, R. (2008). Assisting a concerned person to motivate someone experiencing cybersex into treatment. J. Marital Fam. Ther. 34, 498–511. doi: 10.1111/j.1752-0606.2008.00091.x
Latane, B., and Darley, J. (1969). Bystander ‘apathy.’Am. Sci. 57, 244–268.
Latane, B., and Rodin, J. (1969). A lady in distress: Inhibiting effects of friends and strangers on bystander intervention. J. Exp. Soc. Psychol. 5, 189–202. doi: 10.1016/0022-1031(69)90046-8
Laws, K. R. (2012). Twitter Post. Available at: http://twitter.com/Keith_Laws/statuses/163218919449962496
Leahey, T. H. (1980). The myth of operationism. J. Mind Behav. 1, 127–143.
Leo, J., and Lacasse, J. R. (2008). The media and the chemical imbalance theory of depression. Society 45, 35–45. doi: 10.1007/s12115-007-9047-3
Lieberman, J. A. (2015). Shrinks: The Untold Story of Psychiatry. New York: Little, Brown, and Company.
Lilienfeld, S. O. (2007). Cognitive neuroscience and depression: legitimate versus illegitimate reductionism and five challenges. Cogn. Ther. Res. 31, 263–272. doi: 10.1007/s10608-007-9127-0
Lilienfeld, S. O. (2010). Can psychology become a science? Personal. Individ. Diff. 49, 281–288. doi: 10.1016/j.paid.2010.01.024
Lilienfeld, S. O. (2014). The research domain criteria (RDoC): an analysis of methodological and conceptual challenges. Behav. Res. Ther. 62, 129–139. doi: 10.1016/j.brat.2014.07.019
Lilienfeld, S. O., and Arkowitz, H. (2007). Is there really an autism epidemic? Sci. Am. 17, 58–61. doi: 10.1038/scientificamerican1207-58sp
Lilienfeld, S. O., and Lynn, S. J. (2015). “Dissociative identity disorder: a scientific perspective,” in Science and Pseudoscience in Clinical Psychology, 2nd Edn, eds S. O. Lilienfeld, S. J. Lynn, and J. M. Lohr (New York: Guilford Press).
Lilienfeld, S. O., Lynn, S. J., Ruscio, J., and Beyerstein, B. L. (2009). 50 Great Myths of Popular Psychology: Shattering Widespread Misconceptions About Human Behavior. New York: Wiley.
Lilienfeld, S. O., Ritschel, L. A., Lynn, S. J., Cautin, R. L., and Latzman, R. D. (2013). Why many clinical psychologists are resistant to evidence-based practice: root causes and constructive remedies. Clin. Psychol. Rev. 33, 883–900. doi: 10.1016/j.cpr.2012.09.008
Lilienfeld, S. O., Waldman, I. D., and Israel, A. C. (1994). A critical examination of the use of the term and concept of comorbidity in psychopathology research. Clin. Psychol. 1, 71–83.
Linehan, M. M. (1989). Cognitive and behavior therapy for borderline personality disorder. Rev. Psychiatry 8, 84–102.
Loevinger, J. (1957). Objective tests as instruments of psychological theory: monograph supplement 9. Psychological. Rep. 3, 635–694. doi: 10.2466/pr0.19220.127.116.115
Lüthi, A. (2013). Sleep spindles where they come from, what they do. Neuroscientist 20, 243–256. doi: 10.1177/1073858413500854
Lykken, D. T. (1998). A Tremor in the Blood: Uses and Abuses of the Lie Detector. New York: Plenum Press.
Lynn, S. J., Kirsch, I., Knox, J., Fassler, O., and Lilienfeld, S. O. (2007). “Hypnosis and neuroscience: implications for the altered state debate,” in Hypnosis and Conscious States: The Cognitive Neuroscience Perspective, ed. G. A. Jamieson (Oxford: Oxford University Press), 145–165.
Lynn, S. J., Laurence, J.-R., and Kirsch, I. (2015). Hypnosis, suggestion, and suggestibility: an integrative model. Am. J. Clin. Hypn. 57, 314–329. doi: 10.1080/00029157.2014.976783
Macdonald, J. M. (1955). Truth serum. J. Criminal Law Criminol. Police Sci. 46, 259–263. doi: 10.2307/1139862
Mann, C. E., and Himelein, M. J. (2008). Putting the person back into psychopathology: an intervention to reduce mental illness stigma in the classroom. Soc. Psychiatry Psychiatr. Epidemiol. 43, 545–551. doi: 10.1007/s00127-008-0324-2
Markon, K. E. (2009). Hierarchies in the structure of personality traits. Soc. Personal. Psychol. Compass 3, 812–826. doi: 10.1111/j.1751-9004.2009.00213.x
McClelland, G. H., and Judd, C. M. (1993). Statistical difficulties of detecting interactions and moderator effects. Psychol. Bull. 114, 376–380. doi: 10.1037/0033-2909.114.2.376
McComas, W. F. (1996). Ten myths of science: reexamining what we think we know about the nature of science. Sch. Sci. Math. 96, 10–16. doi: 10.1111/j.1949-8594.1996.tb10205.x
McCullough, D. (2002). Interview with NEH chairman Bruce Cole. National Endowment for the Humanities. Available at: http://www.neh.gov/about/awards/jefferson-lecture/david-mccullough-interview (accessed June 23, 2015).
McDermott, R., Tingley, D., Cowden, J., Frazetto, G., and Johnson, D. (2009). The Warrior Gene’ (MAOA) predicts behavioral aggression following provocation. Proc. Natl. Acad. Sci. U.S.A. 106, 2118–2123. doi: 10.1073/pnas.0808376106
Meehl, P. E. (1945). The dynamics of “structured” personality tests. J. Clin. Psychol. 1, 296–303.
Meehl, P. E. (1969). Nuisance Variables and the Expost Facto Design. Report No. PR-69-4. Minneapolis: Department of Psychiatry, University of Minnesota.
Meehl, P. E. (1986). “Diagnostic taxa as open concepts: metatheoretical and statistical questions about reliability and construct validity in the grand strategy of nosological revision,” in Contemporary Directions in Psychopathology, eds T. Millon and G. L. Klerman (New York: Guilford Press), 215–231.
Meehl, P. E. (1995). Bootstraps taxometrics: solving the classification problem in psychopathology. Am. Psychol. 50, 266–275. doi: 10.1037/0003-066X.50.4.266
Melton, G. J. (1999). Brainwashing and the Cults: The Rise and Fall of a Theory. Available at: http://www.cesnur.org/testi/melton.htm.
Merckelbach, H., Devilly, G. J., and Rassin, E. (2002). Alters in dissociative identity disorder: metaphors or genuine entities? Clin. Psychol. Rev. 22, 481–497. doi: 10.1016/s0272-7358(01)00115-5
Merzenich, M. M. (2013). Soft-Wired: How the New Science of Brain Plasticity Can Change Your Life. San Francisco, CA: Parnassus.
Mesholam-Gately, R. I., Giuliano, A. J., Goff, K. P., Faraone, S. V., and Seidman, L. J. (2009). Neurocognition in first-episode schizophrenia: a meta-analytic review. Neuropsychology 23, 315–336. doi: 10.1037/a0014708
Messick, S. (1987). Validity. ETS Res. Rep. Ser. 2, i208. doi: 10.1002/j.2330-8516.1987.tb00244.x
Miller, G. A. (1996). How we think about cognition, emotion, and biology in psychopathology. Psychophysiology 33, 615–628. doi: 10.1111/j.1469-8986.1996.tb02356.x
Miller, G. A. (2010). Mistreating psychology in the decades of the brain. Perspect. Psychol. Sci. 5, 716–743. doi: 10.1177/1745691610388774
Miller, J. (2009). What is the probability of replicating a statistically significant effect? Psychon. Bull. Rev. 16, 617–640. doi: 10.3758/PBR.16.4.617
Miller, N., and Pedersen, W. C. (1999). Assessing process distinctiveness. Psychol. Inq. 10, 150–155. doi: 10.1207/S15327965PL100210
Morin, C. (2011). Neuromarketing: the new science of consumer behavior. Society 48, 131–135. doi: 10.1007/s12115-010-9408-1
Morse, S. (2011). Tax Compliance and the Love Molecule. Arizona State Law Journal. Available at: http://arizonastatelawjournal.org/tax-compliance-and-the-love-molecule
Mulle, J. G. (2012). Schizophrenia genetics: progress, at last. Curr. Opin. Genet. Dev. 22, 238–244. doi: 10.1016/j.gde.2012.02.011
Muller, R. (1992). Is there a neural basis for borderline splitting? Compr. Psychiatry 33, 92–104. doi: 10.1016/0010-440X(92)90004-A
Naliboff, B. D., Frese, M. P., and Rapgay, L. (2008). Mind/body psychological treatments for irritable bowel syndrome. Evid. Based Complementary Altern. Med. 5, 41–50. doi: 10.1093/ecam/nem046
Newman, D. L., Moffitt, T. E., Caspi, A., and Silva, P. A. (1998). Comorbid mental disorders: implications for treatment and sample selection. J. Abnorm. Psychol. 107, 305–311. doi: 10.1037/0021-843X.107.2.305
Nisbett, R. E., Aronson, J., Blair, C., Dickens, W., Flynn, J., Halpern, D. F., et al., (2012). Intelligence: new findings and theoretical developments. Am. Psychol. 67, 130–159. doi: 10.1037/a0026699
Olweus, D. (1977). “A critical analysis of the modern interactionist position,” in Personality at the Crossroads: Current Issues in Interactional Psychology, eds D. Magnusson and N. S. Endler (New York: Wiley), 221–233.
Pant, H., McCabe, B. J., Deskovitz, M. A., Weed, N. C., and Williams, J. E. (2014). Diagnostic reliability of MMPI-2 computer-based test interpretations. Psychol. Assess. 26, 916–924. doi: 10.1037/a0036469
Pap, A. (1958). Semantics and Necessary Truth. New Haven, CT: Yale University Press.
Pashley, M. (1994). A-level students: their problems with gene and allele. J. Biol. Educ. 28, 120–126. doi: 10.1080/00219266.1994.9655377
Pathak, A., and Srivastava, M. (2011). Narcoanalysis: a critical appraisal. Indian J. Forensic Med. Toxicol. 5, 54–57.
Pelphrey, K. A., and McPartland, J. C. (2012). Brain development: neural signature predicts autism’s emergence. Curr. Biol. 22, R127–R128. doi: 10.1016/j.cub.2012.01.025
Peter, J. P. (1981). Construct validity: a review of basic issues and marketing practices. J. Mark. Res. 18, 133–145. doi: 10.2307/3150948
Petrocelli, J. V. (2003). Hierarchical multiple regression in counseling research: common problems and possible remedies. Meas. Eval. Couns. Dev. 36, 9–22.
Pies, R. (2011). Psychiatry’s new brain-mind and the legend of the chemical imbalance. Psychiatric Times. Available at: http://www.psychiatrictimes.com/blog/couchincrisis/content/ (access July 11, 2011).
Pinker, S. (2014). The Sense of Style: The Thinking Person’s Guide to Writing in the 21st Century. New York: Penguin.
Piper, A. Jr. (1993). Truth serum and recovered memories of sexual abuse: a review of the evidence. J. Psychiatry Law 21, 447–471.
Popper, K. R. (1959). The Logic of Scientific Discovery. London: Hutchinson.
Popper, K. R. (1983). Realism and the Aim of Science. London: Routledge.
Proyer, R. T., and Häusler, J. (2007). Assessing behavior in standardized settings: the role of objective personality tests. Int. J. Clin. Health Psychol. 7, 537–546.
Radford, B. (2003). Media Mythmakers: How Journalists, Activists, and Advertisers Mislead Us. Amherst, NY: Prometheus Books.
Raz, A. (2011). Does neuroimaging of suggestion elucidate hypnotic trance? Int. J. Clin. Exp. Hypn. 59, 363–377. doi: 10.1080/00207144.2011.570682
Reynolds, C. F., Frank, E., Thase, M. E., Houck, P. R., Jennings, J. R., Howell, J. R., et al. (1988). Assessment of sexual function in depressed, impotent, and healthy men: factor analysis of a brief sexual function questionnaire for men. Psychiatry Res. 24, 231–250. doi: 10.1016/0165-1781(88)90106-0
Robinson, D. N. (1995). The logic of reductionistic models. New Ideas Psychol. 13, 1–8. doi: 10.1016/0732-118X(94)E0032-W
Rogers, R. (2003). Forensic use and abuse of psychological tests: multiscale inventories. J. Psychiatr. Practice 9, 316–320. doi: 10.1097/00131746-200307000-00008
Ross, C. A. (1994). The Osiris Complex: Case Studies in Multiple Personality Disorder. Toronto, CA: University of Toronto Press.
Rutter, M. (1994). Comorbidity: meanings and mechanisms. Clin. Psychol. 1, 100–103. doi: 10.1111/j.1468-2850.1994.tb00012.x
Sagi-Schwartz, A., Van IJzendoorn, M. H., Grossmann, K. E., Joels, T., Grossmann, K., Scharf, M., et al. (2014). Attachment and traumatic stress in female Holocaust child survivors and their daughters. Am. J. Psychiatry 160, 1086–1092. doi: 10.1176/appi.ajp.160.6.1086
Sapolsky, R. (1997). A gene for nothing. Discover 18, 40–46.
Satel, S., and Lilienfeld, S. O. (2013). Brainwashed: The Seductive Appeal of Mindless Neuroscience. New York: Basic Books.
Saxe, L., Dougherty, D., and Cross, T. (1985). The validity of polygraph testing: scientific analysis and public controversy. Am. Psychol. 40, 355–366. doi: 10.1037/0003-066X.40.3.355
Schmidt, F. L., and Hunter, J. E. (1996). Measurement error in psychological research: lessons from 26 research scenarios. Psychol. Methods 1, 199–223. doi: 10.1037/1082-989X.1.2.199
Shen, H. (2015). Neuroscience: the hard science of oxytocin. Nature 522, 410–412. doi: 10.1038/522410a
Shermer, M. (2015). “Hardwired permanent,” in This Idea Must Die, ed. J. Brockman (New York: Harper), 100–103.
Skeem, J. L., and Cooke, D. J. (2010). Is criminal behavior a central component of psychopathy? Conceptual directions for resolving the debate. Psychol. Assess. 22, 433–445. doi: 10.1037/a0008512
Skitka, L. J., Bauman, C. W., and Mullen, E. (2004). Political tolerance and coming to psychological closure following the September 11, 2001, terrorist attacks: an integrative approach. Personal. Soc. Psychol. Bull. 30, 743–756. doi: 10.1177/0146167204263968
Smith, M. (2015). Explaining Japan’s Feature Phone Fetish. Santa Clara, CA: Engadget. Available at: http://www.engadget.com/2015/03/13/japan-loves-feature-phones/
Smoller, J. W., and Finn, C. T. (2003). “Family, twin, and adoption studies of bipolar disorder,” in American Journal of Medical Genetics Part C: Seminars in Medical Genetics, Vol. 123, eds S. V. Faraone and M. T. Tsuang (New York: Wiley Subscription Services, Inc., A Wiley Company), 48–58.
Song, H., and Kwon, N. (2012). The relationship between personality traits and information competency in Korean and American students. Soc. Behav. Personal. 40, 1153–1162. doi: 10.2224/sbp.2012.40.7.1153
Spanos, N. P. (1996). Multiple Identities and False Memories: A Sociocognitive Perspective. Washington, DC: American Psychiatric Association. doi: 10.1037/10216-000
Spitzer, M., Fischbacher, U., Herrnberger, B., Grön, G., and Fehr, E. (2007). The neural signature of social norm compliance. Neuron 56, 185–196. doi: 10.1016/j.neuron.2007.09.011
Spitzer, R. L. (1994). Psychiatric “co-occurrence”? I’ll stick with “comorbidity.” Clin. Psychol. 1, 88–92.
Stanovich, K. E. (2012). How to Think Straight About Psychology. Boston, MA: Pearson Allyn and Bacon.
Stix, G. (2014). Fact or Fiction? Oxytocin is the “Love Molecule.” Scientific American. Available at: http://www.scientificamerican.com/article/fact-or-fiction-oxytocin-is-the-love-hormone (access September 8, 2014).
Streiner, D. L. (2007). A short cut to rejection: how not to write the results section of a paper. Can. J. Psychiatry 52, 385–389.
Sylvers, P., Lilienfeld, S. O., and LaPrairie, J. L. (2011). Differences between trait fear and trait anxiety: Implications for psychopathology. Clin. Psychol. Rev. 31, 122–137. doi: 10.1016/j.cpr.2010.08.004
Tavris, C., and Aronson, E. (2007). Mistakes Were Made (but not by me): Why we Justify Foolish Beliefs, Bad Decisions, and Hurtful Acts. Boston: Houghton Miﬄin Harcourt.
Taylor, A. K., and Kowalski, P. (2010). Naive psychological science: the prevalence, strength, and source of misconceptions. Psychol. Record 54, 15–25.
Thompson, B. (1989). Why won’t stepwise methods die? Meas. Eval. Couns. Dev. 21, 146–148.
Tortorella, A., Fabrazzo, M., Monteleone, A. M., Steardo, L., and Monteleone, P. (2014). The role of drug therapies in the treatment of anorexia and bulimia nervosa: a review of the literature. J. Psychopathol. 20, 50–65.
Tversky, A., and Kahneman, D. (1971). Belief in the law of small numbers. Psychol. Bull. 76, 105–110. doi: 10.1037/h0031322
Vaillant, G. E. (1977). Adaptation to Life. Cambridge, MA: Harvard University Press.
Ventegodt, S., Andersen, N. J., and Merrick, K. (2009). An ethical analysis of contemporary use of coercive persuasion (“brainwashing”,“mind control”) in psychiatry. J. Alternat. Med. Res. 1, 177–188.
Wampold, B. E., and Freund, R. D. (1987). Use of multiple regression in counseling psychology research: a flexible data-analytic strategy. J. Couns. Psychol. 34, 372–382. doi: 10.1037/0022-018.104.22.1682
Watson, D., and Clark, L. A. (1984). Negative affectivity: the disposition to experience aversive emotional states. Psychol. Bull. 96, 465–490. doi: 10.1037/0033-2909.96.3.465
Weinberger, D. A., and Gomes, M. E. (1995). Changes in daily mood and self-restraint among undercontrolled preadolescents: a time-series analysis of “acting out”. J. Am. Acad. Child Adolesc. Psychiatry 34, 1473–1482. doi: 10.1097/00004583-199511000-00014
Weinstein, H. M. (2011). Editorial note: the myth of closure, the illusion of reconciliation: final thoughts on five years as co-editor-in-chief. Int. J. Trans. Justice 5, 1–10. doi: 10.1093/ijtj/ijr002
Weiss, D. J. (1970). Factor analysis and counseling research. J. Couns. Psychol. 17, 477–485. doi: 10.1037/h0029894
Weissman, M. M. (1993). Family genetic studies of panic disorder. J. Psychiatr. Res. 27, 69–78. doi: 10.1016/0022-3956(93)90018-W
Wheeler, K. (2011). A relationship-based model for psychiatric nursing practice. Perspect. Psychiatr. Care 47, 151–159. doi: 10.1111/j.1744-6163.2010.00285.x
Wing, D. M. (1995). Transcending alcoholic denial. Image 27, 121–126. doi: 10.1111/j.1547-5069.1995.tb00834.x
Wolsko, P. M., Eisenberg, D. M., Davis, R. B., and Phillips, R. S. (2004). Use of mind–body medical therapies. J. Gen. Int. Med. 19, 43–50. doi: 10.1111/j.1525-1497.2004.21019.x
Wong, E. (2012). One Molecule for Love, Morality, and Prosperity? Slate. Available at: http://www.slate.com/articles/health_and_science/medical_examiner/2012/07/oxytocin_is_not_a_love_drug_don_t_give_it_to_kids_with_autism_.html (access July 17, 2012).
Zak, P. J. (2013). The Moral Molecule: The New Science of What Makes us Good or Evil. New York: Random House.
Zimbardo, P. G. (1997). What messages are behind today’s cults? Am. Psychol. Assoc. Monit. 28, 14.
Zurbriggen, E. L., Ramsey, L. R., and Jaworski, B. K. (2011). Self-and partner-objectification in romantic relationships: associations with media consumption and relationship satisfaction. Psychiatry 132, 911–914. doi: 10.1007/s11199-011-9933-4
Keywords: scientific thinking, misconceptions, misunderstandings, terminology, jingle and jangle fallacies
Citation: Lilienfeld SO, Sauvigné KC, Lynn SJ, Cautin RL, Latzman RD and Waldman ID (2015) Fifty psychological and psychiatric terms to avoid: a list of inaccurate, misleading, misused, ambiguous, and logically confused words and phrases. Front. Psychol. 6:1100. doi: 10.3389/fpsyg.2015.01100
Received: 13 April 2015; Accepted: 17 July 2015;
Published: 03 August 2015.
Edited by:Jason W. Osborne, University of Louisville, USA
Reviewed by:Kathy Ellen Green, University of Denver, USA
Thomas James Lundy, Independent Researcher, USA
Copyright © 2015 Lilienfeld, Sauvigné, Lynn, Cautin, Latzman and Waldman. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Scott O. Lilienfeld, Department of Psychology, Emory University, Room 473, 36 Eagle Row, Atlanta, GA 30322, USA, email@example.com