Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Sleep, 23 July 2025

Sec. Sleep, Behavior and Mental Health

Volume 4 - 2025 | https://doi.org/10.3389/frsle.2025.1625185

This article is part of the Research TopicUnderstanding conscious experiences during sleepView all articles

Dreams are more “predictable” than you think

  • 1European Commission, Joint Research Centre (JRC), Ispra, Italy
  • 2Department of Informatics, University of Sussex, Brighton, United Kingdom

Introduction: A growing body of work has used machine learning and AI tools to analyse dream reports, and compare them to other textual content. Since these tools are usually trained on text from the web, researchers have speculated they might not be suited to model dreams reports, often labeled as “unusual” and “bizarre” content.

Methods: We used a set of large language models (LLMs) to encode dream reports from DreamBank and Wikipedia. To estimate the ability of LLMs to model and predict textual reports we adopted perplexity, a measure based on entropy, formally, the exponentiated log-likelihood of a sequence. Intuitively, perplexity indicates how “surprising” a sequence of words is to a model.

Results: In most models, perplexity scores for dream reports were significantly lower than those for Wikipedia articles. Moreover, we found that perplexity scores were significantly different in reports produced by male vs female participants, and between blind and normally sighted individuals. In one case, we found this difference to be significant between clinical and healthy subjects.

Discussion: Dream reports were found to be generally easier to model and predict than Wikipedia articles. LLMs were also found to implicitly encode group differences previously observed in the literature based on gender, visual impairment, and clinical population.

1 Introduction

Dream reports describe the content of the conscious experiences we had while asleep. Through the years, researchers have used these transcripts to connect dreams with awakened states (Blagrove et al., 2004; Skancke et al., 2014; Andrews and Hanna, 2020), and to study consciousness (Nir and Tononi, 2010; Siclari et al., 2017) and pathological conditions (Kobayashi et al., 2008; Skancke et al., 2014; Thompson et al., 2015; Andrews and Hanna, 2020). For these reasons, both researchers and practitioners have been consistently interested in dream reports, and have developed a variety of frameworks to study, analyse, and annotate their content in a systematic way (Hall and Van De Castle, 1966; Hauri, 1975; Schredl, 2010).

The analysis and annotation processes of dream reports can be extremely time-consuming and rely upon human experts who usually undergo long training, which has limited the growth and reproducibility of research around dreams and dream reports (Elce et al., 2021). As a result, researchers have shown a growing interest in adopting automatic analysis of dream reports' content and structure, based on machine learning and natural language processing (NLP) (see Elce et al., 2021 for a review). Many of these approaches use models that have been fully, or partially, trained on large amounts of rather standardized text from the internet, such as Wikipedia (Nadeau et al., 2006; Razavi et al., 2013; Altszyler et al., 2017; Sanz et al., 2018; McNamara et al., 2019; Bertolini et al., 2024b,a; Cortal, 2024).

Since a vast body of work identifies dream reports as being more bizarre than wakeful experience (Rosen, 2018), one might assume that training a model on more structured and formal textual data might limit the ability of the said model to deal with reports from dreams—a position informally held by multiple researchers in the community. While the extent to which dream reports quantitatively differ from other forms of textual transcripts remains a matter of significant debate (Kahan and LaBerge, 2011; Domhoff, 2017; Zheng and Schweickert, 2023), multiple studies have indeed shown that their semantic content and word use can significantly diverge from other forms of textual items. Many of these studies are based on dictionary-based frequency analysis of content words (e.g., Bulkeley and Graves, 2018; Mallett et al., 2021; Zheng and Schweickert, 2021; Yu, 2022; Zheng and Schweickert, 2023; Zheng et al., 2024). While fully transparent and computationally efficient, dictionary-based approaches such as LIWC (Pennebaker et al., 2015) do present some critical issues (Bulkeley and Graves, 2018; Zheng and Schweickert, 2023; Bertolini et al., 2024a), such as typographical errors, or limited access to a broader context and syntactic structure. However, multiple works have shown how these methods could be used to discover differences between different types of dreams, such as nightmares, lucid dreams, and baseline dream reports (Bulkeley and Graves, 2018; Zheng and Schweickert, 2023). A partial solution was proposed by Zheng and Schweickert (2023), which expanded on the previous literature by studying the differences between dream reports and other types of textual transcripts, using both LIWC and support vector machines (SVM) (Cortes and Vapnik, 1995). The LIWC approach found a large set of categories that significantly differ between dream and non-dream reports, and the proposed SVM approach could successfully discriminate between the two categories of reports. However, the adopted dataset was quite limited in magnitude—around 800 instances, balanced between dream and non-dream reports. This constraints the generalisability of the findings, largely grounding the observed difference to the dataset of choice. Altszyler et al. (2017) introduced an approach more rooted in the overall semantic content of the textual report, by comparing two word-embedding approaches (namely Latent Semantic Analysis (LSA) (Landauer and Dumais, 1997) and word2vec's skip-gram with negative samples (Mikolov et al., 2013) to investigate how the relationship between a content word like run changes in large web corpora compared to a large collection of dream reports from DreamBank (Domhoff and Schneider, 2008). In this work the authors discovered that LSA better encodes the difference in the type of contexts such words appearing in the two types of corpora.

Since sleep and dream research are witnessing an increasing amount of NLP-based approaches, investigating whether these qualitative differences might have a quantifiable impact on NLP models is of crucial importance, as it might limit the ability of such tools to model dream reports, particularly if these methodologies utilize unsupervised techniques. This work proposes to address this specific issue directly. Unlike previous work, which focused on qualitatively identifying what content makes a (limited set of) dream and waking reports different (Zheng and Schweickert, 2023; Zheng et al., 2024), we study in a quantitative manner how much a (large) set of dream reports appears to be “surprising” to a model that has seen a huge amount of non-dream-based text. To do so, we adopt a fully unsupervised solution based on pre-trained autoregressive large language models (LLMs), and on perplexity, a popular NLP metric, intuitively indicating how well an LLM can predict a sequence of words. The proposed approach has found similar application in Colla et al. (2022) work, where authors showed how perplexity scores from GPT2 (Radford et al., 2019) and n-grams can be used to discriminate between healthy participants and patients with Alzheimer's disease.

This work makes four main contributions. First, it shows that, when considered as a continuous string of text, (a large proportion of) DreamBank is only marginally harder to predict than (a comparable section of) Wikipedia. Second, and most importantly, dream reports are on average significantly more predictable than Wikipedia articles when considered as single textual units. Third, it identifies a negative correlation between the number of words in a report/article and how “surprising” such a report/article appears to the model. Fourth, it provides preliminary evidence suggesting that gender and visual impairment can significantly impact how “surprising” a report appears to the model, providing the first evidence that modern NLP tools such as LLMs internally and implicitly replicate group differences previously observed in the literature.

2 Materials and methods

2.1 Metric and models

The primary interest of this work is to quantitatively assess whether dream reports are in fact harder to model and predict for a pre-trained large language model (LLM), the current tool of choice in most NLP research and applications. To measure this phenomenon, we adopt perplexity (PPL) (Huyen, 2019). Intuitively, perplexity can be seen as a measures of how “unpredictable” or “surprising” a given string of text is for a model. In other words, given a target word i, and a sequence of words (c, for context) preceding i, perplexity measures the ability of an LLM to predict i, given its context c. Lower the perplexity scores, higher is the ability of a model to predicting how a sentence evolves. In other words, low perplexity indicates low surprisal. Formally speaking, perplexity is the exponentiated log-likelihood of a sequence X and is computed using Equation 1:

PPL(X)=epx{-1titlogpθ(xi|x<i)},    (1)

where X = (x0, x1, ..., xt) is the sequence of words, logpθ(xi|x < i) is the log-likelihood, of the ith word conditioned by the preceding context (x < i). While many other solutions have been proposed to evaluate how well a language model can capture different linguistic phenomena, perplexity is still widely used and can inform us on how well a model reflects natural language by measuring how distant a string is to a more “natural” sequence (Meister and Cotterell, 2021). Hence, our goal can be stated as understanding whether a machine trained on a very large amount of textual data “perceives” dream reports as “surprising” (i.e., as having a high perplexity).

For models without computational constraints, perplexity should be evaluated using a sliding-window approach. This method slides the context window across the text, ensuring the model has sufficient context for each prediction. The process sums negative log-likelihoods for all word-context pairs and averages across total words, as shown in Figure 1.

Figure 1
Text displays the phrase “This is a study on dreams and LLMs” repeated eight times. Words are highlighted in alternating blue and orange colors, shifting to emphasize different words in each repetition.

Figure 1. Schematic representation of perplexity computation, using a sliding fixed window context of three words.

This approach better approximates true sequence probability decomposition and typically produces more favorable scores. However, it requires a separate forward pass for each token, making it computationally expensive. A practical solution uses strided sliding windows, moving the context by larger steps rather than single tokens. This maintains a large context while significantly reducing computation time. Following Hugging Face implementation1, we use a stride of 512 tokens with each model's maximum sequence length as input size (context plus target word). These settings surpass the results reported in the original GPT-2 papers.

This strided approach efficiently computes perplexity for large datasets that cannot fit entirely in model memory. For shorter sequences that fit within the model limits, we can process them entirely at once, obtaining a single perplexity score per sequence. Our work primarily uses this single-sequence approach, focusing on individual dream reports and Wikipedia articles. As detailed below, these texts never exceed the maximum input length for any model investigated.

To model our textual data, we adopt models from two series of autoregressive pre-trained LLMs: GPT2 and OLM0 (Groeneveld et al., 2024). The GPT2 family consists of GPT2 (137 million (M) parameters), GPT2-Medium (380 M), GPT2-Large (812 M), and GPT2-XL (1,610 M). On the other hand, the OLMo family presents two models: OLMo-1B (1,180 M), and OLMo-7B (6,890 M).

Indeed, the current landscape of autoregressive LLMs offers a suite of impressive alternatives, such as GPT-4 (OpenAI et al., 2024), Gemini (Team, 2024), or Llama 3 (Dubey et al., 2024). However, our selection of models allows us to control for multiple interesting factors, namely the impact of model size, training data, and the evolving state of the art. While it might seem an era ago—and certainly was in AI terms...—GPT2 once was (at) the pinnacle of the LLM leader-board. Indeed, OLMo's performance is not extremely representative of the state of the art. However, at its release time, it was on par with the highest-end competitors, such as Llama 2. Aside from their performance, these two families share an important factor, which makes them more suitable for our experiments than more recent and powerful models: the extent to which we know their training data. Contrary to its more recent siblings, we have quite some information on the data used to train GPT2. Most importantly, on what was not used for its training, namely, Wikipedia (Radford et al., 2019). On the other hand, and even more unusual for the current standard, OLMo's training data is fully open source. Not only do we know Wikipedia was used for its training, but we can search which articles were used in the model training. While documents from Wikipedia compose a little over .1% of the overall documents in training set (Soldaini et al., 2024), this is extremely relevant to our experiments as it allows us to frame the results of the models with respect to their pre-training procedure. Lastly, both families, which have a convenient point of contact in the two one-billion-parameters models, present a set of models growing in size, which nicely reflects the capabilities and approach of the time frame they were built in, and can allow us to study how increasing the number of parameters in a model impacts its ability to model dream and other textual data, depending on its training data. In summary, if the hypothesis that dream reports are harder to model for LLMs, we should find that average PPL scores for dream reports should be, on average, significantly higher than PPL scores for Wikipedia articles, especially in LLMs exposed to Wikipedia's articles during training.

2.2 Dataset

2.2.1 Dream dataset

Similarly to previous work (Fogli et al., 2020; Gutman Music et al., 2022; Bertolini et al., 2024a,b; Cortal, 2024), we adopt a set of dream reports extracted from DreamBank (Domhoff and Schneider, 2008)2, an online collection of dream reports from different people and scientific studies. The original dataset contains approximately 22k reports in the English language, annotated with respect to gender, year of collection, and series—the specific subsets of DreamBank representing (groups of) individuals from which dreams are collected.

2.2.2 Text dataset

We use Wikipedia as the source of our baseline text. More specifically, we consider the WikiText2 dataset (Merity et al., 2017)3, an open-source dataset containing approximately 20k articles from Wikipedia. This specific baseline choice is motivated by two main reasons. First, and more specifically to our models of choice, (part of) Wikipedia is included of OLMo's training set. Moreover, WikiText2 was entirely excluded from GPT2 pre-training, and was instead used as one of the testing benchmarks in the original paper. Second, and on a general stance, adopting Wikipedia allows for a strict comparison with a standardized text, in terms of syntactic and semantic structure. This is due to the fact that large portions of Wikipedia are formally and heavily curated, and can hence work as a “stress” test for the hypothesis that dream reports are notably different.

2.2.3 Sampling

Given the discrepancy between the datasets' magnitude and some of their specific content, we use a filtering and sampling procedure over the original datasets. We begin by filtering out from Wikipedia all those instances that do not contain an article's body—that is, instances consisting of only titles or empty strings. To limit the possibility that a variable such as the number of words might impact our experiments, we further extract from both DreamBank and Wikipedia the set of items laying that contain between 30 and 250 words. The remaining datasets consist of approximately 13 k Wikipedia articles and 17 k dream reports. To generate a test set with a similar distribution in the number of words per instance, we interactively sample a subset of dream reports of the same magnitude as the remaining Wikipedia set (i.e., 13 k) for 250 iterations. We then run a random permutation test comparing the Wikipedia set against each sample dream set and select the least diverging one. The final distributions are described in Figure 2, and are made freely available (see link in the “data availability statement” section).

Figure 2
Box plot comparing the number of words in two datasets DreamBank and WikiText2. DreamBank has a median around 100, with a range from approximately 50 to 200 words. WikiText2 has a similar range but the median is slightly higher. Both datasets show similar variability.

Figure 2. Word-count distributions of the number of words (No. Words) per instance in the final Wikipedia (WikiText2) and dream (DreamBank) test sets used in the experiments.

2.3 Statistical analyses

We compare one-dimensional distributions (e.g., how many words constitute each dream report) with a random permutation test. To assess whether two-dimensional distributions (e.g., the number of words and perplexity scores of each Wikipedia article) are significantly different from one another, we adopt the Peacock test, which is a two-dimensional non-parametric generalization of the Kolmogorov-Smirnov test (Peacock, 1983; Fasano and Franceschini, 1987). Correlation analyses are based on Spearman's coefficient. All p values in the work refer to scores obtained after applying Holm correction (Holm, 1979), a method used to adjust p-values for multiple comparisons to minimize Type I errors, by sequentially adjusting the significance threshold as a function of the number of tests performed. Experiments were run with the support of an NVIDIA H100 80GB HBM3 GPU. The code and data to replicate the experiments are freely available at https://github.com/jrcf7/report_perplexity.

3 Results

3.1 Comparing dream reports and Wikipedia articles

Table 1 gives an overview of the overall perplexities produced by the different versions of GPT2 and OLMo on the two test sets, namely DreamBank and WikiText2. The table further contains the respective lengths of the datasets, in terms of the total number of tokens, and the size of each model (in millions of learnable parameters). Based on the results in the table, we can make three main observations. First, the perplexity scores for WikiText2 from our experiments closely resemble those of the original paper that introduced the GPT2 models (Radford et al., 2019). Second, compared to DreamBank, each model seems to produce lower perplexity scores for Wikipedia. Third, while the perplexity scores produced by GPT2 on WikiText2 and DreamBank appear close to each other (31.9 vs 27.4), the distance grows with model size. Although in a smaller magnitude, a similar trend is observed for the two variants of OLMo. Interestingly, in these models, the differences between datasets are not as marked as they are for GPT2 models. This behavior is unexpected since (part of) Wikipedia is included in OLMo's training set, and should hence have a significant advantage over out-of-distribution data like dream reports. This evidence could suggest that WikiText2, or part of it, might not be part of the Wikipedia subset used to train OLMo models. Overall, the differences remain relatively small across the board of the GPT2 models. Moreover, whilst the perplexity of DreamBank is overall higher than that of WikiText2, this discrepancy might be explained by the fact that DreamBank is built from collections of very different individuals, from (very) different time periods. In other words, while Wikipedia articles tend to follow a more unified language type and structure, DreamBank's reports can suddenly and significantly vary from one line to the other.

Table 1
www.frontiersin.org

Table 1. Whole corpora results.

These results suggest that, considered as a whole corpus (i.e., a subsequent and unique string of text), Wikipedia is slightly easier to predict for all selected models. However, the main focus of our work is to understand if single dream reports are harder to model—i.e., are less predictable—than single Wikipedia articles, as these would generally be the input to any given LLM. Figure 3 offers a rather intuitive and straightforward answer to this question, by plotting the average perplexity score produced by each model (Y axis), given an instance with a defined number of words (X axis). In each diagram, the continuous blue line represents dream reports from DreamBank, while the dashed orange line represents articles from Wikipedia. Our analysis reveals that for all GPT2 models, the two two-dimensional distributions are significantly different from one another (p < 0.0001), and a random permutation test further showed that the one-dimensional distribution of the perplexity scores alone is too (p < 0.01). As the figure intuitively suggests, our analysis also conforms to the fact that, while always significant, the differences tend to fade as the model size increases. Looking at the Peacock test (Peacock, 1983; Fasano and Franceschini, 1987), we see how the score of GPT2, D=.34, slowly reduces passing from GPT2-Medium, D = 0.23, GPT2-Large, D = 0.19, and reaching D=.15 for GPT2-XL.

Figure 3
Comparison of perplexity across six models GPT2, GPT2-Medium, GPT2-Large, GPT2-XL, OLMo-1B, and OLMo-7B, shown in line graphs. Each plot displays perplexity against the number of words, with two datasets, DreamBank and WikiText2, represented by solid blue and dashed orange lines, respectively. Perplexity generally decreases as the number of words increases.

Figure 3. Perplexities by number of words per single item. Visualization of the interaction (mean and standard error, described by the shades) between the number of words (x-axis) and the perplexity scores (y-axis) produced by GPT2 for single dream reports and WikiText2 articles.

For the OLMo models, we observe a rather different trend. The two lines appear to overlap under OLMo-1B, and the GPT2 tendency seems inverted for OLMo-7B, with DreamBank's scores surpassing Wikipedia ones. This interpretation is confirmed by the statistical analysis. Under both models, we found a significant overall difference with the Peacock test (p < 0.0001). However, the random permutation analysis showed that the difference in perplexity scores is not significant for OLMo-1B. Moreover, the distance between the two distributions in OLMo-7B is notably small (D = 0.17), a rather surprising result considering that both OLMo have been exposed to (part of) Wikipedia during their training phase.

In summary, our results show that a large proportion of the LLMs under investigation found dream reports to be significantly more predictable than a more formal and structured text, such as Wikipedia articles. Most importantly, these models, namely the GPT2 ones, did not include any Wikipedia article in their training data, and hence have no advantages over dream reports. In the case of the two OLMo models, that did include some Wikipedia articles in their training data, and should hence have a clear advantage over dream reports, only the 7B version is overall better at modeling Wikipedia articles over dream reports.

3.2 Group analysis

The previous section provides consistent evidence that dream reports might be easier to model than more “standardised” strings of texts, such as Wikipedia articles. In this section, we study whether three directly measurable macro factors previously studied in the relevant literature also impact how well an LLM can model dream reports. The analysis takes into consideration five factors. The number of words per report (No. Words), year of collection, and three variables that were previously observed in the literature to produce qualitative changes in the content and structure of dream reports, namely gender, vision impairment, and clinical patients (Hall and Van De Castle, 1966; Schrdel and Reinhard, 2008; Wong et al., 2016; Kirtley, 1975; Hurovitz et al., 1999; Meaidi et al., 2014; Mota et al., 2014; Zheng et al., 2024). Lastly, this section focuses solely on GPT2 and OLMo-7B. This choice is motivated by the fact that they represent the models with the most marked preference for one of the two datasets.

As hinted by Figure 3, our analysis has found a negative correlation between the number of words per report, and perplexity scores, both for GPT2 (ρ = - 0.33) and OLMo-7B (ρ = - 0.38), and both strongly significant (p < 0.0001). Observing lower perplexity scores for larger documents is not unexpected, since predicting a given word becomes easier as the context to guess said word becomes more abundant. While rather expected and explainable, this (co)relationship is likely more complicated than expected, as suggested by the relation between perplexity and word count in dream reports from participants of different genders. As shown in Figure 4, the perplexity scores produced by participants who identify themselves as male are significantly (both p < 0.01) lower, and are hence easier to model and predict for both GPT2 and OLMo-7B. However, as clearly shown in Figure 5, these reports are also significantly (p < 0.01) shorter than the ones produced by participants who identify themself as female, as already observed in other work (e.g., Mathes and Schredl, 2013). In other words, while from a general stance, shorter reports appear to entail higher perplexity, the trend seems to invert when taking into account the gender subgroup.

Figure 4
Two side-by-side bar charts compare the perplexity scores by gender for two language models. Chart A shows GPT2 with higher perplexity for females (around 43) than males (around 38). Chart B shows OLMo-7B with higher perplexity for females (around 23) than males (around 21). Error bars indicate variability.

Figure 4. Per-report perplexities: gender. Distributions of perplexity scores obtained by GPT2 and OLMo-7B on DreamBank single dream reports grouped by the gender of the participants.

Figure 5
Bar chart showing the average number of words spoken by gender. Males, in light pink, average around 115 words. Females, in darker pink, average about 125 words. Error bars are included.

Figure 5. Per-report number of words: gender. Distributions of the word count of DreamBank single dream reports divided by the gender of the participants.

The effect of the year of data collection on the perplexity scores is also assessed with a correlation analysis. To this end, we converted the categorical framing of some instance (e.g., “1980s–1990s”), by simply finding the average a time-span (e.g., 1985). Instances with non-available dates were excluded from the analysis. The obtained dates, together with the original ones, are presented in Table 2. The result of the analysis suggests a negative correlation between the year of collection and the perplexity scores. In other words, as one might expect, reports produced in more recent years appear easier to model for GPT2, and hence tend to produce lower perplexity scores. However, while strongly significant (p < 0.0001), the effect was very weak for both GPT2 (ρ = -0.13) and OLMo-7B (ρ = -0.16).

Table 2
www.frontiersin.org

Table 2. Conversion table for DreamBank's year of collection variable.

Among DreamBank's series, there are two that collect reports from several blind participants, both males and females, for a total of 285 dream reports. To compare this restricted set of reports with the one produced by normally-sighted individuals, we have sampled a set of reports from DreamBank that has the same range of perplexity scores observed for blind participants. Just like for the general and gender-based results, these two sets show to be significantly different (p < 0.0001) when considered as two-dimensional distributions (as in Figure 3); however, when taken separately, only the perplexity score turned out to be significantly different (p < 0.01). Figure 6 summarizes the differences in the perplexity scores distributions obtained for reports produced by visually impaired and normally sighted participants. As shown, even when sampling from a limited range of items, perplexity scores for visually impaired participants are on average considerably lower and have a remarkably smaller variance, especially when encoded with GPT2.

Figure 6
Box plots labeled A and B compare the perplexity of two models, GPT2 and OLMo-7B, for impaired and normal vision groups. Both plots show higher perplexity in impaired vision (yellow box) compared to normal vision (blue box). Outliers are present in both groups, with the impaired vision group displaying greater variance in perplexity.

Figure 6. Per-report perplexities: vision impairment. Distributions of perplexity scores obtained by GPT2 and OLM0-7B on DreamBank single dream reports grouped by vision impairment of the participants.

Lastly, we consider a very small set (circa 70 instances) of reports belonging to a subject diagnosed with post-traumatic stress disorder (PTSD), a veteran of the Vietnam War. We follow the same sampling procedure and overall analysis described in the previous paragraph for the visually impaired participants, summarized in Figure 7. As suggested by the two diagrams, the difference in perplexity scores is significant only under the OLMo-7B model (p < .01). Indeed, these results are limited by the small size of sample, and are hence harder to frame and contextualize. We note that the results from GPT2 appear in line with the last experiment in Bertolini et al. (2024a), where the authors showed that a small LLM trained to classy dream for emotional content, using a report from healthy participants, performed well on this same set, despite being out of distribution. In contrast, the results from OLMo-7B appear in line with the work suggesting that clinical participants produce dream reports that significantly differ from healthy participants, as suggested by Mota et al. (2014). It is interesting to note that the work from Mota et al. (2014) largely relies on graphs-based analysis and patterns, and that transformers (Vaswani et al., 2017), the neural network at the base of most LLMs, can be considered as a special case of graph neural networks (Veličković, 2023). It is possible that a large enough model could locally and implicitly represent the same type of graph that is useful to distinguish between clinical and healthy participants.

Figure 7
Box plots comparing perplexity scores for two models, GPT2 (A) and OLMo-7B (B), based on PTSD clinical status. The “Yes” group is in orange, and the “No” group is in green. Data points show variability, with potential outliers present.

Figure 7. Per-report perplexities: clinical patient. Distributions of perplexity scores obtained by GPT2 and OLMo-7B on DreamBank single dream reports grouped by clinical condition.

4 Discussion

A growing amount of work has adopted NLP tools to investigate and annotate dream reports (see Elce et al., 2021; Bertolini et al., 2024a; Cortal, 2024 for more details). Many of these approaches rely on neural models of various dimensions, trained on large text corpora scraped from the web (Radford et al., 2019). Since a consistent body of evidence has shown that the structure and semantic content of dream reports can significantly differ from other types of textual transcripts (see Altszyler et al., 2017; Bulkeley and Graves, 2018; Zheng and Schweickert, 2023, inter alia), it is important to understand, and possibly quantify, if and how much these differences impact the ability of NLP tools to model and interpret rather specific strings of text, such as dream reports. This is especially relevant and important when adopting off-the-shelf and unsupervised models and methods, as already hinted by Bertolini et al. (2024a).

In this work, we adopted a set of large language models (LLMs) from the GPT2 and OLMo family, to investigate these issues. More specifically, we studied how well they can model and predict dream reports, compared to a more “standard” text, like Wikipedia articles, using perplexity as a measure of uncertainty. Our results have shown how most LLMs produce significantly lower perplexity scores (hence better) for single dream reports than for single Wikipedia articles. The only exceptions to this trend were observed in the two OLMo models. However, these models contained part of Wikipedia in their training data, and hence had a notable advantage. Moreover, we found only a partial significance in the smaller model (OLMo-1B), and a marginal advantage for the larger model (OLMo-7B).

These findings paint a clear picture. A picture where LLMs, or at least the ones tested in this work, do not seem to struggle at all with processing dream reports, nor do they seem more “surprised” by dream reports than they are by Wikipedia articles. In the literature, a consistent research line tends to associate dreams and their reports with bizarreness (Rosen, 2018), entailing a significant deviation from normal experience, whatever that might be. This view appears in clear contrast with our findings, as they indicate that for LLMs, dream reports are as “predictable” as the “the norm”, at least in the form of Wikipedia text. This might come as unexpected but it is likely due to what is our compass for dream reports bizarreness: reality. Aside very specific pathological cases and scenarios, we are generally capable of distinguishing a bizarre or absurd event from reality. LLMs, on the other hand, are machines designed and trained to encode or generate text, regardless of its truthiness or correctness, and can in fact frequently struggle even with identifying simple and established facts (Wang et al., 2024). This does not mean that LLMs can not or should not be used in dream research. On the contrary, our work suggests they can handle these unique strings of text. However, these results show that when using LLMs in this line of research, we should be extremely careful in projecting our definition and understanding of the mind and world onto these tools. One of these definitions might in fact be bizarreness, for which humans and LLMs might have a very different “concept”. Future work will have to focus on providing more insight into the existing relation between perplexity, or other mathematically measurable metrics, to carefully operationalised human concepts such as bizarreness, surprisal, or predicability.

Indeed, the main findings of this work suggest that dream reports are not a unique and unpredictable class of textual strings per se. However, just like Wikipedia's articles, some can be harder to model and predict. The second part of the work has hence proposed a set of analyses to understand which features of a report—or their author—might have an impact on the model of choice. The focus was on two specific models, namley GPT2 and OLMo-7B, and five variables immediately measurable from the adopted dataset: word count (No.Words), gender, year of collection, visual impairments, and mental health. We focused our attention on these models since they produced the most marked preference for dream reports or Wikipedia, while the choice of variables was based on group differences previously found in the literature.

The correlation analysis found a (rather expected) negative interaction between the number of tokens contained in a report (No.Words), and the perplexity scores produced by each model, which was also found for Wikipedia articles. However, further analyses suggested that the observed effect might be largely influenced by a consistent set of outliers, with very low perplexity scores. In other words, there appears to be another mediating variable influencing how challenging is for an LLM to model a dream report. Overall, the analysis further weakened the hypothesis that dream reports are rather unique strings of texts. All DreamBank's results, from the negative correlation to the outliers' effect and the shape of the distribution, found a strong match in the results produced by the models when tested on Wikipedia data.

The results based on gender and visual impairment further challenged the strength of the negative correlation between perplexity and word count. On the one hand, under both models, reports from blind participants did resulted in significantly lower perplexity, but no significant effect was found between the two groups in terms of reports' length. Even more strikingly, in the case of gender, the group with significantly lower perplexity scores (i.e., male) turned out to produce also significantly shorter reports. Again, these results patterns were stable across the two models. These discrepancies suggest that what really has an impact on the ability of the model to process a given report might have less to do with the number of words and more with the type of words in a report. A similar conclusion was also proposed in Bertolini et al. (2024a). Using an out-of-distribution ablation experiment, it was shown that leaving a specific DreamBank series out of training made it difficult for the model to handle a specific emotion (e.g., “happiness” for the Bea 1 series.). The authors noted that this could not be simply explained by the number of instances in the training data, and was likely related to the specific vocabulary used in that specific series to describe that particular emotion.

The work also adds more evidence to the existing body of scientific knowledge showing how the gender of a participant might impact the related dream report (Hall and Van De Castle, 1966; Schrdel and Reinhard, 2008; Wong et al., 2016; Zheng et al., 2024). While repeatedly observed, these differences were mainly constrained to a report's semantic content and/or grammatical structure, such as a reference to a specific emotion, use of violent language, or part-of-speech use. This work suggests that the observed distinction might have a very tangible effect since reports produced by male dreamers were found to be significantly easier, on average, to model by both GPT2 and OLMo. This likely suggests that the distinction is even deeper than previously noted, and might include a combination of content, vocabulary, and structure.

A possible explanation for the observed gender-based difference might come from the data used to train these models, which is largely scraped from the internet. Multiple reports and preliminary studies have identified a worldwide disproportion in internet usage that disadvantages female users (Breen et al., 2025). This disproportion might not be limited to internet usage. For instance, in 2012, a Wikipedia blogpost estimated that up to 90% of its editors were men4, a number later confirmed by a survey in 20185. More recently, researchers have used corpus-linguistics and word embedding to show that within (a large English-based corpus extracted from) the internet, the concepts of “people” and “person” do not appear to be gender neutral, but are more aligned with the concept of “men” (Bailey et al., 2022). This misalignment was also observed in machine-human interaction. A preliminary work found that ChatGPT was more frequently perceived as male rather than female on a variety of tasks (Wong and Kim, 2023). In other words, it is possible that LLMs might find male-generated dream reports easier to model and predict because they have been primarily trained on male-generated data.

Results suggesting that blind dreamers produced more predictable reports seem more difficult to frame in the current literature and knowledge. Multiple pieces of evidence across time have shown how blind participants express a significantly lower amount of visual features in their reports, predominantly presenting auditory, tactile and olfactory reference (Kirtley, 1975; Hurovitz et al., 1999; Meaidi et al., 2014; Zheng et al., 2024). However, Meaidi et al. (2014) showed that these differences can significantly vary between congenitally and late blind participants, and both series contain a mixture of congenitally and non-congenitally blind participants (although most have been for more than 20 years). A possibility might be that maintaining access to the visual modality while dreaming allows for a larger degree of abstraction and variance of dream content, leading sighted participants to generate more diverse reports, that can result in harder sequences to predict for the model. Regardless of this hypothesis, it is important to notice that, since the two series contain reports produced by several individuals—approximately thirty—with an age window spanning from 24 to 70, and remarkably different backgrounds, it is unlikely that a single participant drives the observed difference in perplexity scores.

Concerning the year of collection of each report, one might find the observed small effect as unexpected, considering that many reports were collected at a time when the internet existed only in the minds of visionary scientists and writers. However, this might be explained by the fact that the internet is a collection of extremely heterogeneous documents, that obviously include very old textual instances. It is hence possible that, while specific reports did not leak into the training data, their vocabulary and style might very well have. In other words, the model might have also been exposed to the form and vocabulary used in older reports.

Overall, we believe that this work adds an important piece of evidence to the literature investigating differences in dream experiences from different groups. We have long been aware that reports produced by participants with different gender or visual impairment tend to present significantly different content—and hence different word distributions. The experiments proposed in this work, however, further suggest that these differences are not limited to which words these groups use, but also how these groups use words, and that these differences in word usage as a measurable impact on current NLP tools.

To conclude, it is important to notice that this work has three main limitations. First of all, while OLMo training set is fully open-source, WebText, the dataset used to train GPT2, is not, and it is hence harder to estimate possible data leakage from DreamBank. That is, whether a part of the test data used in this work was also included in the training data for the model. In their work, Radford et al. (2019) note that training text for GPT2 was scraped following outbound links from Reddit, with at least 3 karma, and one link connecting Reddit to DreamBank. However, the link reached the main page of DreamBank, which does not allow scraping dream reports. As shown by example codes (e.g., here6), the main solution to acquire dream reports from DreamBank is to iteratively sample them via the random sample page, which requires actively entering specific settings—such as series or number of words—to print out a set of reports. In other words, it seems quite unlikely that a consistent part of the test data for this work was in fact also included in the training data for GPT2. Future work will have to focus on models like OLMO, where the full extent of the training data is available. This would ensure better comparison and understanding of other relevant phenomena, such as whether the difference in perlocutionary scores might be connected to a specific type of documents, like Wikipedia articles of web-scraped dialogue, and with what strength. Second, the language of tested items was limited to English. Third, the adopted dream report dataset, DreamBank, is not fully transparent about the extent to which the reports were manipulated. The extended amount of grammatical errors and informal structures/forms found upon a manual inspection of a (limited) set of reports suggested that the data went through a very limited manipulation, but this can not be widely confirmed. Future work will have to investigate how strongly these findings can be generalized to other languages and dream datasets, as well as to provide a more detailed explanation of what might make a report more complex to predict for a current LLMs, taking more into consideration semantic content and syntactic structures.

5 Conclusion

This study has provided compelling evidence that dream reports are not the unpredictable textual entities they were once thought to be. By employing a set of large language models to analyze and predict the textual content of dream reports and compare it with standardized texts from Wikipedia, the research has shown that dream reports are, on average, more predictable than Wikipedia articles. This finding challenges the assumption that dream content is too peculiar or bizarre for models trained on web-based corpora. Additionally, the study has uncovered intriguing differences in predictability related to the gender and visual impairment of dream report authors, suggesting that these factors significantly influence the language models' performance. These results not only contribute to our understanding of dream report characteristics but also have implications for the use of natural language processing tools in dream research. The insights of the presented study into the predictability of dream reports and the factors that affect it open the path for future research into the complex ways in which different groups express their dream experiences.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Ethics statement

Ethical approval was not required for the study involving humans in accordance with the local legislation and institutional requirements. Written informed consent to participate in this study was not required from the participants or the participants' legal guardians/next of kin in accordance with the national legislation and the institutional requirements.

Author contributions

LB: Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Project administration, Software, Validation, Visualization, Writing – original draft, Writing – review & editing. SC: Formal analysis, Methodology, Writing – review & editing. JW: Project administration, Supervision, Writing – review & editing.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. This research was partially conducted while the LB was at the University of Sussex. This research was partially supported by the EU Horizon 2020 project HumanE-AI (grant no. 952026).

Acknowledgments

We would like to thank the colleagues of the Digital Health Unit (JRC.F7) at the Joint Research Centre of the European Commission for their helpful guidance and support.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Gen AI was used in the creation of this manuscript.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Author disclaimer

The views expressed are purely those of the authors and may not in any circumstance be regarded as stating an official position of the European Commission.

Footnotes

References

Altszyler, E., Ribeiro, S., Sigman, M., and Slezak, D. F. (2017). The interpretation of dream meaning: resolving ambiguity using latent semantic analysis in a small corpus of text. Consciousn. Cognit. 56, 178–187. doi: 10.1016/j.concog.2017.09.004

PubMed Abstract | Crossref Full Text | Google Scholar

Andrews, S., and Hanna, P. (2020). Investigating the psychological mechanisms underlying the relationship between nightmares, suicide and self-harm. Sleep Med. Rev. 54:101352. doi: 10.1016/j.smrv.2020.101352

PubMed Abstract | Crossref Full Text | Google Scholar

Bailey, A. H., Williams, A., and Cimpian, A. (2022). Based on billions of words on the internet, PEOPLE = MEN. Sci. Adv. 8:eabm2463. doi: 10.1126/sciadv.abm2463

PubMed Abstract | Crossref Full Text | Google Scholar

Bertolini, L., Elce, V., Michalak, A., Widhoelzl, H.-S., Bernardi, G., and Weeds, J. (2024a). “Automatic annotation of dream report's emotional content with large language models,” in Proceedings of the 9th Workshop on Computational Linguistics and Clinical Psychology (CLPsych 2024), eds. A. Yates, B. Desmet, E. Prud'hommeaux, A. Zirikly, S. Bedrick, S. MacAvaney, et al. (St. Julians: Association for Computational Linguistics), 92–107.

Google Scholar

Bertolini, L., Michalak, A., and Weeds, J. (2024b). Dreamy: a library for the automatic analysis and annotation of dream reports with multilingual large language models. Sleep Med. 115, 406–407. doi: 10.1016/j.sleep.2023.11.1092

Crossref Full Text | Google Scholar

Blagrove, M., Farmer, L., and Williams, E. (2004). The relationship of nightmare frequency and nightmare distress to well-being. J. Sleep Res. 13, 129–136. doi: 10.1111/j.1365-2869.2004.00394.x

PubMed Abstract | Crossref Full Text | Google Scholar

Breen, C., Fatehkia, M., Yan, J., Zhao, X., Leasure, D. R., Weber, I., and Kashyap, R. (2025). Mapping Subnational Gender Gaps in Internet and Mobile Adoption Using Social Media Data. Center for Open Science. doi: 10.31235/osf.io/qnzsw_v2

Crossref Full Text | Google Scholar

Bulkeley, K., and Graves, M. (2018). Using the LIWC program to study dreams. Dreaming 28, 43–58. doi: 10.1037/drm0000071

Crossref Full Text | Google Scholar

Colla, D., Delsanto, M., Agosto, M., Vitiello, B., and Radicioni, D. P. (2022). Semantic coherence markers: The contribution of perplexity metrics. Artif. Intellig. Med. 134:102393. doi: 10.1016/j.artmed.2022.102393

PubMed Abstract | Crossref Full Text | Google Scholar

Cortal, G. (2024). “Sequence-to-sequence language models for character and emotion detection in dream narratives,” in Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024), eds. N. Calzolari, M. Y. Kan, V. Hoste, A. Lenci, S. Sakti, and N. Xue (Torino: ELRA and ICCL), 14717–14728.

Google Scholar

Cortes, C., and Vapnik, V. (1995). Support-vector networks. Mach. Learn. 20, 273–297. doi: 10.1007/BF00994018

Crossref Full Text | Google Scholar

Domhoff, G. W. (2017). The Emergence of Dreaming: Mind-Wandering, Embodied Simulation, and the Default Network. Oxford: Oxford University Press.

Google Scholar

Domhoff, G. W., and Schneider, A. (2008). Studying dream content using the archive and search engine on DreamBank.net. Consciousn. Cognit. 17, 1238–1247. doi: 10.1016/j.concog.2008.06.010

PubMed Abstract | Crossref Full Text | Google Scholar

Dubey, A., Jauhri, A., Pandey, A., Kadian, A., Al-Dahle, A., Letman, A., et al. (2024). The LLAMA 3 Herd of Models. CoRR abs/2407.21783. doi: 10.48550/arXiv.2407.21783

Crossref Full Text | Google Scholar

Elce, V., Handjaras, G., and Bernardi, G. (2021). The language of dreams: application of linguistics-based approaches for the automated analysis of dream experiences. Clocks & Sleep 3, 495–514. doi: 10.3390/clockssleep3030035

PubMed Abstract | Crossref Full Text | Google Scholar

Fasano, G., and Franceschini, A. (1987). A multidimensional version of the kolmogorov test. Monthly Notices Royal Astronom. Soc. 225, 155–170. doi: 10.1093/mnras/225.1.155

PubMed Abstract | Crossref Full Text | Google Scholar

Fogli, A., Aiello, L. M., and Quercia, D. (2020). Our dreams, our selves: automatic analysis of dream reports. Royal Soc. Open Sci.7:192080. doi: 10.1098/rsos.192080

PubMed Abstract | Crossref Full Text | Google Scholar

Gemini Team Google: Petko Georgiev, Lei, V. I., Burnell, R., Bai, L., Gulati, A., Tanzer, G., et al. (2024). Gemini 1.5: unlocking multimodal understanding across millions of tokens of context. CoRR abs/2403.05530. doi: 10.48550/arXiv.2403.05530

Crossref Full Text | Google Scholar

Groeneveld, D., Beltagy, I., Walsh, P., Bhagia, A., Kinney, R., Tafjord, O., et al. (2024). OLMo: Accelerating the science of language models. arXiv [Preprint]. doi: 10.18653/v1/2024.acl-long.841

PubMed Abstract | Crossref Full Text | Google Scholar

Gutman Music, M., Holur, P., and Bulkeley, K. (2022). Mapping dreams in a computational space: A phrase-level model for analyzing fight/flight and other typical situations in dream reports. Consciousn. Cognit. 106:103428. doi: 10.1016/j.concog.2022.103428

PubMed Abstract | Crossref Full Text | Google Scholar

Hall, C. S., and Van De Castle, R. L. (1966). The Content Analysis of Dreams. Norwalk, CT: Appleton-Century-Crofts.

Google Scholar

Hauri, P. (1975). “Categorization of sleep mental activity for psychophysiological studies,” in The Experimental Study of Sleep: Methodological Problems, 271–281.

PubMed Abstract | Google Scholar

Holm, S. (1979). A simple sequentially rejective multiple test procedure. Scand. J. Statist. 6, 65–70.

Google Scholar

Hurovitz, C. S., Dunn, S., Domhoff, G. W., and Fiss, H. (1999). The dreams of blind men and women: A replication and extension of previous findings. Dreaming 9, 183–193. doi: 10.1023/A:1021397817164

Crossref Full Text | Google Scholar

Huyen, C. (2019). Evaluation Metrics for Language Modeling. Stanford, CA: The Gradient.

Google Scholar

Kahan, T. L., and LaBerge, S. P. (2011). Dreaming and waking: similarities and differences revisited. Consciousn. Cognit. 20, 494–514. doi: 10.1016/j.concog.2010.09.002

PubMed Abstract | Crossref Full Text | Google Scholar

Kirtley, D. D. (1975). The Psychology of Blindness. Chicago, IL: Nelson-Hall.

Google Scholar

Kobayashi, I., Sledjeski, E. M., Spoonster, E., Fallon Jr, W. F., and Delahanty, D. L. (2008). Effects of early nightmares on the development of sleep disturbances in motor vehicle accident victims. J. Traumatic Stress 21, 548–555. doi: 10.1002/jts.20368

PubMed Abstract | Crossref Full Text | Google Scholar

Landauer, T. K., and Dumais, S. T. (1997). A solution to plato's problem: the latent semantic analysis theory of acquisition, induction, and representation of knowledge. Psychol. Rev. 104, 211–240. doi: 10.1037/0033-295X.104.2.211

Crossref Full Text | Google Scholar

Mallett, R., Picard-Deland, C., Pigeon, W., Wary, M., Grewal, A., Blagrove, M., and Carr, M. (2021). The relationship between dreams and subsequent morning mood using self-reports and text analysis. Affect. Sci. 3, 400–405. doi: 10.1007/s42761-021-00080-8

PubMed Abstract | Crossref Full Text | Google Scholar

Mathes, J., and Schredl, M. (2013). Gender differences in dream content: Are they related to personality? Int. J. Dream Res. 6, 104–109. doi: 10.11588/ijodr.2013.2.10954

Crossref Full Text | Google Scholar

McNamara, P., Duffy-Deno, K., Marsh, T., and Marsh, T. (2019). Dream content analysis using artificial intelligence. Int. J. Dream Res. 12:1. doi: 10.11588/ijodr.2019.1.48744

Crossref Full Text | Google Scholar

Meaidi, A., Jennum, P., Ptito, M., and Kupers, R. (2014). The sensory construction of dreams and nightmare frequency in congenitally blind and late blind individuals. Sleep Med. 15, 586–595. doi: 10.1016/j.sleep.2013.12.008

PubMed Abstract | Crossref Full Text | Google Scholar

Meister C. and Cotterell, R.. (2021). Language model evaluation beyond perplexity. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers) (Online: Association for Computational Linguistics), 5328–5339. doi: 10.18653/v1/2021.acl-long.414

PubMed Abstract | Crossref Full Text | Google Scholar

Merity, S., Xiong, C., Bradbury, J., and Socher, R. (2017). “Pointer sentinel mixture models,” in 5th International Conference on Learning Representations (ICLR), 1–13. Available online at: https://openreview.net/forum?id=Byj72udxe

Google Scholar

Mikolov, T., Chen, K., Corrado, G., and Dean, J. (2013). Efficient Estimation of Word Representations in Vector Space.

PubMed Abstract | Google Scholar

Mota, N. B., Furtado, R., Maia, P. P. C., Copelli, M., and Ribeiro, S. (2014). Graph analysis of dream reports is especially informative about psychosis. Scient. Reports 4:1. doi: 10.1038/srep03691

PubMed Abstract | Crossref Full Text | Google Scholar

Nadeau, D., Sabourin, C., Koninck, J. D., Matwin, S., and Turney, P. D. (2006). “Automatic dream sentiment analysis,” in Proc. of the Workshop on Computational Aesthetics at the Twenty-First National Conf. on Artificial Intelligence (Washington, DC: AAAI).

Google Scholar

Nir, Y., and Tononi, G. (2010). Dreaming and the brain: from phenomenology to neurophysiology. Trends Cognit. Sci. 14, 88–100. doi: 10.1016/j.tics.2009.12.001

PubMed Abstract | Crossref Full Text | Google Scholar

OpenA, I., Achiam, J., Adler, S., Agarwal, S., Ahmad, L., Akkaya, I., et al. (2024). Gpt-4 Technical Report.

Google Scholar

Peacock, J. A. (1983). Two-dimensional goodness-of-fit testing in astronomy. Monthly Notices Royal Astronom. Soc. 202, 615–627. doi: 10.1093/mnras/202.3.615

PubMed Abstract | Crossref Full Text | Google Scholar

Pennebaker, J. W., Boyd, R. L., Jordan, K., and Blackburn, K. (2015). “The development and psychometric properties of LIWC2015,” in Technical Report. Austin TX: University of Texas at Austin.

PubMed Abstract | Google Scholar

Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., and Sutskever, I. (2019). “Language models are unsupervised multitask learners,” in OpenAI Blog.

PubMed Abstract | Google Scholar

Razavi, A. H., Matwin, S., Koninck, J. D., and Amini, R. R. (2013). Dream sentiment analysis using second order soft co-occurrences (SOSCO) and time course representations. J. Intellig. Inform. Syst. 42, 393–413. doi: 10.1007/s10844-013-0273-4

Crossref Full Text | Google Scholar

Rosen, M. G. (2018). How bizarre? A pluralist approach to dream content. Consciousn. Cognit. 62, 148–162. doi: 10.1016/j.concog.2018.03.009

PubMed Abstract | Crossref Full Text | Google Scholar

Sanz, C., Zamberlan, F., Erowid, E., Erowid, F., and Tagliazucchi, E. (2018). The experience elicited by hallucinogens presents the highest similarity to dreaming within a large database of psychoactive substance reports. Front. Neurosci. 12:7. doi: 10.3389/fnins.2018.00007

PubMed Abstract | Crossref Full Text | Google Scholar

Schrdel, M., and Reinhard, I. (2008). Gender differences in dream recall: a meta-analysis. J. Sleep Res. 17, 125–131. doi: 10.1111/j.1365-2869.2008.00626.x

PubMed Abstract | Crossref Full Text | Google Scholar

Schredl, M. (2010). Dream content analysis: Basic principles. Int. J. Dream Res. 3:1. doi: 10.11588/ijodr.2010.1.474

Crossref Full Text | Google Scholar

Siclari, F., Baird, B., Perogamvros, L., Bernardi, G., LaRocque, J. J., Riedner, B., et al. (2017). The neural correlates of dreaming. Nat. Neurosci. 20, 872–878. doi: 10.1038/nn.4545

PubMed Abstract | Crossref Full Text | Google Scholar

Skancke, J. F., Holsen, I., and Schredl, M. (2014). Continuity between waking life and dreams of psychiatric patients: a review and discussion of the implications for dream research. Int. J. Dream Res. 7, 39–53. doi: 10.11588/ijodr.2014.1.12184

Crossref Full Text | Google Scholar

Soldaini, L., Kinney, R., Bhagia, A., Schwenk, D., Atkinson, D., Authur, R., et al. (2024). Dolma: 559 an open corpus of three trillion tokens for language model pretraining research.

Google Scholar

Thompson, A., Lereya, S. T., Lewis, G., Zammit, S., Fisher, H. L., and Wolke, D. (2015). Childhood sleep disturbance and risk of psychotic experiences at 18: UK birth cohort. Br. J. Psychiat. 207, 23–29. doi: 10.1192/bjp.bp.113.144089

PubMed Abstract | Crossref Full Text | Google Scholar

Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., et al. (2017). “Attention is all you need,” in Advances in Neural Information Processing Systems, eds. I. Guyon, U. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett (Red Hook, NY: Curran Associates, Inc).

Google Scholar

Veličković, P. (2023). Everything is connected: Graph neural networks. Curr. Opin. Struct. Biol. 79:102538. doi: 10.1016/j.sbi.2023.102538

PubMed Abstract | Crossref Full Text | Google Scholar

Wang, Y., Wang, M., Manzoor, M. A., Liu, F., Georgiev, G. N., Das, R. J., and Nakov, P. (2024). “Factuality of large language models: A survey,” in Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, eds. Y. Al-Onaizan, M. Bansal, and Y. N. Chen (Miami, FL: Association for Computational Linguistics), 19519–19529.

Google Scholar

Wong, C., Amini, R., and Koninck, J. D. (2016). Automatic gender detection of dream reports: A promising approach. Consciousn. Cognit. 44, 20–28. doi: 10.1016/j.concog.2016.06.004

PubMed Abstract | Crossref Full Text | Google Scholar

Wong, J., and Kim, J. (2023). ChatGPT is more likely to be perceived as male than female. arXiv [Preprint]. Available online at: https://arxiv.org/abs/2305.12564

Google Scholar

Yu, C. K.-C. (2022). Automated analysis of dream sentiment royal road to dream dynamics? Dreaming 32, 33–51. doi: 10.1037/drm0000189

Crossref Full Text | Google Scholar

Zheng, X., and Schweickert, R. (2021). Comparing hall van de castle coding and linguistic inquiry and word count using canonical correlation analysis. Dreaming 31, 207–224. doi: 10.1037/drm0000173

Crossref Full Text | Google Scholar

Zheng, X., and Schweickert, R. (2023). Differentiating dreaming and waking reports with automatic text analysis and support vector machines. Conscious. Cognit. 107:103439. doi: 10.1016/j.concog.2022.103439

PubMed Abstract | Crossref Full Text | Google Scholar

Zheng, X., Schweickert, R., and Song, M. (2024). Automatic dream content analysis finds effects of gender, age, and blindness on word use. Dreaming. 35, 68–85. doi: 10.1037/drm0000287

Crossref Full Text | Google Scholar

Keywords: dream report analysis, dream reports modeling, gender difference, dreaming in blind participants, machine learning, large language models, natural language processing

Citation: Bertolini L, Consoli S and Weeds J (2025) Dreams are more “predictable” than you think. Front. Sleep 4:1625185. doi: 10.3389/frsle.2025.1625185

Received: 08 May 2025; Accepted: 30 June 2025;
Published: 23 July 2025.

Edited by:

Claudia Picard-Deland, Montreal University, Canada

Reviewed by:

Don Kuiken, University of Alberta, Canada
Kristoffer Appel, Institute of Sleep and Dream Technologies, Germany

Copyright © 2025 Bertolini, Consoli and Weeds. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Lorenzo Bertolini, bG9yZW56by5iZXJ0b2xpbmlAZWMuZXVyb3BhLmV1

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.