Skip to main content

ORIGINAL RESEARCH article

Front. Artif. Intell., 23 February 2022
Sec. Language and Computation
Volume 5 - 2022 | https://doi.org/10.3389/frai.2022.796793

Exploring the Representations of Individual Entities in the Brain Combining EEG and Distributional Semantics

  • Cognitive Science Research Group, School of Electronic Engineering and Computer Science, Queen Mary University of London, London, United Kingdom

Semantic knowledge about individual entities (i.e., the referents of proper names such as Jacinta Ardern) is fine-grained, episodic, and strongly social in nature, when compared with knowledge about generic entities (the referents of common nouns such as politician). We investigate the semantic representations of individual entities in the brain; and for the first time we approach this question using both neural data, in the form of newly-acquired EEG data, and distributional models of word meaning, employing them to isolate semantic information regarding individual entities in the brain. We ran two sets of analyses. The first set of analyses is only concerned with the evoked responses to individual entities and their categories. We find that it is possible to classify them according to both their coarse and their fine-grained category at appropriate timepoints, but that it is hard to map representational information learned from individuals to their categories. In the second set of analyses, we learn to decode from evoked responses to distributional word vectors. These results indicate that such a mapping can be learnt successfully: this counts not only as a demonstration that representations of individuals can be discriminated in EEG responses, but also as a first brain-based validation of distributional semantic models as representations of individual entities. Finally, in-depth analyses of the decoder performance provide additional evidence that the referents of proper names and categories have little in common when it comes to their representation in the brain.

1. Introduction

As the idiom goes, people and places can be one of a kind—but could it be that our brains actually mean it? Thinking about Jacinda Ardern involves inevitably bringing to mind the information that she's a politician, aside from her face and her voice. But together with this may come to mind much more: other fellow politicians, other people she is often portrayed with, or places where she's often found at. This bundle of information encompasses disparate pieces of knowledge about the person herself, her kind (or category), other people belonging to the same kind, and other parts of the world which are associated with her. But how is this bundle structured, and how are the various bits wound together in the brain?

It has been shown in cognitive neuroscience that relevant differences exist between the processing of semantic information regarding unique, individual entities, which are entities indicated by proper names such as people and places, and that related to generic entities, such as referents of common nouns (Semenza and Zettin, 1989; Gorno-Tempini and Price, 2001; Semenza, 2006, 2009; Martins and Farrajota, 2007; Olson et al., 2013; Fairhall et al., 2014; Brédart, 2017; Schneider et al., 2018; Morton et al., 2021). This distinction reflects two ontological distinctions. The first one holds between instances and categories, with roots in philosophy (Lowe, 2003; Murez and Recanati, 2016), formal linguistics (Carlson and Pelletier, 1995) and cognitive psychology (Rosch, 1975; Kahneman et al., 1992; Leslie et al., 1998; Carey and Xu, 2001). Intuitively, instances are entities which are perceived as unique, whereas categories are classes of individuals, grouped together in order to distinguish them according to their class, and not to their individual identity (Klapper et al., 2017). This is the sense in which categories and categorization will be intended in this work. The second distinction is a further qualification of the status of instances. It starts from the observation that whereas certain instances are generally given a proper name (for example people, places, monuments, pets), others typically are not (for instance, instances of chairs, pans, windows). The availability of a proper name as a label, according to this theory, reflects cognitive and social constraints: only individual, unique entities which are sufficiently cognitively and socially salient can receive a proper name (Strawson, 1950; Kripke, 1972; Jeshion, 2009). This is the sense in which we refer to individual entities.

Semantics investigations in cognitive neuroscience have mostly focused on generic entities, and individual entities are usually treated as special cases: early neuroimaging studies found a strong involvement of the anterior temporal lobes (ATLs) when processing individual entities (Gorno-Tempini et al., 1998; Grabowski et al., 2001), but the ATLs have since clearly emerged as a hub for semantic processing in general (Ralph et al., 2017). This puts into question the existence of separate loci of processing for individual and generic entities, as processing of the referents of proper names may just require more (and more wide-spread) resources, because of their high specificity (Borghesani et al., 2019), and because of their social and emotional features (Olson et al., 2007, 2013). Research on individual entities has then focused on finding the neural correlates of a possible supramodal representation of individual entities, specific to this kind of entity (Fairhall et al., 2014; Schneider et al., 2018; Tsantani et al., 2019); or, restricting the analysis to people, on understanding timing and location of uni- and multi- modal processes such as face and voice recognition (Campanella and Belin, 2007; Anzellotti and Caramazza, 2017; Young et al., 2020); on trying to tease apart the processes related to social and general semantic cognition (Olson et al., 2013; Rice et al., 2018; Binney and Ramsey, 2020); on the structure of the representations of people, by comparing associative and categorical priming for faces and names (Schweinberger, 1996; Wiese and Schweinberger, 2008; Wiese, 2011). The results in the literature with respect to this last question are contradictory, as it remains unclear whether the categorical priming effect is weaker (Young et al., 1994; Barry et al., 1998; Carson and Mike Burton, 2001; Vladeanu et al., 2006; Bazzanella and Bouquet, 2011; Germain-Mondon et al., 2011) or on a par with associative priming (Darling and Valentine, 2005; Stone and Valentine, 2007; Stone, 2008; Wiese and Schweinberger, 2011). More in general, it remains open to debate whether categorical information plays a significant role in the structuring the semantic representations of individuals, or not; and if it does, to what extent (Turk et al., 2005).

In artificial intelligence, and in particular in computational linguistics and NLP, recent advances have provided researchers with models, based on distributional properties of words in texts, which capture very subtle semantic knowledge (Erk, 2012; Camacho-Collados and Pilehvar, 2018). In particular, knowledge about individual entities can be now modeled in much greater detail and with impressive results in NLP benchmarks (Lenci et al., 2021). These representations of individual entities, however, have mostly been tested on traditional NLP tasks (Chen et al., 2019). Unlike with generic concepts (Mandera et al., 2017), very little attention has been paid to how well these models capture human processing of the semantics of individual entities. Distributional models of individual entities could, on the one hand, benefit from cognitive neuroscience research, which could provide a guide and an evaluation for the models (Günther et al., 2019; Hollenstein et al., 2019); and on the other, help it, by offering powerful vectorial models which can isolate processing of individual entities in the brain (Bruffaerts et al., 2019).

In this work we bring together brain data (in the form of electroencephalography, or EEG, data) and computational, distributional models of semantics. From the point of view of neuroscience, we investigate the way in which entity-exclusive and categorical (relating the entity to its category) semantic knowledge structures the semantic representations of both individual entities and their categories. From the point of view of computational linguistics, we look for the first time at whether distributional models encode semantic information about individual entities as it is processed in the brain.

In order to do this, we collected EEG data from 33 participants using linguistic stimuli. During the experiment we recorded responses to proper names of famous people and places, belonging to different categories (e.g., “Scarlett Johansson,” “Eiffel Tower”), as well as to the common nouns referring to their socially-relevant classes (e.g., “actor,” “monument”). This matched set of stimuli allowed us to investigate four questions: whether we could find distinctive signatures for each individual entity in the brain; to what extent these representations were shaped according to category; whether the evoked responses to individual entities and categories shared representational information, as picked up by machine learning models; and finally, to what extent distributional models map onto brain processing of individual entities.

We approach our research questions in two ways: one based on classification of EEG responses according to their category, both at a coarse- and fine-grained level of categorization; and another one based on learning regression models that perform decoding from brain data to word vectors (learning a linear map from brain activations to the true values of the word vectors' dimensions). We show that it is possible to obtain above-chance performance in both cases, finding traces of individualization and categorization, overcoming the high noise present in the EEG data. This also confirms that distributional models of individual entities, despite relying exclusively on textual data, encapsulate information matching brain processing. Furthermore, results indicate that evoked responses to nouns for categories and proper names for their instances have little representational information in common.

2. Background

2.1. Distributional Semantics

Distributional models of semantics represent the meaning of words in the form of vectors, by looking at co-occurrences between words as they are found in large collections of texts, called corpora (Lenci, 2008; Boleda, 2020).

The theoretical underpinning of these models dates back to Wittgenstein's considerations on language use (Wittgenstein, 1953). In his view, an important part of a word's meaning can be understood by looking at the way in which it is used in actual language. This so-called distributional hypothesis is best-known in the version formulated by Firth (1957): “You shall know a word by the company it keeps”–i.e., words which are found in similar contexts have similar meaning (see also Harris, 1954).

Vector-space models of lexical meaning based on the distributional hypothesis started to appear, and to be evaluated as cognitive models of lexical semantics, little more than 20 years ago, when both computational power and the availability of textual resources improved dramatically (Lund and Burgess, 1996; Landauer and Dumais, 1997; Schütze, 1997). The success of these models in, e.g., predicting human synonymy patterns (Turney, 2001) motivated a great deal of research (Lin, 1998; Finkelstein et al., 2001; Curran and Moens, 2002; Almuhareb and Poesio, 2004; Agirre and Edmonds, 2007; Bullinaria and Levy, 2007; Padó and Lapata, 2007; Baroni and Lenci, 2010).

After a first period in which the dominant paradigm was to learn such models out of word co-occurrences, the field moved toward deriving distributional models of words as by-products of neural networks whose objective was to learn language models (Mikolov et al., 2013; Baroni et al., 2014; Pennington et al., 2014), which is currently the dominant paradigm. These models frame word vector learning as a machine learning problem. The goal is that of learning, by looking at corpora, to predict word vectors such that words with similar meanings should have similar vectors, and vice versa. The latest models, which use deep neural networks (Peters et al., 2018; Devlin et al., 2019), take into account the fact word meaning changes slightly depending on the context where a word is found. In these models, there are no “static” word vectors; instead, they provide, given a sentence or a paragraph, word vectors specific to that context—and because of this, they are often called “contextualized” language models.

All along, it has been shown that these models capture semantic information about words as it is stored and processed in human cognition (Bruffaerts et al., 2019). This is quite surprising, given that these models rely exclusively on textual data, whereas humans have much broader sources of information—for recent overviews of what phenomena can be modeled, see Günther et al. (2019) and Lenci et al. (2021). In particular, research in cognitive neuroscience has shown the important role of sensory information in brain semantic representations (Barsalou, 2008; Ralph et al., 2017). In order to account for this type of signal, which cannot be found directly in text, another kind of distributional semantics models has been developed, adding also visual (Bruni et al., 2012) and auditory (Kiela and Clark, 2015) features to the vectors created from corpora. These multimodal distributional semantics models can, as expected, improve results on tasks involving concrete concepts (Bruni et al., 2014).

2.2. Cognitive Data and Distributional Semantics

Much early work on distributional models was evaluated in a purely qualitative matter, by showing that the lexical vectors clustered “in an intuitive way.” This informal sort of evaluation however was quickly replaced by attempts to introduce more quantitative forms of evaluation. One popular approach was to extract from a lexical database such as WordNet (Fellbaum, 1998) test words belonging to different classes (e.g., animals, tools) and then evaluate the extent to which the learned vectors clustered into clusters matching the original classes (Lin, 1998; Curran and Moens, 2002; Almuhareb and Poesio, 2004). However, this approach to evaluation made the results entirely depend on the cognitive plausibility of the classes in the target lexical database, which was problematic. Thus, while the knowledge-based approach was not completelely abandoned, it was quickly supplemented with forms of evaluation which tested more directly the extent to which the learned representations encoded linguistic and conceptual knowledge.

One type of approach is to use distributional models to predict human behavior in tests such as the already mentioned TOEFL test (Turney, 2001). In fact, this is the form of evaluation most used in the psychological literature from which distributional models originated (Lund and Burgess, 1996; Landauer and Dumais, 1997). Benchmarks for distributional models combining clustering evaluation with tests were proposed, e.g., by Baroni et al. (2010, 2014), and Lenci et al. (2021).

Another evaluation framework is that of similarity and relatedness tasks: given a set of words (for instance, concrete or abstract nouns), measures of similarity or relatedness between all possible pairs of words in the list are first elicited from human subjects (as simple examples, “cheese” and “yogurt” are very similar, whereas “cheese” and “France” are strongly related) and, in parallel, from a word vector model. For humans, some quantitative scale is used, whereas, for distributional models, a vector distance measure such as cosine similarity among word vectors is employed. Finally, the correlation between human judgments and the model's matched predictions is computed—the assumption being that the higher the distributional model's correlation with human ratings, the higher its quality (Finkelstein et al., 2001; Bruni et al., 2012; Hill et al., 2015). Yet another method is based on quantifying the models' ability to replicate cognitive phenomena such as priming and reading times (Jones and Mewhort, 2007; Mandera et al., 2017; Günther et al., 2019).

An exciting alternative to this approach was evaluating distributional models using evidence about semantic categorization in the brain (Warrington and Shallice, 1984; Caramazza and Shelton, 1998; Haxby et al., 2001). One of the earliest proposals exploring both brain data and distributional models of word meaning was Mitchell et al. (2008). That paper showed that, for a selected set of common nouns referring to concrete concepts, co-occurrence patterns as extracted from corpora could successfully predict brain activity evoked by drawings of the referent of the words and recorded with fMRI. fMRI stands for functional magnetic resonance imaging, and is a technique which provides images of neuronal activity as reflected by cerebral blood flow (Friston et al., 1998).

The approach of Mitchell et al. (2008), called brain encoding, was interesting for a number of reasons. First of all, it indicated that purely textual co-occurrence captured important features of semantic processing in the brain. Secondly, it pioneered an approach to isolate brain signatures of semantic processing of individual words using computational linguistics models—essentially turning what could be seen as a classification, categorical problem (recognizing the word which evoked the brain activity) to a regression problem (finding a mapping between one vectorial representation, previously obtained by way of a model, to another, the brain data). Finally, this work presented a new way of evaluating vectorial models of word meaning as created in computational linguistics, looking at how well they modeled brain processing. This effectively provided one of the most direct possible evaluations of the cognitive plausibility of the models.

After, Mitchell et al. (2008), several lines of work expanded the approach further. One approach focused on using different models of semantics on the same set of fMRI images provided by Mitchell et al. (2008). The goal was to try to find if, and how, performance differed across models: distributional models other than the ones originally used (Murphy et al., 2012), models incorporating knowledge base information (Jelodar et al., 2010), models based on Wikipedia definitions (Pereira et al., 2013), multimodal models incorporating both textual and visual features (Anderson et al., 2013), models based on word associations from a thesaurus (Akama et al., 2015). Another line of research involved obtaining original fMRI data, applying a similar encoding analysis, but widening the scope of the approach to other linguistic phenomena: phrases (Chang et al., 2009), sentences (Anderson et al., 2021), naturalistic processing of visually (Wehbe et al., 2014a) and orally (Huth et al., 2016; Zhang et al., 2020) presented stories.

A mirror approach was also proposed, that of decoding: the mapping is learnt from the fMRI brain data to the word vectors. In Anderson et al. (2017) the authors showed that not only representations of concrete concepts, but also abstract ones, could be successfully decoded to distributional word vectors. And in Pereira et al. (2018) a much more extensive set of stimuli was used, encompassing both abstract and concrete concepts, presented as words, definitional sentences and pictures during the fMRI scans. Expanding even further the use of word vectors to isolate semantic processing in the brain, Djokic et al. (2020) found evidence, through decoding to word vectors, that metaphorical and literal readings of sentences have different brain signatures, and Nishida and Nishimoto (2018) showed that it was possible to learn a mapping from videos (and their matched descriptive textual annotations) to distributional word vectors.

2.3. EEG and Distributional Semantics

Another type of brain data, whose main members are electroencephalography (EEG) and magnetoencephalography (MEG), has also been used for brain encoding/decoding studies from/to computational word vectors. They measure different signals coming from the brain (electrical fields for EEG, and magnetic fields for MEG), but their source is the same—ionic currents generated by biochemical processes in neurons (da Silva, 2013).

Although research on using EEG to evaluate distributional models started at about the same time as research using fMRI (Murphy et al., 2008, 2011), they have been used to a lesser extent, because of both conceptual and practical issues. fMRI captures extremely detailed brain images every 1 or 2 seconds. These are the ideal form of concept representation as retrieved from semantic memory, and can be straightforwardly matched to word vectors, which are taken to model lexical items in semantic memory. Instead, M/EEG provide snapshots of brain processing which have much better time resolution—in the milliseconds range—but poorer spatial definition. This is especially true in the case of EEG, which not only has the lowest spatial resolution of all, but also the lowest signal-to-noise ratio. Conceptually, word vectors do not model the temporal dimension of semantic processing, but rather the spatial one, in the form of distributed vector spaces: and spatial analyses in cognitive neuroscience research have traditionally involved primarily fMRI, due to its high spatial resolution (Kemmerer, 2014).

Both EEG and MEG come with their own advantages, however. EEG is particularly cheap and portable, two reasons which make it also the preferred choice for brain-computer interfaces (Allison et al., 2020) and for working with elderly patients (Gu et al., 2013); and, provided one uses enough channels, it can be used as a sort of brain imaging tool (Michel and Murray, 2012). MEG, despite being expensive, provides much better signal and spatial resolution, and more channels by default.

The few studies using EEG data implemented the encoding setup, as in Murphy et al. (2009), where pictures of concrete entities where used as stimuli; and both encoding and decoding in Sassenhagen and Fiebach (2020), a study where both concrete and abstract common nouns were used. MEG data, which grants higher machine learning performances given the superior signal quality, was instead used for decoding to word vectors from brain processing of pictures referring to concrete concepts (Sudre et al., 2012) and visually presented stories (Wehbe et al., 2014b).

With respect to the literature here reported, the main innovation of our work is showing that word vectors can be used for isolating extremely fine-grained brain signatures of individual entities indicated by proper names. By doing so, we go beyond knowledge about generic concepts indicated by common nouns, which was instead the main focus of the works reported above. Also, we show that it is possible to do so using the available technique with the lowest signal-to-noise ratio, EEG, provided that the experimental design is carefully designed.

3. Materials and Methods

3.1. Stimuli

In order to be able to investigate the structural properties of the representation of individual entities from both an entity-level and a categorical perspective, we collected evoked responses not only to proper names of people and places, but also to words referring to their main categories. Importantly, the set of stimuli is hierarchically structured, and carefully matched in terms of semantic categories and entities. There are two “coarse” categories (people and places); for each of these we considered four fine-grained categories (musicians, politicians, actors, writers as people, and bodies of water, monuments, cities and countries as places); and for each fine-grained category we included four individuals among the stimuli, as well as the nouns for the categories. The result is a total of 40 stimuli: 8 nouns for the fine-grained categories, and 32 proper names for the individual entities. The hierarchical structure of the stimuli and the way they are used in the experimental paradigm (Section 3.2) are presented in Figure 1, and the stimuli selection procedure is described below.

FIGURE 1
www.frontiersin.org

Figure 1. Experimental design and hierarchical stimuli organization. Our set of stimuli was organized symmetrically, so that exactly the same number of stimuli was present for each coarse- and fine-grained category—as shown in (A). Stimuli included both proper names of individual entities and their fine-grained categories. In (B) we present the experimental paradigm. Each stimulus was projected in isolation on the screen for 750 ms, and subjects then had to mentally visualize its referent while a cross remained on screen for 1,000 ms. Participants then answered a question, involving the fine- or coarse-grained level of categorization.

People and places were chosen a priori as top categories, following previous work on individual entities in cognitive neuroscience (Gorno-Tempini and Price, 2001; Grabowski et al., 2001; Ross and Olson, 2012; Fairhall and Caramazza, 2013; Leonardelli et al., 2019). In this field, people and places are the most common choice of categories to be contrasted when investigating the semantics of individual entities, since a long tradition of studies on patients has shown that, within human semantic knowledge of individual entities, these two categories can be selectively impaired (della Rocchetta et al., 1998; Miceli et al., 2000; McCarthy and Warrington, 2016). This seems to entail that semantic knowledge about these two kinds of individual entities can be teased apart in the brain—a finding which has consistently been confirmed by results from healthy subjects (Fairhall et al., 2014; Rice et al., 2018).

It should be noticed that in NLP, where textual resources can be obtained or created on a very large scale with limited effort, it is commonplace to consider a wider range of coarse-grained categories of entities (organizations, dates, events, products, or categories of biomedical entities such as genes and diseases Goyal et al., 2018). However, in cognitive neuroscience, it is much harder to obtain large scale datasets. This is due to the combination of two factors. The first one is practical: experimental sessions are particularly intense for subjects, imposing strong constraints on the number of trials that can be obtained and analyzed. The second one is methodological: in most cases, like ours, analyses are conducted at the level of individual subjects, and then aggregated at a later stage. This entails that the amount of experimental items that can be analyzed in the context of an experiment is limited to the number that can be considered for a single participant. The role of having multiple participants, instead, is not that of increasing the number of experimental items or trials, but that of testing the generalizability of the results.

For the fine-grained categories and the individuals, we followed a two-step, mixed approach, guiding our manual selection with data-driven observations. The aim of the first selection step was providing an initial, large set of individual entities and matched categories, to be further reduced during the second phase, using familiarity ratings provided by independent subjects.

For the first step, we came up with a list of ten fine-grained categories for people and places. We followed principles of economy and discriminativeness (Rosch, 1975), selecting, as fine-grained category stimuli, categories for people and places which represented economic ways of describing and distinguishing subclasses of individual entities. In the case of people, we used occupations as fine-grained categories (e.g., “politician” for individual concept “Barack Obama”), which is one of most basic, socially shared and clear-cut way of categorizing people (Cantor and Mischel, 1979; Mason and Macrae, 2004; Turk et al., 2005), whereas in the case of places we employed their taxonomic hypernym (e.g., “city” for individual entities such as “Istanbul”). The final list of ten categories included the eight categories reported above (musician, politician, actor, writer as people, and body of water, monument, city and country as places), as well as two additional entries, athlete and geographic area. Then, from this list of fine-grained categories, we manually picked a preliminary set of 100 well-known individual entities—10 names for each of 10 fine-grained categories. We made sure that each individual entity had a dedicated page in both English and Italian Wikipedia, since we needed that source of textual data in both languages for the extraction of word vectors, as discussed in Section 3.5.

For the second, and final, stimuli selection step, we chose familiarity as the main criterion, as it is one of the most important variables affecting the processing of proper names (Valentine, 1998; Smith-Spark et al., 2006; Brédart, 2017; Moore and Valentine, 2020), as well as a necessary requirement for our experiment (in order to capture the neural representations for an individual entity, we needed to ensure that subjects had a previous representation in semantic memory to be retrieved). Familiarity was defined, following Moore and Valentine (2020) as cumulative encounters with (representations of) that individual entity, across time and media. At the beginning of the second stimuli selection step, we collected familiarity ratings from 30 Italian subjects, none of which took part to the subsequent EEG experiment. We made sure that this sample was matched in nationality and close in age to the sample of the EEG experiment (mean age: 29), in order to ensure that the entities and the fine-grained categories used as stimuli for the data acquisition procedure would be as familiar as possible to our sample of participants in the EEG study. Note that this procedure was fully independent from EEG data collection: participants in the EEG experiment had no role in the stimuli selection procedure, and they were not asked to norm the stimuli according to familiarity. In the norming experiments, subjects were asked to rate on a Likert-type scale, from 1 to 5, their familiarity with each individual entity, where familiarity was defined as reported above.

After having collected the familiarity ratings, we retained, for the EEG study, only the individual entities and the fine-grained categories which, on average, had the highest familiarity scores. At the end of this procedure, the four most familiar fine-grained categories for each coarse category were chosen as fine-grained categories for the study; and the four most familiar individual entities belonging to those categories where selected as the individuals entities to be used. We report the set of stimuli in the Supplementary Materials.

Psycholinguistic variables such as orthographic complexity and length are particularly hard to match for proper names, especially in the case of our experimental setup. First of all, in languages such as Italian and English, they are morphologically and orthographically immediately recognizable from common names, because of features such as upper case initial letter, and non-applicability of number and case information (Peressotti et al., 2003). Secondly, proper names of people are in most cases longer than places' names, because they require both name and surname in order to be correctly disambiguated. Thirdly, there is no hope of matching proper names and words of fine-grained categories. And finally, since we use a decoding analysis centered around semantics, we assume that the decoders will automatically learn to focus on ERP signatures of semantic, not orthographic, information. Because of these reasons, we chose not to control rigidly orthographic variables, and in particular stimulus length, during data collection, taking for granted that such differences between stimuli would inevitably be present. Instead, we added word length as a control variable in the classification analyses (Chyzhyk et al., 2018), where it could play in principle a confounding role due to the differences in mean lengths across categories—9 letters for places, 12 letters for people (see Section 3.4 for the details, and the tables in the Supplementary Materials for the full list of the experimental stimuli, together with their familiarity scores and average word lengths).

3.2. Experimental Procedure

Thirty-tree right-handed subjects (age from 18 to 35 years old, with 20 female participants) took part to the experiment. The subjects were all native Italian speakers, and the experiment was conducted in Italian. All experimental procedures were approved by the Ethical Committee of SISSA, Trieste, where the data were collected, and subjects gave their written informed consent.

The experimental procedure involved 24 short runs of 40 trials. In each run, each of the 40 stimuli (both proper names and fine-grained category nouns) would appear once, in a randomized order, resulting, at the end of the session, in 24 evoked responses to each name and noun. We chose to repeat the stimuli 24 times, since (Grootswagers et al., 2017) clearly demonstrate that, in order to reach optimal decoding and classification results using evoked potentials, between 16 and 32 trials for each stimulus are needed. Each trial consisted of two parts, as shown in Figure 1: first, the presentation of a stimulus name or noun for 750 ms (word presentation times in EEG experiments are kept below 1 s, as word processing begins already at 150 ms after the stimulus appears Simanova et al., 2010; Wehbe et al., 2014b; Sassenhagen and Fiebach, 2020); the stimulus was preceded and followed by the presentation of a white fixation cross for 1,000 ms at the center of the screen. Participants were instructed to read the word and then visualize mentally the referent of the stimulus, until the cross was on screen. We decided to leave a relatively short time for the mental imagery task, in order to keep the process as much as possible time-locked to stimulus appearance (Bastiaansen et al., 2011; Shatek et al., 2019) and avoid mind wandering, while allowing subjects to quickly picture the stimulus' referent (cfr. Section 5.3).

Afterwards, a question appeared on screen, which always involved pressing either “s” or “k” on an external keyboard, placed in front of the subject. “s” and “k” were chosen simply because of their position on the keyboard's QWERTY layout: they can be easily reached with left and right index finger, respectively, while keeping a comfortable, yet fixed, position on the chair. There were two possible types of questions, either a coarse-level question (“people or place?”) or fine-level (e.g., “the name referred to a musician,” with possible answers “correct” or “wrong”), similarly to Leonardelli et al. (2019). The mapping between response keys and answers was fully randomized across trials. Within each run, questions were randomized, but balanced across question type and, when applicable, answer type: coarse-level questions appeared 16 times, whereas fine-level questions appeared 24 times, of which 12 times required a “correct” answer. The questions were added to the experimental paradigm in order to keep participants attentive and to ensure that they focused on their semantic representation of the individual entity and its category. The aim of randomizing different questions was avoiding as much as possible strategic preparation for the coming question.

3.3. EEG Recording and Preprocessing

For data acquisition, we employed a BIOSEMI ActiveTwo system with 128 channels 1, recording signals at a sampling rate of 2048 hz. We employed a fully automatized preprocessing pipeline adapted from Jas et al. (2018), and implemented with the MNE Python package (Gramfort et al., 2013), in order to improve replicability. Results were then visually inspected in order to check for the quality of the preprocessing, and we found that no amendment was required.

We first downsampled the data to 256 hz, a recommended choice for cognitive neuroscience experiments (Luck, 2014), leading to a sampling resolution of 3.9 ms. This reduces the amount of data points to be processed, speeding up the analyses, at no cost: as pointed out in Luck (2014), relevant cognitive activity happens at frequencies below 80 hz, and therefore, following Nyquist's theorem, sampling at 256hz is a safe choice (Nyquist's theorem states that, in order to capture in a digital format an analog signal, the sampling rate has to be more than twice as great as the highest frequency in the signal, here around 80 hz). Given the considerations above, we subsequently applied a low-pass filter to the data at 80hz so as to minimize the effect of irrelevant signal interferences.

A further source of noise in EEG recordings are voltage artifacts that have nothing to do with cognitive activity, such as slow voltage drifts. These are due to skin-related potentials, that generate noise in the recorded waveform because they have a shifted phase with respect to the evoked responses to the experimental stimuli (Luck, 2014). To remove them, we applied baseline correction, which consists in subtracting the average of the pre-stimulus potentials, in our case from –100 to 0 ms, from the whole epoch. This is the recommended alternative to the conceptually simpler high-pass filtering, which may instead induce undesired artifacts (Tanner et al., 2016; van Driel et al., 2021).

Finally, we epoched the data to –0.100 and 1,200 ms after stimulus onset, and used the AutoReject algorithm to find and interpolate bad channels and to remove bad epochs, excluding from the analyses spans of recorded data that contained excessive noise or artifacts (Jas et al., 2017).

We also recorded two electrooculogram (EOG) channels, which record electrical potentials produced by eye movements. These eye-related potentials, which may be picked up in parallel by EEG electrodes, interfere with the signal of interest coming from the brain, and are therefore considered as artifacts to be removed. The most common way of dealing with this kind of noise employs the recorded EOG channels together with Independent Component Analysis (ICA) (Urigüen and Garcia-Zapirain, 2015). ICA is used to separate linearly mixed signals, estimating their independent source signals from recorded data. In our case, we take the EEG recordings to contain a mixture of eye- and brain- related signals, that we would like to disentangle. In order to remove ocular artifacts, we used the standard procedure of Jas et al. (2018). Using the MNE implementation, we fit ICA on our data, then we found and excluded automatically the components which correlated the most with the EOG signal (Jas et al., 2018), the assumption being that these components capture eye-related signal sources.

To reduce the impact of noise, which can be quite severe in EEG recordings, we averaged all evoked responses corresponding to a stimulus within a subject, as it has been shown to improve the signal-to-noise ratio (Grootswagers et al., 2017). Before the analysis, we standardized the data for each channel using MNE's scaling method, which standardizes electric potentials channel by channel. For the standardization, mean and standard deviation are computed for each channel from all time points within all epochs.

This preprocessing pipeline provided us with 40 evoked responses for each stimulus per subject. Conceptually, these correspond to snapshots of semantic processing across 1,200 ms for the thirty-two proper names and their eight fine-grained categories.

3.4. Coarse- and Fine-Grained Decoding

Given the structure of our dataset, and our interest in uncovering the structure of semantic representations of individual entities at different levels of granularity, we carried out two separate classification analyses for the coarse-grained level (people and places) and the fine-grained level (the eight fine-grained categories). For both setups we employed a SVM with default parameters (C=1.0, l2 regularization). This is standard procedure in cognitive neuroscience, and it has been shown that, with l2-regularized SVM, fine-tuning the C parameter does not impact results (Grootswagers et al., 2017; Varoquaux et al., 2017). The main difference between coarse- and the fine-grained setups was that, whereas in the former case we set up a binary classification procedure, with random, baseline accuracy at 50%, in the latter we used multiclass, one-vs-all classification with random baseline at 12.5%. We will also present results for within-coarse categories (i.e., results obtained when using only evoked responses to either people or places). In this case too we used a one-vs-all classifier, this time with a baseline of 25%, since there are only four fine-grained categories to consider.

We took a time-resolved decoding approach (Grootswagers et al., 2017) for our classification, training and testing separately on each time point. We tested statistical significance at each time point using threshold-free cluster enhancement (TFCE), a permutation-based, non-parametric test of statistical significance proposed in Smith and Nichols (2009) and then widely adopted in the neuroscientific literature (Helwig, 2019), also for classification of EEG signals (Grootswagers et al., 2019; Kaiser et al., 2020; Petit et al., 2020). The main advantages of this procedure are its sensitivity, due to the fact that it can take into account the fact that brain signals are clustered both in space and in time; its avoidance of parametric test assumptions (Mensen and Khatami, 2013); and finally, the fact that it inherently counters the multiple comparisons problem (the inflated risk of finding false positives) that arises from testing so many data points (in our case, time points; for details on the general procedure, see Mensen and Khatami, 2013). We used the TFCE implementation of MNE (Gramfort et al., 2013), with default parameters. Since our classification took place in the time domain (time-point by time-point), the TFCE procedure could take into account only temporal adjacency when looking for potential clusters. Time-points were considered to be adjacent if they fell within a 10 ms window (remember that our resolution is 3.9 ms—Section 3.3), as post-synaptic potentials, the kind of potentials captured by EEG recordings, do not last less than 10 ms (Luck, 2014), and therefore the EEG signal can be assumed to be smoothed within that time window.

Given the nested categorical structure of the labels, we paid special attention to data splits among training and testing. We generally took a leave-4-out evaluation approach (see exceptions at the end of this section), which trains on 87.5% of the full original data, leaving out 12.5% for testing. We did not use random folds, which would end up giving unreliable results, because of unbalanced splits at the fine-grained level. Instead we first computed all possible combinations of sample labels to be used as test sets which respect the following criteria: sample labels should be balanced across coarse categories (two people and two places), and the test set should not contain more than one exemplar from each fine-grained categories. An example of a test set could be “Barack Obama,” “Scarlett Johansson,” “Eiffel Tower,” “South Africa,” with different labels depending on the analysis: “people,” “people,” “places,” “places” for the coarse-grained analysis, and “politician,” “actress/actor,” “monument,” “country” for the fine-grained one. This procedure can be considered as a balanced implementation of the ShuffleSplit data splitting technique (Pedregosa et al., 2011; Varoquaux et al., 2017).

Given that word length was not strictly controlled across categories in stimuli selection, giving rise to different average word lengths for each category (see Section 3.1 and Supplementary Materials), word length could in principle act as a confound in classification, where categories are used as target labels. In order to control for its effect, we adapted to our classification analyses the confound variable control procedure presented in Chyzhyk et al. (2018). This method was specifically validated by the authors in the case of neuroscience classification analyses where an individual variable has to be controlled, showing that it avoids both overly pessimistic and optimistic accuracies (Chyzhyk et al., 2018; More et al., 2020). The intuition it follows is that, to control for a potential confound in classification, given a large enough pool of candidate train-test splits, it is enough to select and use only the splits where, in the test sets, the outcome is independent from the confound: with this procedure, the classification analysis provides an evaluation of whether the brain data determines successful prediction beyond the confounding effect. In our setup, controlling for word length in this way is straightforward since, as discussed above, we generate as candidate train-test splits all the possible balanced combinations of stimuli. First, for each possible candidate test set, we transform stimuli and target labels (the categories' nouns), that are categorical variables, to numeric values: words become word lengths, and target labels become the average word lengths of the label's entities (as reported in the Supplementary Materials; e.g., “Madonna” is encoded as 7, and “politico” is encoded as 13, the average of all politicians' names in our set of stimuli). Then, we compute the Spearman correlation (indicated by ρ) between the encoded stimuli and target labels: this quantifies the confounding effect of word length on that candidate test set. Having computed the correlations for all the candidate test sets, we sort them in ascending order by their absolute value. Finally, following Varoquaux et al. (2017), where authors recommend using 50 random train-test splits when working with brain data, we retain as final test sets the 50 candidates having the lowest correlation. In the case of the coarse-grained classification, we could use 50 test sets completely uncorrelated from word length (ρ = 0.0), whereas in the case of fine-grained classification the highest correlation is not 0.0, but still close to it (ρ = 0.162) 2.

In order to get insights regarding the structure of the representations of proper names, we further exploited the hierarchical structure of the dataset. In a separate set of analyses, we trained on evoked responses for individuals, and we only tested on those for the fine-grained categories. This can be seen as some sort of transfer learning, looking at whether, and when, the two kinds of representations converge. In this case, the dataset was used in its entirety. Furthermore, since we reckoned that there could be some differences in terms of discriminability between people and places, we also ran the analyses separately first on people only, and then on places. In this case we used a leave-2-out classification setup, corresponding again to a 87.5% train–12.5% test split.

3.5. Distributional Semantics Models

One of the key goals of our work is exploiting an array of recent, cutting-edge computational models of semantics in order to isolate semantic information in evoked responses in the brain, at the level of unique, individual entities. We selected a set of models which can be connected as transparently as possible to cognitive theories of the semantics of individual entities. At the broadest level, we use three kinds of models. All of them represent individual entities as vectors, but the way in which these vectors are created differs significantly across models.

The first kind of word vector models relies exclusively on distributional information about words as observed in linguistic corpora (BERT Devlin et al., 2019, ELMO Peters et al., 2018, Word2Vec Mikolov et al., 2013; see Section 3.5.1). Within this family, all models follow, with various degrees of sophistication, the distributional principle, according to which words which are found in similar contexts have similar meaning. With respect to the way they represent individual entities, they can be understood as putting them on the same level as all other words—individual entities are not given a special treatment, and their representation follows the same pathway as that of all other words. Distributional models have been often taken as models of human semantic memory, both because of principled reasons (they embody the assumption that word meaning can be retrieved from their use, and that it can be defined in a distributed fashion, across various dimensions) and because of their empirical success at modeling human data.

The second type of models (TransE Bordes et al., 2013; see Section 3.5.2) comes from a very different tradition in artificial intelligence, that of ontologies and structured representations of entities (Guarino, 1995; Ji et al., 2021). Representations of these kinds follow a distinct principle, that of creating models of the world that an artificial intelligence agent could use in an external environment (Guarino, 1995), and have most commonly taken the form of knowledge bases—graphs where entities (both unique and generic) are the nodes, and the links among nodes capture their relationships. In this case, entities—and individual entities in particular—are at the very core of the representational structure.

The third kind of models (LUKE Yamada et al., 2020b, Wikipedia2Vec Yamada et al., 2020a; see Section 3.5.3) is a combination of the two approaches presented above, and it has recently received much attention in NLP because it promises to overcome the limitations inherent to each methodology: for distributional models, the lack of precise entity-level information; and for knowledge bases, their structural rigidity and their difficult integration with generic linguistic knowledge, as well as their costly creation and maintenance (Peters et al., 2019; Sun et al., 2020b; Yamada et al., 2020b). We will call these models, as is commonplace in the NLP literature, entity-aware embeddings. From a cognitive point of view, these computational models may be interpreted as implementing the intuition that individual entities are represented at a separate level than generic entities. Some models introduced this idea in cognitive psychology (Bruce and Young, 1986; Burton and Bruce, 1992; Young, 1999), stating that individual entities bearing a proper name have a dedicated identity node, which is then integrated, at a later stage, with generic semantic knowledge.

For all models we use the version trained on English. This is the language over which the models were developed originally, because of resource availability—as a matter of fact, importantly, for most models the Italian version is not available at all. Also, in terms of performance on NLP tasks in different languages, English consistently presents the best performances available (Bender, 2011; Pires et al., 2019). We used English vectors for many reasons. First of all, our stimuli can be considered to be largely language-independent (Van Langendonck and Van de Velde, 2016), at least across the two languages involved in our experiment (Italian, the language in which the experiment was carried out, and English), and strongly referential in nature: the referents of proper names are specific people and places in the world, which only require to be familiar with them (Kripke, 1972), and the category nouns that we employed (occupations and type of place) are shared across Italian and English cultures. The level of representation of interest is exclusively semantic, thus ruling out orthographic and phonetic language-specific phenomena.

To support our argument, we checked empirically whether using Italian models, instead of English models, would give rise to relevant differences in decoding performance. We ran the analyses with the models available also in Italian (Word2Vec, Wikipedia2Vec and BERT base), and then compared results across languages with a two-tailed non-parametric Wilcoxon statistical significance test, correcting for multiple comparisons using Benjamini and Hochberg (1995)'s False Detection Rate (FDR) procedure, which is a standard procedure in both computational linguistics and neuroscience (Groppe et al., 2011; Dror et al., 2017). Difference in scores was not significant for BERT (p = 0.97), whereas it was significant for Word2Vec (p = 0.043, with the Italian model performing better—average Italian model accuracy: 0.619, average English model accuracy 0.595) and Wikipedia2Vec (p = 0.017, where the opposite was true—average Italian model accuracy: 0.609, average English model accuracy: 0.639; cfr. Figure 8). Given that results did not show a consistent pattern of advantage for one language or the other, and that most of the available models were trained on English data, we report and discuss results for the models trained on English.

3.5.1. Static and Contextualized Distributional Models

Distributional models can further be subdivided into static and contextualized. Static models, such as Word2Vec (Mikolov et al., 2013), work at the level of individual lexical items. They are essentially vector spaces, where each vector captures the meaning of an individual word. Contextualized models, such as the widely used ELMO (Peters et al., 2018) and BERT (Devlin et al., 2019), are more recent. They instead focus on sentences, and not on individual lexical items (Camacho-Collados and Pilehvar, 2018). Given a sentence, their goal is that of creating representations of the words contained in the sentence which reflect their current idiosyncratic, context-dependent meaning. Contextual distributional models have been generally shown to improve on static models on most tasks in Natural Language Processing (Rogers et al., 2020), but it should be noticed that they are dramatically more complex than static models (Lenci et al., 2021).

As our static distributional model, we choose Word2Vec, a very well-known model, which has also been shown to be a good model of human semantic memory (Mandera et al., 2017). Word2Vec vectors are created by training a feedforward neural network on a word prediction task. In the case of the model we use here, it is called the “skip-gram” task (Mikolov et al., 2013). It requires to learn to predict whether, given a word in a sentence and another target word from the vocabulary, the target actually comes from the actual set of words surrounding the query, or whether it was randomly sampled among the words not appearing in the window. In this experiment we use a model pre-trained by the authors of the original papers on a corpus of news articles, where individual entities have been marked as individual words so that they end up having their own vector. The vectors were created by the authors of Mikolov et al. (2013), optimizing the learning parameters and the dimensionality (1,000) for the representation of entities.

There is a very large number of contextualized models, often specialized for specific NLP tasks. We chose to use two of the “basic,” vanilla models (which are in any case quite complex) on top of which most of subsequent research has been built, ELMO (Peters et al., 2018) and BERT (Devlin et al., 2019). Despite being created for generic language processing—actually, the adaptability of the word vectors they create is one of the reasons of their success—these models can in most cases compete with specialized models in terms of performance, and they are very often used as a strong benchmark. ELMO is a bidirectional LSTM neural network, which learns to predict a given word conditioned on the previous words, as well as the next ones. BERT, instead, adopts the Transformer architecture (Vaswani et al., 2017), which relies heavily on the computational mechanism of attention (Lindsay, 2020), in order to encode a sentence into a set of vectors capturing both lexical and contextual meaning and structure. Importantly, BERT is a deep architecture (it exists in two flavors, one with 12 layers and one with 24 layers), with different layers encoding—at least partially—different kinds of linguistic information (Rogers et al., 2020).

One of the challenges posed by contextual models, as opposed to static word vectors, is that they leave many free choices to the experimenter when it comes to extracting the vectors: which layer to use; whether to use vectors for word mentions, or instead those for full sentences; what model dimensionality to choose. For both ELMO and BERT, we follow a methodology proposed in Bommasani et al. (2020) and refined in Lenci et al. (2021), which has been show to capture effectively lexical meaning. We extract many vectors for separate mentions of each proper name, and then average them. The mentions are taken from the Wikipedia pages of each individual entity, in order to encode definitional information about the entities themselves. From each Wikipedia page we take up to the first 32 sentences. We experimented with various layers, but finally we decided to use an average of the final layers, as in Lenci et al. (2021) (the last four for BERT, and the last one for ELMO), since they have already been shown to perform well with neural data (Jat et al., 2019). For BERT we used Huggingface's Transformers implementation (Wolf et al., 2020) and the original pre-trained weights, both in their so-called “base” and “large” versions (1,024 and 2,048 dimensions, respectively). For ELMO we used the pretrained original weights (1,024 dimensions) and the AllenNLP implementation (Gardner et al., 2018).

Our method for extracting static representations from contextualized models has not been tested specifically on entities, but only on common nouns. Therefore, we validate it by measuring its performance on WikiSRS, a similarity and relatedness task (cfr. Section 2.2) created specifically for entities. This benchmark was introduced in Newman-Griffis et al. (2018), and it was recently used for the evaluation of contextualized models (Chen et al., 2019). The dataset was created by crowd-sourcing similarity and relatedness judgments for 688 pairs of named entities. To evaluate the performance of our vector extraction methodology, we followed the procedure of Chen et al. (2019): first we obtained the vectors for the entities appearing in WikiSRS; then we computed the pairwise cosine similarities between vectors corresponding to the dataset's pairs; and finally we looked at the Spearman correlation between model similarities and human judgments. We carried out separate evaluations for the two portions of the dataset—once for similarity and another for relatedness. As a baseline, we computed the scores obtained when using the vector extraction methodology proposed in Chen et al. (2019), which differs from ours in two respects: first, they only employed the first sentence from Wikipedia as input for the creation of the entity vector; and second, for BERT, as entity representation, they used the special [CLS] token, used by BERT to represent the whole input sentence. For BERT large, our methodology improves on the baseline for relatedness from ρ = 0.297 to ρ = 0.423, and for similarity from ρ = 0.424 to ρ = 0.491. For BERT base, the model's fit with relatedness scores improves from ρ = 0.287 to ρ = 0.375, and for similarity from ρ = 0.401 to ρ = 0.468. For ELMO, the baseline score for relatedness is ρ = 0.372, while with our method it improves to ρ = 0.399; and for similarity, correlation goes from ρ = 0.424 to ρ = 0.4399. We take this consistent pattern of improvements to confirm the validity of our proposed methodology for the extraction of static entity vectors from contextualized language models.

3.5.2. Knowledge-Base Models

As a knowledge base model, we used vectors obtained using TransE (Bordes et al., 2013), a method to translate a knowledge graph into a set of entity and relationship vectors (we only used the entity vectors). Roughly, TransE starts from random vectors, and goes through the knowledge base, optimizing along the way entity and relation vectors. Vectors are tuned so that, given a triplet of entity1, entity2 and a relation holding between the two, the sum of the vectors for entity1 and the relation should return a vector which is among the nearest neighbors of entity2; and the opposite for randomly sampled negative relationships. We employed a pre-trained model, published in Han et al. (2018), which was trained on WikiData, an open knowledge base that can be seen as a structured sibling to Wikipedia (Vrandečić and Krötzsch, 2014). The final vectors have 100 dimensions.

3.5.3. Entity-Aware Models

Finally, as entity-aware embeddings, we use both a static and a contextualized model, built on top of Word2Vec and BERT, that were shown to both improve on their basic version with respect to entity-related tasks in NLP. The first one is called Wikipedia2Vec (Yamada et al., 2020a), and the second one is LUKE (Yamada et al., 2020b). We chose these two models because they both adopt a conceptually similar approach, and they use the same training data, which reduces possible confounds—the English text of Wikipedia, complemented by the underlying graph of the hyperlinks contained in each page. Both models modify their basic training regimes, storing and processing separate representations for individual entities and common nouns. Their strategy consists of exploiting hyperlinks in Wikipedia pages as annotations of entity mentions, using these mentions for an additional entity-specific task which is added to the training. In this task, given an entity mention in a sentence within a Wikipedia page, the model has to learn to predict other entities found in that sentence. We use the pre-trained versions of the models published by the authors. Wikipedia2Vec comes with 500 dimensions, and was trained from scratch with a window size of 10 words and 15 negative examples, for 10 iterations. The entity-specific task is a skip-gram prediction task for mentions of entities. LUKE, both in the base and large versions, has the same dimensionality as BERT (1,024 and 2,048) and was trained on top of pre-trained BERT weights, which were used as initialization weights for the part of the model dedicated to generic word representations. Representations for the individual entities are learnt with the same masked language modeling objective of BERT, just applied separately to entities. For LUKE, which works in practice as a contextualized language model, we follow the same word vector extraction procedure validated in Section 3.5.1.

3.6. Decoding to Distributional Word Vectors

In order to find out whether we can distinguish between evoked responses at the finest level of granularity, that of individual identities, we exploit our entity vectors as mapping targets. In other words, for each entity, we use a regularized linear regression model to learn to predict the true values of the dimensions of each entity vector from the corresponding EEG representation of that entity.

As inputs (i.e., as our EEG representations) we use whole epochs, from 100 to 1,200 ms after stimulus onset, which we collapse to one vector, so that for each evoked response to a name or a noun we have one vector. To learn the mapping, we use Ridge regression with default parameters, following recent work (Pereira et al., 2018; Sassenhagen and Fiebach, 2020), implemented in the Python package Scikit-learn (Pedregosa et al., 2011).

We adopt for evaluation the pairwise evaluation approach proposed by Mitchell et al. (2008). This is a leave-2-out training regime, repeated for all possible pairs of stimuli. Note that the model has to learn to predict the true value of each vector dimension for entities which are completely unseen during training—a form of zero-shot machine learning task, which requires the model to isolate and exploit at test time the signatures in the brain of semantic processing of individual entities.

In pairwise evaluation, at test time the model first predicts two entity vectors, e1^ and e2^, using the corresponding evoked responses; then the respective Spearman correlations to the original entity vectors e1 and e2 are computed. At the end there will be four correlation measures, two of which are for the matching vectors and are expected to be, taken together, higher than the correlations for the mismatched vectors. The decoding accuracy evaluation is based on this expectation, in that it is considered successful, with accuracy = 1, if ρ(e1,e1^)+ρ(e2,e2^)>ρ(e1,e2^)+ρ(e2,e1^); else, decoding is considered unsuccessful, and accuracy = 0.

We controlled for statistical significance using a one-sample, one-tailed Wilcoxon test, because of the minimal assumptions made regarding the distribution underlying the scores (Grootswagers et al., 2017). We also applied FDR to control for multiple comparisons, given that we were running statistical analyses for eight models.

With respect to word length, we could not apply straightforwardly the confound control procedure presented in Section 3.4, since decoding to word vector is not a classification task. However, we point out that, in this case, word length should not even be considered a confound variable for principled reasons: word vectors do not encode any information about word length at all. To show that this is the case, we compute the Spearman correlations between pairwise vector similarities and the matched pairwise differences in character length among words (the intuition being that, if word vectors encoded information regarding word length, their distance in vector space should correlate with the number of additional letters required to match two words in length). Results confirm that word length should not be considered as a confound variable for the word vector decoding analyses: for all models except one we found extremely low correlations (in all cases, ρ <0.08); the only exception is LUKE, where correlation is anyways quite mild (LUKE base: ρ = 0.144, LUKE large: ρ = 0.326).

In order to prove that the word vectors actually captured the kind of semantic information regarding entities that we wanted to isolate in EEG data, we carried out some clustering analyses as a validation procedure, as is customary with distributional models (Almuhareb and Poesio, 2004; Baroni and Lenci, 2010; Lenci et al., 2021), reported in Figure 2. We measure to what extent we can cluster word vectors according to their coarse-grained and fine-grained semantic class, when considering individuals only, categories only, or both individuals and categories; and when restricting the analyses to people and places only, or using both people and places together. As a clustering algorithm, we use K-Means, a common choice in computational linguistics (Lenci et al., 2021). As an evaluation measure we use the adjusted Rand Index, which looks at all possible pairs of samples, measuring how many pairs are correctly assigned to the same or a different cluster, then correcting for chance (Hubert and Arabie, 1985). Both for the clustering algorithm and the evaluation measures we use their Scikit-learn implementation (Pedregosa et al., 2011).

FIGURE 2
www.frontiersin.org

Figure 2. Clustering results on word vectors, with various possible assignments of categorical labels. We test the performance of each model on a completely unsupervised clustering task. In this evaluation, we measure to what extent word vectors can be clustered according to their category, evaluating separately coarse-grained and fine-grained categories. The evaluation metric that we used, Adjusted Rand Index, involves a correction for chance, so results above 0.0 show some sensitivity to categorical structure. We report results for all possible combinations of stimuli: clustering using only the vectors for individuals; only the vectors for categories; both categories and individual entities; people and places separately. Most models perform above chance in all clustering tasks.

The results show that clustering is above chance (>0) for almost all models. Discrimination is easiest for coarse categories, and particularly so when individuals and categories are clustered separately. Performance is above chance for most models in all possible labellings, for fine-grained categorical labels as well. The toughest discrimination is the one where both individual and category vectors for people are used, and in this respect others have already found that social categories for people are not well captured by distributional models (Westera et al., 2021). Also, in general, performance worsens when clustering vectors for individuals and categories together, as already observed by Gupta et al. (2018). Overall, however, indicate that the distributional models encode the categorical structure that we need in order to use them for decoding.

4. Results

4.1. Coarse-Grained Semantic Category Classification

We report in Figure 3 the averaged results for the time-resolved classification of coarse-grained categories, where we classify evoked responses at time t into two classes, either person or place. Classification accuracy raises intermittently above the baseline level from around 150 ms after stimulus onset, which is compatible with visual word recognition processes starting at around 150 ms (Carreiras et al., 2014; Ling et al., 2019)—but never reaches significance while the stimulus in on screen. We interpret this as an effect of the variable control procedure for word length described in Section 3.4, given that it is in this time range that semantics and word-reading processes can get confounded. Scores reach statistical significance later on, starting at around 800 ms, indicating that it is possible to decode coarse-grained categories of individual entities from EEG data during the mental imagery task.

FIGURE 3
www.frontiersin.org

Figure 3. Classification of coarse-grained categories from individuals. We run a time-resolved binary classification on the EEG data. For each time point, we learn to classify evoked responses according to their coarse semantic category (either person or place). We control for word length employing the test sets where the correlation between labels and stimulus length is lowest (see Section 3.4). The green line is the average of results across the 33 participants, and the shaded areas correspond to the standard error of the mean. The random baseline is at 0.5, given that this is a binary classification problem, and is represented by a dotted horizontal line; we also plot as dotted vertical lines the time-points when stimuli appear and disappear. Statistically significant points (p < 0.05 corrected for multiple comparisons by TFCE; see Section 3.4) are reported both on the averaged lines and, to make them easier to read, below the x axis.

One of our objectives was trying to understand whether, and when, semantic information about the coarse-grained class is shared in brain processing between proper names and the names of the categories these individual entities belong to. In Figure 4 we report classification scores obtained when training on individuals, but then testing on categories only—effectively looking at how much information about coarse-grained semantic categories can be transferred from instances of categories (the individual entities) to the categories themselves. Although statistical significance is never reached, decoding performance goes above baseline after 800 ms, during the mental imagery task. The lack of significant decoding accuracies in Figure 4, which are instead reached when using only evoked responses to stimuli at the same ontological level (instances, instead of categories; Figure 3), suggests that coarse-level semantic information does not seem to be strongly shared between entities and categories.

FIGURE 4
www.frontiersin.org

Figure 4. Classification of coarse-grained categories, transferring information from individual entities to categories. Here we report time-resolved classification scores, following the same structure as Figure 3. However, in this case we look explicitly at whether discriminative representational information is shared across individuals and entities, training on individuals, and testing on categories. The results show that this is not the case: scores go past the random baseline only after the stimulus disappears, and never reach statistical significance.

4.2. Fine-Grained Semantic Category Classification

4.2.1. Aggregate Results

When it comes to classification of fine-grained categories, as it can be seen in Figure 5, classification accuracy is again above chance starting after around 150 ms. It reaches statistically significant discriminability between 300 and 400 ms, then it drops until 800 ms, when scores start being statistically significant again during the mental imagery task, until 1,100 ms. A peak of classification accuracy close to 400 ms is to be expected, as this has been consistently found to be the time frame where word-level semantic processing happens (Hauk et al., 2006; Simanova et al., 2010; Frank et al., 2015; Sassenhagen and Fiebach, 2020). For instance, the N400, a negative deflection in recorded brain potentials at around 400 ms after stimulus appearance, is considered to be a signature of semantic processing in EEG responses, although it is still debated precisely what semantic process it should reflect (Lau et al., 2008; Kutas and Federmeier, 2011; Rabovsky et al., 2018). Concerning mental imagery, the scores strongly indicate that fine-grained categorical information can be discriminated during this experimental task, converging with the results of Figure 3.

FIGURE 5
www.frontiersin.org

Figure 5. Classification of fine-grained categories from individuals. We run a time-resolved multi-class classification analysis, trying to decode at each time point the fine-grained semantic category of the stimulus. Figure structure is the same as in Figure 3, but here the random baseline is at 0.125, since there are eight possible categories (four for people and four for places). Scores are statistically significant between 300 and 400 ms, and from 800 to 1,100 ms.

When considering the commonalities between evoked responses for individual entities and categories (Figure 6), we see that decoding accuracy when training on individuals and testing on categories barely makes it past the random baseline at discontinuous time-points, never reaching statistical significance. These results concur with those of Figure 4, in that both seem to indicate that not much in terms of semantic representation is shared across individuals and categories.

FIGURE 6
www.frontiersin.org

Figure 6. Classification of fine-grained categories, transferring information from individual entities to categories. This figure is the equivalent of Figure 4, just referring to the multi-class classification task involving fine-grained categories as labels. The random baseline is set at 0.125, because of the presence of eight possible classes. Notice that decoding scores follow a very different course with respect to coarse-grained categories. Statistical significance is never reached, confirming that little discriminative semantic information is shared between responses to individuals and categories.

4.2.2. Per-Category Results

We also compared, separately, the performance on either people or places. These two coarse-grained categories may, in principle, produce very different results, as their respective fine-grained categories are defined differently, by necessity: for people, based on their occupations; for places, more generically based on their most immediate superordinate category (cfr. Section 3.1).

And indeed, the emerging patterns of results go in different directions. Results for both categories are reported in Figure 7. When using only evoked responses to people, classification performance never reaches significance, and is above the chance baseline only shortly at various time-points (around 300–400 ms, 500 ms, 700 ms, and 800–1,100 ms). On the other end, limiting the analyses to fine-grained place categories results in better classification accuracy, reaching statistical significance between 300 and 400 ms (as in Figure 5). These analyses indicate that fine-grained categories are more easily discriminable in our EEG dataset in the case of places—in lay terms, that it is harder to find in the evoked responses separate traces of the semantic distinctions among musicians, writers, politicians and actors than it is when we consider monuments, cities, countries and bodies of water.

FIGURE 7
www.frontiersin.org

Figure 7. Classification of fine-grained categories, separately for people and places, considering individuals only. We plot classification scores against time, as reported above in Figures 36, with the exception that here we restrict our analyses to evoked responses for either people or places. Since there are only four possible classes, random baseline is at 0.25. Interestingly, for people results never reach significance, suggesting that people categories are hard to decode from EEG data. Instead, when decoding fine-grained semantic categories of places, results are strikingly different: classification reaches statistical significance between 300 and 400 ms.

It is impossible to reach strong conclusions given the differences in the nature of the fine-grained categories of people and places (see above and Section 3.1), but it can be noticed that these results converge with previous results showing a distinction, within individual entities, between semantic processing for proper names of conspecifics and other kinds of entities such as places (Miceli et al., 2000; Lyons et al., 2002; Caramazza and Mahon, 2003; Mahon and Caramazza, 2009; Fairhall et al., 2014).

4.3. Decoding to Distributional Word Vectors

In Figure 8 we report the results obtained when decoding from evoked responses to word vectors: training and testing is limited to individual entities only, since we are looking at whether it is possible to discriminate among individual entities in the brain.

FIGURE 8
www.frontiersin.org

Figure 8. Decoding to word vectors, considering only individual entities. We report the distribution of the scores for 33 subjects when predicting word vector dimensions from EEG data. In this case, we employ evoked responses and word vectors for individual entities. Evaluation is carried out using the leave-two-out methodology proposed in Mitchell et al. (2008). For each computational model, we plot the distributions as violin plots, where the white dots indicate the model average score, and gray bars are used for the 95% confidence interval. We run a one-tailed Wilcoxon statistical significance test against the chance baseline of 0.50. Stars indicate the resulting p-value. In general, results are well above random performance, providing evidence that it is possible to distinguish representations of individual entities in the brain, by exploiting computational models of language. Also, contextualized models perform better than their static counterpart.

The best average performance, 0.683, is reached by the largest version of BERT, which in NLP often provides excellent performances (Rogers et al., 2020). Overall, contextualized models show higher accuracies (average scores: BERT large = 0.683, BERT base = 0.67, LUKE large = 0.667, LUKE base = 0.666, ELMO = 0.658), providing better fits than static (average scores: Wikipedia2Vec = 0.639, Word2Vec = 0.595) and graph-based models (average score for TransE = 0.622). In order to test the statistical significance of results, we compare all possible pairs of models using a two-sided Wilcoxon test with FDR control of multiple comparisons. All contextualized models show statistically significant improvement on Word2Vec (in all cases, p < 0.0005); and very similar results emerge regarding TransE, since the difference in scores is statistically significant for all models (BERT base: p = 0.0021, BERT large: p = 0.0004, LUKE base: p = 0.0054, LUKE large: p = 0.0085) but ELMO, which is approaching significance (p = 0.057). Difference with Wikipedia2Vec is statistically significant for LUKE base (p = 0.024), BERT base (p = 0.035) and BERT large (p = 0.0054).

When comparing contextualized models with one another, statistical significance tests do not provide any p-value below 0.05, indicating that no reliable difference among the performances of these models exist; nevertheless, it should be noticed that BERT large is not only the model showing the best average results, but also the only model to get close to significance against other contextualized models (against BERT base: p = 0.064; against ELMO: p = 0.088; against LUKE large and base: p = 0.132), while the p-values for comparisons involving the other contextualized models are well above 0.3. Within static models, statistical significance is only reached when comparing Wikipedia2Vec and Word2Vec (p = 0.0062). Taken together, these results also suggest that adding knowledge graph information to word vectors does not make them better matches for their brain counterparts, neither for contextualized nor for static models.

In order to better understand how each type of semantic organizational principle (person or place, individual or category) affects decoding scores, and how semantic representations may differ or converge across the various categorizations, we exploit our full set of evoked responses, including both individual entities and categories for training and testing. Doing so while using the leave-two out evaluation of Mitchell et al. (2008), gives us the possibility to break down scores into separate bins, depending on the two evoked responses which compose the test set. We could further subdivide results into accuracies when both test samples are individual entities, or categories, or mixed; and even further, into separate scores for test instances having words referring to two people, or two places, or a mixture of the two. We report in Figure 9 a breakdown of the results, when using BERT large, our best model (results with other models are strongly correlated, on average ρ = 0.93), on the full dataset of individual entities and categories.

FIGURE 9
www.frontiersin.org

Figure 9. Breakdown of results when decoding to word vectors, considering both individuals and categories. Making the most of the leave-two-out evaluation methodology, which allows to split results according to the semantic and ontological category of the two words, we break down the results for the model with the highest average score, which is BERT large. Each violin plots the distribution of scores for a given combination of categories across subjects. Scores are statistically significant in all cases, except for the case when only categories are used. There, only one set of scores is statistically significant at p < 0.05, and the 95% confidence interval always include the baseline, indicating that performance is not reliably above chance. Confirming that individual entities and categories are easiest to discriminate, scores improve dramatically when an individual entity and a category are left out for testing.

As is to be expected, better scores are obtained when comparing, in the pairwise evaluation, one person and one place, with just one exception: representations for people's categories and entities are easiest to discriminate when compared with one another. High performances, well above 0.7, are reached when the test instances are one category and one entity—confirming that representations of entities and categories seem to be segregated in the brain. The worse results come from decodings of categories alone, rather surprisingly given that previous literature has shown that common nouns can be decoded successfully. Importantly, in these cases the 95% confidence intervals always include the random baseline, a sign that, even in the only case where a significant p-value is reached (p < 0.05, people against places), performance cannot reliably be said to be above chance. In this respect, we are not able to draw conclusions given that our set of categories is small (8 in total). However, we can point out that, when considering that most of the training set is made of individual entities, these results seem to suggest that the decoder could not find much information that could be transferred between the representations of individual entities and their categories.

5. Discussion

5.1. The Advantages of Bringing Together Two Approaches

Some of the results of our work are primarily of interest from a cognitive neuroscience perspective, whereas others are of interest mainly to practitioners of computational linguistics, NLP and artificial intelligence. Nevertheless, each of these perspectives should be of interest to the other side, and in our experiments we brought them together in order to maximize the reciprocal relevance of, on the one hand, brain data, and on the other, distributional models of word meaning. This interplay of cognitive neuroscience and computational linguistics in our work can be seen in many directions.

First of all, we did not employ specialized experimental tasks, such as semantic priming. These may induce task-related, strategic biases in the results, as argued separately in Wiese and Schweinberger (2011), Wiese (2011), and Adorni et al. (2014), which makes them harder to use for artificial intelligence research. Instead, we aimed to capture snapshots of brain signals for semantic processing of names and their categories that could be aligned with representations of the same names and nouns obtained from models of meaning in artificial intelligence. We achieved this by using a nested set of stimuli and a paradigm which motivates the participants to retrieve and focus mentally on their representations of individual entities and the respective categories.

We then carried out two types of machine learning-based analysis of the relationship between the representation individual entities and that of categories. The first group of analyses, summarized in Figures 37, focuses on brain data only, but with results that should have an import on computational models–namely, that categories are represented at a separate level from entities. These constraints should be taken into account when developing computational models of individual entities. In a second set of analyses, reported in Figures 8, 9, we directly aligned brain data and distributional models, using the latter to isolate pieces of semantic information in the brain with great precision, by way of decoding. In this case the connection between the two approaches is particularly evident: computational models find traces of semantic processing of individual entities in the brain, as it can be captured by word distributions in text, without confounds from other linguistic processes such as orthography and morphology; and conversely, measures of the fit of each model with brain data quantify the amount of cognitively-relevant information contained in each distributional semantics model.

5.2. Decoding Individual Entities Using Word Vectors

The aspect of this work that is likely to be of the greatest interest to the NLP and AI communities are our results regarding the decoding of responses to individual entities to word vectors.

First of all, we have shown that it is possible to map from brain representations of individual entities to their distributional vectors. So far, this had only been achieved for common nouns (Mitchell et al., 2008; Anderson et al., 2017; Pereira et al., 2018).

These results are all the more surprising because individual entities are semantically much more fine-grained than generic entities, and their meaning is traditionally taken to be determined to a much greater extent by their real world reference, rather than their distributional behavior (Kripke, 1972). Other proposals argue that the meaning of proper names is determined socially (Jeshion, 2009). Neither type of information is easy to extract from text alone, although much research on multimodal models can be seen as providing a framework for the direct reference issue (Bruni et al., 2014), and selected types of text, such as social media posts and novels, may provide enough data for the extraction of social networks from word distributions (Dunbar et al., 2015; Hutchinson and Louwerse, 2018).

Another result of interest to the community of computational linguistics and NLP is the fact that our decoding results, and the associated statistical significance tests, provide some evidence that contextualized models represent individual entities in a way that is closer to what the brain does, compared with statistical distributional models or knowledge-graph methods (see Section 4.3). These results add to a recent body of work which also finds this advantage for contextualized models for both sentence processing decoding and common noun decoding (Jat et al., 2019; Schwartz and Mitchell, 2019; Sun et al., 2020a; Anderson et al., 2021), and worse performance for category-based models (Sassenhagen and Fiebach, 2020). However, we did not intend to provide an in-depth evaluation of distributional models with respect to their ability to capture semantic information about individual entities. Extensive evaluations of distributional models regarding their knowledge of entities are becoming increasingly important in NLP, testing the models' abilities to capture factual and relational knowledge (Petroni et al., 2019), similarity among entities (Newman-Griffis et al., 2018), information about entity types (Choi et al., 2018), or co-reference (Sorodoc et al., 2020) and disambiguation (He et al., 2013) phenomena (for a comprehensive set of tests, see Chen et al., 2019). In this work, however, we only used widely adopted models, whose performance rank among the best in their family (although not necessarily the best), making as clear as possible their theoretical assumptions with respect to cognitive theories of representations of entities. We assume that the core result patterns would translate also to the other models in the family. In this respect, the most similar approach to ours is that of Westera et al. (2021), where the authors show, by using just one model, Word2Vec, that better representations for categories of entities can be obtained by averaging exemplars instead of acquiring separate vectors for the categories, by evaluating the vectors on a set of human judgments.

One surprising result is that the decoding performance for common nouns referring to categories, that we reported in Figure 9 in the second, third and fourth violin plots from the left, is low. Decoding is statistically significant only when the categorical distinction between people and places makes the task easier—but even in that case, the 95% confidence interval includes the baseline, indicating that performance is not reliably above chance.

Part of the reason is clearly that much lower accuracy is obtained for decoding from EEG than for decoding from fMRI. Also, we do not have many categorical stimuli. This leaves unanswered whether our conclusions would apply also to other kinds of categories often employed in computational linguistics. To our knowledge, only one other kind of individual entities has been studied in cognitive neuroscience—brand names, whose closest notion in computational linguistics is that of organizations. Although very few studies exist (Gontijo et al., 2002; Crutch and Warrington, 2004; Cheung et al., 2010), proper names of brands seem to have similar brain processing to those of famous people; if this is the case, then we would expect that, in principle, the two should behave similarly in an experimental setup like ours.

A different question is why should categories of people be harder to decode than categories of places in classification analyses (Figure 7). We do not have a full story here, but we will just point out that social concepts are an idiosyncratic type of category, falling in between abstract and concrete concepts (Anderson et al., 2014; Rice et al., 2018; Conca et al., 2021), and that we found, consistently with the literature (Westera et al., 2021), a similar pattern of results in our clustering analysis for distributional models (Figure 2).

5.3. Methodological Innovations

From a methodological point of view, it is important to notice that we always employed a zero-shot paradigm: i.e., we always used as test samples the evoked response to stimuli not seen at training time, a procedure which ensures that we were not overfitting (Varoquaux et al., 2017; McCartney et al., 2019). As argued in Mitchell et al. (2008) and Pereira et al. (2018), this approach results in robust model which is expected to generalize well to other unseen stimuli with similar semantic properties–common nouns in their case, proper names of invididual entities in ours.

A second important characteristic of our approach is that we did not employ images as stimuli–the most common choice in EEG research for decoding studies– but words. We did this knowing that using written words results in lower classification scores compared to either spoken words or images (Simanova et al., 2010). The advantage of written words is that we could compare individual entities belonging to different coarse-grained categories (people and places), avoiding confounds such as face recognition processes (Rossion, 2014) or low-level image features (Rossion and Caharel, 2011). The fact that we nevertheless achieved above chance discrimination can be explained by the fact that we made sure to engage semantic processing by adding two experimental tasks: the mental imagery task, and the categorical question.

Regarding the mental imagery task, results in Figures 3, 5, 8, 9 suggest that it can be used, when induced by written stimuli, as a reliable way to capture semantic processing in the brain. In previous work (Shatek et al., 2019), decoding from mental imagery proved difficult to accomplish. An explanation for this variability in results, also discussed in Shatek et al. (2019), could be that whereas stimulus (word or image) processing is tightly time- and phase-locked to the stimulus presentation (it is strictly speaking “evoked” Bastiaansen et al., 2011), the mental imagery task is less bound to the stimulus presentation: the subjects do not have a clear cue or stimulus to constrain them, and the recorded signals are therefore potentially prone to showing more variance across subjects, which could ultimately make aggregated accuracies lower (Dijkstra et al., 2018; Shatek et al., 2019). Our results suggest that it is possible to induce mental imagery from written words and successfully decode it; however, how to most effectively time-lock mental imagery to experimental stimuli, as well as its temporal dynamics and the effect of imageability (Rofes et al., 2018) remain open questions.

An important feature of our experiment is that, by using written stimuli, we could directly compare semantic processing for individual entities and categories. Nevertheless, our results with transfer classification analyses (Figures 4, 6) should be interpreted with caution, because of two reasons: first, the fact of having used written words as stimuli may still induce differences in representations; second, the two levels of semantic specificity and social relevance may inherently involve different brain processing (Ross and Olson, 2012).

Regarding the former, we believe that we were able to mitigate the confounding effects of written stimuli through our control procedure for word length in the classification analyses (see Sections 3.4, 4.1), and by using machine learning algorithms which can isolate semantic processing beginning immediately after visual word recognition (Hauk et al., 2006; Penolazzi et al., 2007) and running in parallel to visual word processing. With respect to the latter, we notice that such an issue is inevitable in a setup like ours where we wanted to directly compare two inherently different levels of representation (Dehaene, 1995; Proverbio et al., 2009)—but also that, in cognitive neuroscience, this has been shown to be an approach which can reveal common representational properties across disparate kinds of brain responses (e.g., mapping between visual and auditory modalities King and Dehaene, 2014; Leonardelli et al., 2019). A different experimental paradigm could have avoided this concern; however, our analysis, due to its straightforwardness, at least suggests quite clearly that strong commonalities are not present.

5.4. Structural Properties of the Representations of Individual Entities in the Brain

From a neuroscientific point of view, throughout our analyses, we have shown that the evoked responses to individual entities and categories were consistently both hard to bring together when it came to find similarities (Figures 4, 6), and easy to tease apart when the goal was that of distinguishing them in pairs (Figure 9). These results suggest that the representations of individual entities and those of the categories involve limited common semantic processing and information—a position advocated in the past by, among others, Young et al. (1994), Barry et al. (1998), Carson and Mike Burton (2001), Turk et al. (2005), and Germain-Mondon et al. (2011). It remains an open question whether using a different experimental paradigm, and less noisy brain data acquisition methods such as fMRI, may shed some more light over the extent, as well as the location, of the interactions between these two interrelated pieces of semantic knowledge.

6. Conclusion

In this paper we explored the representation of individual entities—entities referred to by proper names—both from the point of view of cognitive neuroscience (acquiring data about their representation in the brain and investigating the structure of these representations) and from the point of view of computational linguistics and NLP (investigating the extent to which distributional representations of individual entities can be aligned to their brain representations).

More precisely, we tackled four research questions. First of all, we were able to isolate, for each individual entity, distinctive signatures in the brain (Section 4.3), and to classify them according to both their coarse and fine-grained categories (Sections 4.1, 4.2). We also found that it is difficult to transfer representational information learnt from the evoked responses to individual entities (e.g., Johnny Depp) to (the nouns for) the categories to which they belong (e.g., actor; Sections 4.2, 4.3). Finally, we provided evidence that distributional models can be mapped with statistically significant performance onto brain representation of individual entities (Section 4.3). Crucially, we were able to obtain these results using EEG, which has inherently lower signal-to-noise ratio than fMRI, but is cheaper and much more portable. This suggests that EEG can act as a useful source of data to investigate jointly brain and artificial intelligence models of language, even for extremely fine-grained semantic processes.

Data Availability Statement

The original contributions presented in the study are included in the article/Supplementary Materials, further inquiries can be directed to the corresponding author/s.

Ethics Statement

The studies involving human participants were reviewed and approved by SISSA Ethics Committee (SISSA-Via Bonomea, 265, 34136 Trieste TS, Italia) the participants provided their written informed consent to participate in this study.

Author Contributions

AB: conceptualization, methodology, software, formal analysis, investigation, data curation, visualization, and writing–original draft. MP: conceptualization, methodology, writing–review and editing, supervision, project administration, and funding acquisition. Both authors contributed to the article and approved the submitted version.

Funding

AB is supported by a doctoral studentship from the School of Electronic Engineering and Computer Science, Queen Mary University of London.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Acknowledgments

We would like to thank Prof. Davide Crepaldi, head of the Language, Learning and Reading lab at SISSA, who provided the facilities for collecting the EEG data while the first author was visiting, and Marjina Bellida for helping out with subject preparation procedures.

Supplementary Material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/frai.2022.796793/full#supplementary-material

Footnotes

1. ^https://www.biosemi.com/activetwo_full_specs.htm

2. ^Our implementation differs from the original implementation in two respects: first, we use Spearman correlation, a linear measure, whereas in Chyzhyk et al. (2018) the authors ranked and selected candidate test sets based on mutual information, which is non-linear; and secondly, here we have individual trials as candidate test items, while in Chyzhyk et al. (2018) the candidates for testing were whole subjects to be left out.

References

Adorni, R., Manfredi, M., and Proverbio, A. M. (2014). Electro-cortical manifestations of common vs. proper name processing during reading. Brain Lang. 135, 1–8. doi: 10.1016/j.bandl.2014.05.002

PubMed Abstract | CrossRef Full Text | Google Scholar

Agirre, E., and Edmonds, P. (2007). Word Sense Disambiguation: Algorithms and Applications, Vol. 33. Berlin; Heidelberg: Springer Science &Business Media.

Google Scholar

Akama, H., Miyake, M., Jung, J., and Murphy, B. (2015). Using graph components derived from an associative concept dictionary to predict fmri neural activation patterns that represent the meaning of nouns. PLoS ONE 10, e0125725. doi: 10.1371/journal.pone.0125725

PubMed Abstract | CrossRef Full Text | Google Scholar

Allison, B. Z., Kübler, A., and Jin, J. (2020). 30+ years of p300 brain-computer interfaces. Psychophysiology 57, e13569. doi: 10.1111/psyp.13569

PubMed Abstract | CrossRef Full Text | Google Scholar

Almuhareb, A., and Poesio, M. (2004). “Attribute-based and value-based clustering: an evaluation,” in Proceedings of the 2004 Conference on Empirical Methods in Natural Language Processing (Barcelona), 158–165.

Google Scholar

Anderson, A., Murphy, B., and Poesio, M. (2014). Discriminating taxonomic categories and domains in mental simulations of concepts of varying concreteness. J. Cogn. Neurosci. 26, 658–681. doi: 10.1162/jocn_a_00508

PubMed Abstract | CrossRef Full Text | Google Scholar

Anderson, A. J., Bruni, E., Bordignon, U., Poesio, M., and Baroni, M. (2013). “Of words, eyes and brains: correlating image-based distributional semantic models with neural representations of concepts,” in Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing, 1960–1970.

Google Scholar

Anderson, A. J., Kiela, D., Binder, J. R., Fernandino, L., Humphries, C. J., Conant, L. L., et al. (2021). Deep artificial neural networks reveal a distributed cortical network encoding propositional sentence-level meaning. J. Neurosci. 41, 4100–4119. doi: 10.1523/JNEUROSCI.1152-20.2021

PubMed Abstract | CrossRef Full Text | Google Scholar

Anderson, A. J., Kiela, D., Clark, S., and Poesio, M. (2017). Visually grounded and textual semantic models differentially decode brain activity associated with concrete and abstract nouns. Trans. Assoc. Comput. Linguist. 5, 17–30. doi: 10.1162/tacl_a_00043

CrossRef Full Text | Google Scholar

Anzellotti, S., and Caramazza, A. (2017). Multimodal representations of person identity individuated with fmri. Cortex 89, 85–97. doi: 10.1016/j.cortex.2017.01.013

PubMed Abstract | CrossRef Full Text | Google Scholar

Baroni, M., Dinu, G., and Kruszewski, G. (2014). “Don't count, predict! a systematic comparison of context-counting vs. context-predicting semantic vectors,” in Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Baltimore, MD: Volume 1: Long Papers), 238–247.

Google Scholar

Baroni, M., and Lenci, A. (2010). Distributional memory: a general framework for corpus-based semantics. Comput. Linguist. 36, 673–721. doi: 10.1162/coli_a_00016

CrossRef Full Text | Google Scholar

Baroni, M., Murphy, B., Barbu, E., and Poesio, M. (2010). Strudel: a distributional semantic model based on property and types. Cogn. Sci. 34, 222–254. doi: 10.1111/j.1551-6709.2009.01068.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Barry, C., Johnston, R. A., and Scanlan, L. C. (1998). Are faces" special" objects? associative and sem antic priming of face and object recognition and naming. Q. J. Exp. Psychol. A 51, 853–882. doi: 10.1080/713755783

CrossRef Full Text | Google Scholar

Barsalou, L. W. (2008). Grounded cognition. Annu. Rev. Psychol. 59, 617–645. doi: 10.1146/annurev.psych.59.103006.093639

PubMed Abstract | CrossRef Full Text | Google Scholar

Bastiaansen, M., Mazaheri, A., and Jensen, O. (2011). “Beyond erps: oscillatory neuronal,” in The Oxford Handbook of Event-Related Potential Components (Oxford), 31–50.

Google Scholar

Bazzanella, B., and Bouquet, P. (2011). “Associative and categorical priming in recognition of individuals,” in Proceedings of the Annual Meeting of the Cognitive Science Society, Vol. (Boston, MA). 33.

Google Scholar

Bender, E. M. (2011). On achieving and evaluating language-independence in nlp. Linguist. Issues Lang. Technol. 6, 1–26. doi: 10.33011/lilt.v6i.1239

CrossRef Full Text | Google Scholar

Benjamini, Y., and Hochberg, Y. (1995). Controlling the false discovery rate: a practical and powerful approach to multiple testing. J. R. Stat. Soc. B 57, 289–300. doi: 10.1111/j.2517-6161.1995.tb02031.x

CrossRef Full Text | Google Scholar

Binney, R. J., and Ramsey, R. (2020). Social semantics: the role of conceptual knowledge and cognitive control in a neurobiological model of the social brain. Neurosci. Biobehav. Rev. 112, 28–38. doi: 10.1016/j.neubiorev.2020.01.030

PubMed Abstract | CrossRef Full Text | Google Scholar

Boleda, G. (2020). Distributional semantics and linguistic theory. Ann. Rev. Linguist. 6, 213–234. doi: 10.1146/annurev-linguistics-011619-030303

CrossRef Full Text | Google Scholar

Bommasani, R., Davis, K., and Cardie, C. (2020). “Interpreting pretrained contextualized representations via reductions to static embeddings,” in Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 4758–4781.

Google Scholar

Bordes, A., Usunier, N., Garcia-Duran, A., Weston, J., and Yakhnenko, O. (2013). Translating embeddings for modeling multi-relational data. Adv. Neural Inf. Process. Syst. 26, 2787–2795. doi: 10.5555/2999792.2999923

PubMed Abstract | CrossRef Full Text | Google Scholar

Borghesani, V., Narvid, J., Battistella, G., Shwe, W., Watson, C., Binney, R. J., et al. (2019). “looks familiar, but i do not know who she is”: the role of the anterior right temporal lobe in famous face recognition. Cortex 115, 72–85. doi: 10.1016/j.cortex.2019.01.006

PubMed Abstract | CrossRef Full Text | Google Scholar

Brédart, S. (2017). The cognitive psychology and neuroscience of naming people. Neurosci. Biobehav. Rev. 83, 145–154. doi: 10.1016/j.neubiorev.2017.10.008

PubMed Abstract | CrossRef Full Text | Google Scholar

Bruce, V., and Young, A. (1986). Understanding face recognition. Br. J. Psychol. 77, 305–327. doi: 10.1111/j.2044-8295.1986.tb02199.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Bruffaerts, R., De Deyne, S., Meersmans, K., Liuzzi, A. G., Storms, G., and Vandenberghe, R. (2019). Redefining the resolution of semantic knowledge in the brain: advances made by the introduction of models of semantics in neuroimaging. Neurosci. Biobehav. Rev. 103, 3–13. doi: 10.1016/j.neubiorev.2019.05.015

PubMed Abstract | CrossRef Full Text | Google Scholar

Bruni, E., Boleda, G., Baroni, M., and Tran, N.-K. (2012). “Distributional semantics in technicolor,” in Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics, Vol. 1 (Jeju Island: Association for Computational Linguistics), 136–145.

Google Scholar

Bruni, E., Tran, N.-K., and Baroni, M. (2014). Multimodal distributional semantics. J. Artif. Intell. Res. 49, 1–47. doi: 10.1613/jair.4135

CrossRef Full Text | Google Scholar

Bullinaria, J. A., and Levy, J. P. (2007). Extracting semantic representations from word co-occurrence statistics: a computational study. Behav. Res. Methods 39, 510–526. doi: 10.3758/BF03193020

PubMed Abstract | CrossRef Full Text | Google Scholar

Burton, A. M., and Bruce, V. (1992). I recognize your face but i can't remember your name: a simple explanation? Br. J. Psychol. 83, 45–60. doi: 10.1111/j.2044-8295.1992.tb02424.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Camacho-Collados, J., and Pilehvar, M. T. (2018). From word to sense embeddings: a survey on vector representations of meaning. J. Artif. Intell. Res. 63, 743–788. doi: 10.1613/jair.1.11259

CrossRef Full Text | Google Scholar

Campanella, S., and Belin, P. (2007). Integrating face and voice in person perception. Trends Cogn. Sci. 11, 535–543. doi: 10.1016/j.tics.2007.10.001

PubMed Abstract | CrossRef Full Text | Google Scholar

Cantor, N., and Mischel, W. (1979). Prototypes in person perception. Adv. Exp. Soc. Psychol. 12, 3–52. doi: 10.1016/S0065-2601(08)60258-0

CrossRef Full Text | Google Scholar

Caramazza, A., and Mahon, B. Z. (2003). The organization of conceptual knowledge: the evidence from category-specific semantic deficits. Trends Cogn. Sci. 7, 354–361. doi: 10.1016/S1364-6613(03)00159-1

PubMed Abstract | CrossRef Full Text | Google Scholar

Caramazza, A., and Shelton, J. R. (1998). Domain-specific knowledge systems in the brain: the animate-inanimate distinction. J. Cogn. Neurosci. 10, 1–34. doi: 10.1162/089892998563752

PubMed Abstract | CrossRef Full Text | Google Scholar

Carey, S., and Xu, F. (2001). Infants' knowledge of objects: Beyond object files and object tracking. Cognition 80, 179–213. doi: 10.1016/S0010-0277(00)00154-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Carlson, G. N., and Pelletier, F. J. (1995). The Generic Book. Chicago, IL: University of Chicago Press.

Google Scholar

Carreiras, M., Armstrong, B. C., Perea, M., and Frost, R. (2014). The what, when, where, and how of visual word recognition. Trends Cogn. Sci. 18, 90–98. doi: 10.1016/j.tics.2013.11.005

PubMed Abstract | CrossRef Full Text | Google Scholar

Carson, D. R., and Mike Burton, A. (2001). Semantic priming of person recognition: categorial priming may be a weaker form of the associative priming effect. Q. J. Exp. Psychol. A 54, 1155–1179. doi: 10.1080/713756003

PubMed Abstract | CrossRef Full Text | Google Scholar

Chang, K.-M. K, Cherkassky, V. L., Mitchell, T., and Just, M. A. (2009). “Quantitative modeling of the neural representation of adjective-noun phrases to account for fmri activation,” in Proceedings of the Joint Conference of the 47th Annual Meeting of the ACL and the 4th International Joint Conference on Natural Language Processing of the AFNLP (Singapore: Association for Computational Linguistics), 638–646.

Google Scholar

Chen, M., Chu, Z., Chen, Y., Stratos, K., and Gimpel, K. (2019). “Enteval: a holistic evaluation benchmark for entity representations,” in Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP) (Hong Kong: Association for Computational Linguistics), 421–433.

Google Scholar

Cheung, M.-,c., Chan, A. S., and Sze, S. L. (2010). Electrophysiological correlates of brand names. Neurosci. Lett. 485, 178–182. doi: 10.1016/j.neulet.2010.09.006

PubMed Abstract | CrossRef Full Text | Google Scholar

Choi, E., Levy, O., Choi, Y., and Zettlemoyer, L. (2018). “Ultra-fine entity typing,” in Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Melbourne, VIC: Association for Computational Linguistics), 87–96.

Google Scholar

Chyzhyk, D., Varoquaux, G., Thirion, B., and Milham, M. (2018). “Controlling a confound in predictive models with a test set minimizing its effect,” in 2018 International Workshop on Pattern Recognition in Neuroimaging (PRNI) (Singapore: IEEE), 1–4.

Google Scholar

Conca, F., Borsa, V., Cappa, S., and Catricalà, E. (2021). The multidimensionality of abstract concepts: a systematic review. Neurosci. Biobehav. Rev. 127, 474–491. doi: 10.1016/j.neubiorev.2021.05.004

PubMed Abstract | CrossRef Full Text | Google Scholar

Crutch, S. J., and Warrington, E. K. (2004). The semantic organisation of proper nouns: the case of people and brand names. Neuropsychologia 42, 584–596. doi: 10.1016/j.neuropsychologia.2003.10.009

PubMed Abstract | CrossRef Full Text | Google Scholar

Curran, J. R., and Moens, M. (2002). “Improvements in automatic thesaurus extraction,” in Proceedings of the ACL Workshop on Unsupervised Lexical Acquisition (Philadelphia, PA: Association for Computational Linguistics), 59–66.

Google Scholar

da Silva, F. L. (2013). Eeg and meg: relevance to neuroscience. Neuron 80, 1112–1128. doi: 10.1016/j.neuron.2013.10.017

PubMed Abstract | CrossRef Full Text | Google Scholar

Darling, S., and Valentine, T. (2005). The categorical structure of semantic memory for famous people: a new approach using release from proactive interference. Cognition 96, 35–65. doi: 10.1016/j.cognition.2004.03.007

PubMed Abstract | CrossRef Full Text | Google Scholar

Dehaene, S. (1995). Evidence for category-specific word processing in the normal human brain. Neuroreport 6, 2153–2157. doi: 10.1097/00001756-199511000-00014

PubMed Abstract | CrossRef Full Text | Google Scholar

della Rocchetta, A. I., Cipolotti, L., and Warrington, E. K. (1998). Countries: Their selective impairment and selective preservation. Neurocase 4, 99–109. doi: 10.1093/neucas/4.2.99

CrossRef Full Text | Google Scholar

Devlin, J., Chang, M.-W., Lee, K., and Toutanova, K. (2019). “Bert: pre-training of deep bidirectional transformers for language understanding,” in Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers) (Minneapolis, MN: Association for Computational Linguistics), 4171–4186.

Google Scholar

Dijkstra, N., Mostert, P., de Lange, F. P., Bosch, S., and van Gerven, M. A. (2018). Differential temporal dynamics during visual imagery and perception. Elife 7:e33904. doi: 10.7554/eLife.33904

PubMed Abstract | CrossRef Full Text | Google Scholar

Djokic, V. G., Maillard, J., Bulat, L., and Shutova, E. (2020). Decoding brain activity associated with literal and metaphoric sentence comprehension using distributional semantic models. Trans. Assoc. Comput. Linguist. 8, 231–246. doi: 10.1162/tacl_a_00307

CrossRef Full Text | Google Scholar

Dror, R., Baumer, G., Bogomolov, M., and Reichart, R. (2017). Replicability analysis for natural language processing: Testing significance with multiple datasets. Trans. Assoc. Comput. Linguist. 5, 471–486. doi: 10.1162/tacl_a_00074

CrossRef Full Text | Google Scholar

Dunbar, R. I., Arnaboldi, V., Conti, M., and Passarella, A. (2015). The structure of online social networks mirrors those in the offline world. Soc. Netw. 43, 39–47. doi: 10.1016/j.socnet.2015.04.005

CrossRef Full Text | Google Scholar

Erk, K. (2012). Vector space models of word meaning and phrase meaning: A survey. Lang. Linguist. Compass 6, 635–653. doi: 10.1002/lnco.362

CrossRef Full Text | Google Scholar

Fairhall, S. L., Anzellotti, S., Ubaldi, S., and Caramazza, A. (2014). Person-and place-selective neural substrates for entity-specific semantic access. Cereb. Cortex 24, 1687–1696. doi: 10.1093/cercor/bht039

PubMed Abstract | CrossRef Full Text | Google Scholar

Fairhall, S. L., and Caramazza, A. (2013). Category-selective neural substrates for person-and place-related concepts. Cortex 49, 2748–2757. doi: 10.1016/j.cortex.2013.05.010

PubMed Abstract | CrossRef Full Text | Google Scholar

Fellbaum, C. (Ed.). (1998). WordNet: An Electronic Lexical Database. Cambridge, MA: The MIT Press.

Google Scholar

Finkelstein, L., Gabrilovich, E., Matias, Y., Rivlin, E., Solan, Z., Wolfman, G., et al. (2001). “Placing search in context: the concept revisited,” in Proceedings of the 10th International Conference on World Wide Web (New York, NY: ACM), 406–414.

Google Scholar

Firth, J. R. (1957). Papers in Linguistics, 1934-1951. London: Oxford University Press.

Google Scholar

Frank, S. L., Otten, L. J., Galli, G., and Vigliocco, G. (2015). The erp response to the amount of information conveyed by words in sentences. Brain Lang. 140, 1–11. doi: 10.1016/j.bandl.2014.10.006

PubMed Abstract | CrossRef Full Text | Google Scholar

Friston, K. J., Fletcher, P., Josephs, O., Holmes, A., Rugg, M., and Turner, R. (1998). Event-related fmri: characterizing differential responses. Neuroimage 7, 30–40. doi: 10.1006/nimg.1997.0306

PubMed Abstract | CrossRef Full Text | Google Scholar

Gardner, M., Grus, J., Neumann, M., Tafjord, O., Dasigi, P., Liu, N. F., et al. (2018). “Allennlp: a deep semantic natural language processing platform,” in Proceedings of Workshop for NLP Open Source Software (NLP-OSS) (Melbourne, VIC: Association for Computational Linguistics), 1–6.

Google Scholar

Germain-Mondon, V., Silvert, L., and Izaute, M. (2011). N400 modulation by categorical or associative interference in a famous face naming task. Neurosci. Lett. 501, 188–192. doi: 10.1016/j.neulet.2011.07.014

PubMed Abstract | CrossRef Full Text | Google Scholar

Gontijo, P. F., Rayman, J., Zhang, S., and Zaidel, E. (2002). How brand names are special: brands, words, and hemispheres. Brain Lang. 82, 327–343. doi: 10.1016/S0093-934X(02)00036-6

PubMed Abstract | CrossRef Full Text | Google Scholar

Gorno-Tempini, M., and Price, C. (2001). Identification of famous faces and buildings: a functional neuroimaging study of semantically unique items. Brain 124, 2087–2097. doi: 10.1093/brain/124.10.2087

PubMed Abstract | CrossRef Full Text | Google Scholar

Gorno-Tempini, M. L., Price, C. J., Josephs, O., Vandenberghe, R., Cappa, S. F., Kapur, N., et al. (1998). The neural systems sustaining face and proper-name processing. Brain 121, 2103–2118. doi: 10.1093/brain/121.11.2103

PubMed Abstract | CrossRef Full Text | Google Scholar

Goyal, A., Gupta, V., and Kumar, M. (2018). Recent named entity recognition and classification techniques: a systematic review. Comput. Sci. Rev. 29, 21–43. doi: 10.1016/j.cosrev.2018.06.001

CrossRef Full Text | Google Scholar

Grabowski, T. J., Damasio, H., Tranel, D., Ponto, L. L. B., Hichwa, R. D., and Damasio, A. R. (2001). A role for left temporal pole in the retrieval of words for unique entities. Hum. Brain Mapp. 13, 199–212. doi: 10.1002/hbm.1033

PubMed Abstract | CrossRef Full Text | Google Scholar

Gramfort, A., Luessi, M., Larson, E., Engemann, D. A., Strohmeier, D., Brodbeck, C., et al. (2013). MEG and EEG data analysis with MNE-Python. Front. Neurosci. 7, 267. doi: 10.3389/fnins.2013.00267

PubMed Abstract | CrossRef Full Text | Google Scholar

Grootswagers, T., Robinson, A. K., Shatek, S. M., and Carlson, T. A. (2019). Untangling featural and conceptual object representations. Neuroimage 202, 116083. doi: 10.1016/j.neuroimage.2019.116083

PubMed Abstract | CrossRef Full Text | Google Scholar

Grootswagers, T., Wardle, S. G., and Carlson, T. A. (2017). Decoding dynamic brain patterns from evoked responses: a tutorial on multivariate pattern analysis applied to time series neuroimaging data. J. Cogn. Neurosci. 29, 677–697. doi: 10.1162/jocn_a_01068

PubMed Abstract | CrossRef Full Text | Google Scholar

Groppe, D. M., Urbach, T. P., and Kutas, M. (2011). Mass univariate analysis of event-related brain potentials/fields i: a critical tutorial review. Psychophysiology 48, 1711–1725. doi: 10.1111/j.1469-8986.2011.01273.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Gu, Y., Cazzolli, G., Murphy, B., Miceli, G., and Poesio, M. (2013). Eeg study of the neural representation and classification of semantic categories of animals vs tools in young and elderly participants. BMC Neurosci. 14, 318. doi: 10.1186/1471-2202-14-S1-P318

CrossRef Full Text | Google Scholar

Guarino, N. (1995). Formal ontology, conceptual analysis and knowledge representation. Int. J. Hum. Comput. Stud. 43, 625–640. doi: 10.1006/ijhc.1995.1066

PubMed Abstract | CrossRef Full Text | Google Scholar

Günther, F., Rinaldi, L., and Marelli, M. (2019). Vector-space models of semantic representation from a cognitive perspective: a discussion of common misconceptions. Perspect. Psychol. Sci. 14, 1006–1033. doi: 10.1177/1745691619861372

PubMed Abstract | CrossRef Full Text | Google Scholar

Gupta, A., Boleda, G., and Pado, S. (2018). Instantiation. arXiv preprint arXiv:1808.01662.

Google Scholar

Han, X., Cao, S., Lv, X., Lin, Y., Liu, Z., Sun, M., et al. (2018). “Openke: an open toolkit for knowledge embedding,” in Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing: System Demonstrations (Brussels), 139–144.

Google Scholar

Harris, Z. S. (1954). Distributional structure. Word 10, 146–162. doi: 10.1080/00437956.1954.11659520

CrossRef Full Text | Google Scholar

Hauk, O., Davis, M. H., Ford, M., Pulvermüller, F., and Marslen-Wilson, W. D. (2006). The time course of visual word recognition as revealed by linear regression analysis of erp data. Neuroimage 30, 1383–1400. doi: 10.1016/j.neuroimage.2005.11.048

PubMed Abstract | CrossRef Full Text | Google Scholar

Haxby, J. V., Gobbini, M. I., Furey, M. L., Ishai, A., Schouten, J. L., and Pietrini, P. (2001). Distributed and overlapping representations of faces and objects in ventral temporal cortex. Science 293, 2425–2430. doi: 10.1126/science.1063736

PubMed Abstract | CrossRef Full Text | Google Scholar

He, Z., Liu, S., Li, M., Zhou, M., Zhang, L., and Wang, H. (2013). “Learning entity representation for entity disambiguation,” in Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics, Vol. 2 (Sofia), 30–34.

Google Scholar

Helwig, N. E. (2019). Statistical nonparametric mapping: multivariate permutation tests for location, correlation, and regression problems in neuroimaging. Wiley Interdisc. Rev. 11, e1457. doi: 10.1002/wics.1457

CrossRef Full Text | Google Scholar

Hill, F., Reichart, R., and Korhonen, A. (2015). Simlex-999: evaluating semantic models with (genuine) similarity estimation. Computat. Linguist. 41, 665–695. doi: 10.1162/COLI_a_00237

CrossRef Full Text | Google Scholar

Hollenstein, N., de la Torre, A., Langer, N., and Zhang, C. (2019). “Cognival: A framework for cognitive word embedding evaluation,” in Proceedings of the 23rd Conference on Computational Natural Language Learning (CoNLL) (Hong Kong), 538–549.

Google Scholar

Hubert, L., and Arabie, P. (1985). Comparing partitions. J. Classificat. 2, 193–218. doi: 10.1007/BF01908075

CrossRef Full Text | Google Scholar

Hutchinson, S., and Louwerse, M. (2018). Extracting social networks from language statistics. Discourse Process. 55, 607–618. doi: 10.1080/0163853X.2017.1332446

CrossRef Full Text | Google Scholar

Huth, A. G., De Heer, W. A., Griffiths, T. L., Theunissen, F. E., and Gallant, J. L. (2016). Natural speech reveals the semantic maps that tile human cerebral cortex. Nature 532, 453–458. doi: 10.1038/nature17637

PubMed Abstract | CrossRef Full Text | Google Scholar

Jas, M., Engemann, D. A., Bekhti, Y., Raimondo, F., and Gramfort, A. (2017). Autoreject: automated artifact rejection for meg and eeg data. Neuroimage 159, 417–429. doi: 10.1016/j.neuroimage.2017.06.030

PubMed Abstract | CrossRef Full Text | Google Scholar

Jas, M., Larson, E., Engemann, D. A., Leppäkangas, J., Taulu, S., Hämäläinen, M., et al. (2018). A reproducible meg/eeg group study with the mne software: recommendations, quality assessments, and good practices. Front. Neurosci. 12, 530. doi: 10.3389/fnins.2018.00530

PubMed Abstract | CrossRef Full Text | Google Scholar

Jat, S., Tang, H., Talukdar, P., and Mitchell, T. (2019). “Relating simple sentence representations in deep neural networks and the brain,” in Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (Florence: Association for Computational Linguistics), 5137–5154.

Google Scholar

Jelodar, A. B., Alizadeh, M., and Khadivi, S. (2010). “Wordnet based features for predicting brain activity associated with meanings of nouns,” in Proceedings of the NAACL HLT 2010 First Workshop on Computational Neurolinguistics (Los Angeles, CA), 18–26.

Google Scholar

Jeshion, R. (2009). The significance of names. Mind Lang. 24, 370–403. doi: 10.1111/j.1468-0017.2009.01367.x

CrossRef Full Text | Google Scholar

Ji, S., Pan, S., Cambria, E., Marttinen, P., and Philip, S. Y. (2021). A survey on knowledge graphs: representation, acquisition, and applications. IEEE Trans. Neural Netw. Learn. Syst. 33, 494–514. doi: 10.1109/TNNLS.2021.3070843

PubMed Abstract | CrossRef Full Text | Google Scholar

Jones, M. N., and Mewhort, D. J. (2007). Representing word meaning and order information in a composite holographic lexicon. Psychol. Rev. 114, 1. doi: 10.1037/0033-295X.114.1.1

PubMed Abstract | CrossRef Full Text | Google Scholar

Kahneman, D., Treisman, A., and Gibbs, B. J. (1992). The reviewing of object files: Object-specific integration of information. Cogn. Psychol. 24, 175–219. doi: 10.1016/0010-0285(92)90007-O

PubMed Abstract | CrossRef Full Text | Google Scholar

Kaiser, D., Häberle, G., and Cichy, R. M. (2020). Real-world structure facilitates the rapid emergence of scene category information in visual brain signals. J. Neurophysiol. 124, 145–151. doi: 10.1152/jn.00164.2020

PubMed Abstract | CrossRef Full Text | Google Scholar

Kemmerer, D. (2014). Cognitive Neuroscience of Language. New York, NY: Psychology Press.

Google Scholar

Kiela, D., and Clark, S. (2015). “Multi-and cross-modal semantics beyond vision: Grounding in auditory perception,” in Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing (Lisbon), 2461–2470.

Google Scholar

King, J.-R., and Dehaene, S. (2014). Characterizing the dynamics of mental representations: the temporal generalization method. Trends Cogn. Sci. 18, 203–210. doi: 10.1016/j.tics.2014.01.002

PubMed Abstract | CrossRef Full Text | Google Scholar

Klapper, A., Dotsch, R., van Rooij, I., and Wigboldus, D. (2017). Four meanings of “categorization”: a conceptual analysis of research on person perception. Soc. Pers. Psychol. Compass. 11, e12336. doi: 10.1111/spc3.12336

CrossRef Full Text | Google Scholar

Kripke, S. A. (1972). “Naming and necessity,” in Semantics of Natural Language (Berlin; Heidelberg: Springer), 253–355.

Google Scholar

Kutas, M., and Federmeier, K. D. (2011). Thirty years and counting: finding meaning in the n400 component of the event-related brain potential (erp). Annu. Rev. Psychol. 62, 621–647. doi: 10.1146/annurev.psych.093008.131123

PubMed Abstract | CrossRef Full Text | Google Scholar

Landauer, T. K., and Dumais, S. T. (1997). A solution to plato's problem: the latent semantic analysis theory of acquisition, induction, and representation of knowledge. Psychol. Rev. 104, 211. doi: 10.1037/0033-295X.104.2.211

CrossRef Full Text | Google Scholar

Lau, E. F., Phillips, C., and Poeppel, D. (2008). A cortical network for semantics:(de) constructing the n400. Nat. Rev. Neurosci. 9, 920–933. doi: 10.1038/nrn2532

PubMed Abstract | CrossRef Full Text | Google Scholar

Lenci, A. (2008). Distributional semantics in linguistic and cognitive research. Italian journal of linguistics 20, 1–31.

Google Scholar

Lenci, A., Sahlgren, M., Jeuniaux, P., Gyllensten, A. C., and Miliani, M. (2021). A comprehensive comparative evaluation and analysis of distributional semantic models. arXiv preprint arXiv:2105.09825.

Google Scholar

Leonardelli, E., Fait, E., and Fairhall, S. L. (2019). Temporal dynamics of access to amodal representations of category-level conceptual information. Sci. Rep. 9, 1–9. doi: 10.1038/s41598-018-37429-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Leslie, A. M., Xu, F., Tremoulet, P. D., and Scholl, B. J. (1998). Indexing and the object concept: developingwhat'andwhere'systems. Trends Cogn. Sci. 2, 10–18. doi: 10.1016/S1364-6613(97)01113-3

PubMed Abstract | CrossRef Full Text | Google Scholar

Lin, D. (1998). “Automatic retrieval and clustering of similar words,” in Proceedings of COLING-ACL (Montreal, QC).

PubMed Abstract | Google Scholar

Lindsay, G. W. (2020). Attention in psychology, neuroscience, and machine learning. Front. Comput. Neurosci. 14, 29. doi: 10.3389/fncom.2020.00029

PubMed Abstract | CrossRef Full Text | Google Scholar

Ling, S., Lee, A. C., Armstrong, B. C., and Nestor, A. (2019). How are visual words represented? insights from eeg-based visual word decoding, feature derivation and image reconstruction. Hum. Brain Mapp. 40, 5056–5068. doi: 10.1002/hbm.24757

PubMed Abstract | CrossRef Full Text | Google Scholar

Lowe, E. J. (2003). “Individuation,” in The Oxford Handbook of Metaphysics (Oxford).

Google Scholar

Luck, S. J. (2014). An introduction to the event-related potential technique. Cambridge, MA: Mit Press.

Google Scholar

Lund, K., and Burgess, C. (1996). Producing high-dimensional semantic spaces from lexical co-occurrence. Behav. Res. Methods Instruments Comput. 28, 203–208. doi: 10.3758/BF03204766

CrossRef Full Text | Google Scholar

Lyons, F., Hanley, J. R., and Kay, J. (2002). Anomia for common names and geographical names with preserved retrieval of names of people: a semantic memory disorder. Cortex 38, 23–35. doi: 10.1016/S0010-9452(08)70636-1

PubMed Abstract | CrossRef Full Text | Google Scholar

Mahon, B. Z., and Caramazza, A. (2009). Concepts and categories: a cognitive neuropsychological perspective. Annu. Rev. Psychol. 60, 27–51. doi: 10.1146/annurev.psych.60.110707.163532

PubMed Abstract | CrossRef Full Text | Google Scholar

Mandera, P., Keuleers, E., and Brysbaert, M. (2017). Explaining human performance in psycholinguistic tasks with models of semantic similarity based on prediction and counting: a review and empirical validation. J. Mem. Lang. 92, 57–78. doi: 10.1016/j.jml.2016.04.001

CrossRef Full Text | Google Scholar

Martins, I. P., and Farrajota, L. (2007). Proper and common names: A double dissociation. Neuropsychologia 45, 1744–1756. doi: 10.1016/j.neuropsychologia.2006.12.016

PubMed Abstract | CrossRef Full Text | Google Scholar

Mason, M. F., and Macrae, C. N. (2004). Categorizing and individuating others: The neural substrates of person perception. J. Cogn. Neurosci. 16, 1785–1795. doi: 10.1162/0898929042947801

PubMed Abstract | CrossRef Full Text | Google Scholar

McCarthy, R. A., and Warrington, E. K. (2016). Past, present, and prospects: Reflections 40 years on from the selective impairment of semantic memory (warrington, 1975). Q. J. Exp. Psychol. 69, 1941–1968. doi: 10.1080/17470218.2014.980280

PubMed Abstract | CrossRef Full Text | Google Scholar

McCartney, B., Martinez-del Rincon, J., Devereux, B., and Murphy, B. (2019). A zero-shot learning approach to the development of brain-computer interfaces for image retrieval. PLoS ONE 14, e0214342. doi: 10.1371/journal.pone.0214342

PubMed Abstract | CrossRef Full Text | Google Scholar

Mensen, A., and Khatami, R. (2013). Advanced eeg analysis using threshold-free cluster-enhancement and non-parametric statistics. Neuroimage 67, 111–118. doi: 10.1016/j.neuroimage.2012.10.027

PubMed Abstract | CrossRef Full Text | Google Scholar

Miceli, G., Capasso, R., Daniele, A., Esposito, T., Magarelli, M., and Tomaiuolo, F. (2000). Selective deficit for people's names following left temporal damage: an impairment of domain-specific conceptual knowledge. Cogn. Neuropsychol. 17, 489–516. doi: 10.1080/02643290050110629

PubMed Abstract | CrossRef Full Text | Google Scholar

Michel, C. M., and Murray, M. M. (2012). Towards the utilization of eeg as a brain imaging tool. Neuroimage 61, 371–385. doi: 10.1016/j.neuroimage.2011.12.039

PubMed Abstract | CrossRef Full Text | Google Scholar

Mikolov, T., Sutskever, I., Chen, K., Corrado, G. S., and Dean, J. (2013). “Distributed representations of words and phrases and their compositionality,” in Advances in Neural Information Processing Systems, 3111–3119.

PubMed Abstract | Google Scholar

Mitchell, T., Shinkareva, S. V., Carlson, A., Chang, K.-M., Malave, V. L., Mason, R. A., et al. (2008). Predicting human brain activity associated with the meanings of nouns. Science 320, 1191–1195. doi: 10.1126/science.1152876

PubMed Abstract | CrossRef Full Text | Google Scholar

Moore, V., and Valentine, T. (2020). “The effects of age of acquisition in processing famous faces and names: exploring the locus and proposing a mechanism,” in Proceedings of the Twenty First Annual Conference of the Cognitive Science Society (New York, NY: Psychology Press), 416–421.

Google Scholar

More, S., Eickhoff, S. B., Caspers, J., and Patil, K. R. (2020). Confound removal and normalization in practice: a neuroimaging based sex prediction case study. Mach. Learn. Knowl. Discovery Databases Appl. Data Sci. Demo Track 12461, 3. doi: 10.1007/978-3-030-67670-4_1

CrossRef Full Text | Google Scholar

Morton, N. W., Zippi, E. L., Noh, S. M., and Preston, A. R. (2021). Semantic knowledge of famous people and places is represented in hippocampus and distinct cortical networks. J. Neurosci. 41, 2762–2779. doi: 10.1523/JNEUROSCI.2034-19.2021

PubMed Abstract | CrossRef Full Text | Google Scholar

Murez, M., and Recanati, F. (2016). Mental files: an introduction. Rev. Philos. Psychol. 7, 265–281. doi: 10.1007/s13164-016-0314-3

CrossRef Full Text | Google Scholar

Murphy, B., Baroni, M., and Poesio, M. (2009). “Eeg responds to conceptual stimuli and corpus semantics,” in Proceedings of the 2009 Conference on Empirical Methods in Natural Language Processing (Singapore), 619–627.

Google Scholar

Murphy, B., Dalponte, M., Poesio, M., and Bruzzone, L. (2008). “Distinguishing concept categories on the basis of single-participant electrophysiological activity,” in Proceedings of The Annual Meeting of the Cognitive Science Society (Washington, DC).

Google Scholar

Murphy, B., Poesio, M., Bovolo, F., Bruzzone, L., Dalponte, M., and Lakany, H. (2011). Eeg decoding of semantic category reveals distributed representations for single concepts. Brain Lang. 117, 12–22. doi: 10.1016/j.bandl.2010.09.013

PubMed Abstract | CrossRef Full Text | Google Scholar

Murphy, B., Talukdar, P., and Mitchell, T. (2012). “Selecting corpus-semantic models for neurolinguistic decoding,” in * SEM 2012: The First Joint Conference on Lexical and Computational Semantics-Volume 1: Proceedings of the Main Conference and the Shared Task, and Volume 2: Proceedings of the Sixth International Workshop on Semantic Evaluation (SemEval 2012) (Montreal, QC), 114–123.

Google Scholar

Newman-Griffis, D., Lai, A. M., and Fosler-Lussier, E. (2018). “Jointly embedding entities and text with distant supervision,” in Proceedings of The Third Workshop on Representation Learning for NLP (Melbourne, VIC), 195–206.

Google Scholar

Nishida, S., and Nishimoto, S. (2018). Decoding naturalistic experiences from human brain activity via distributed representations of words. Neuroimage 180, 232–242. doi: 10.1016/j.neuroimage.2017.08.017

PubMed Abstract | CrossRef Full Text | Google Scholar

Olson, I. R., McCoy, D., Klobusicky, E., and Ross, L. A. (2013). Social cognition and the anterior temporal lobes: a review and theoretical framework. Soc. Cogn. Affect. Neurosci. 8, 123–133. doi: 10.1093/scan/nss119

PubMed Abstract | CrossRef Full Text | Google Scholar

Olson, I. R., Plotzker, A., and Ezzyat, Y. (2007). The enigmatic temporal pole: a review of findings on social and emotional processing. Brain 130, 1718–1731. doi: 10.1093/brain/awm052

PubMed Abstract | CrossRef Full Text | Google Scholar

Padó, S., and Lapata, M. (2007). Dependency-based construction of semantic space models. Comput. Linguist. 33, 161–199. doi: 10.1162/coli.2007.33.2.161

CrossRef Full Text | Google Scholar

Pedregosa, F., Varoquaux, G., Gramfort, A., Michel, V., Thirion, B., Grisel, O., et al. (2011). Scikit-learn: machine learning in python. J. Mach. Learn. Res. 12, 2825–2830. doi: 10.5555/1953048.2078195

CrossRef Full Text | Google Scholar

Pennington, J., Socher, R., and Manning, C. D. (2014). “Glove: global vectors for word representation,” in Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP) (Doha), 1532–1543.

Google Scholar

Penolazzi, B., Hauk, O., and Pulvermüller, F. (2007). Early semantic context integration and lexical access as revealed by event-related brain potentials. Biol. Psychol. 74, 374–388. doi: 10.1016/j.biopsycho.2006.09.008

PubMed Abstract | CrossRef Full Text | Google Scholar

Pereira, F., Botvinick, M., and Detre, G. (2013). Using wikipedia to learn semantic feature representations of concrete concepts in neuroimaging experiments. Artif. Intell. 194, 240–252. doi: 10.1016/j.artint.2012.06.005

PubMed Abstract | CrossRef Full Text | Google Scholar

Pereira, F., Lou, B., Pritchett, B., Ritter, S., Gershman, S. J., Kanwisher, N., et al. (2018). Toward a universal decoder of linguistic meaning from brain activation. Nat. Commun. 9, 1–13. doi: 10.1038/s41467-018-03068-4

PubMed Abstract | CrossRef Full Text | Google Scholar

Peressotti, F., Cubelli, R., and Job, R. (2003). On recognizing proper names: The orthographic cue hypothesis. Cogn. Psychol. 47, 87–116. doi: 10.1016/S0010-0285(03)00004-5

PubMed Abstract | CrossRef Full Text | Google Scholar

Peters, M. E., Neumann, M., Iyyer, M., Gardner, M., Clark, C., Lee, K., et al. (2018). “Deep contextualized word representations,” in Proceedings of NAACL-HLT (New Orleans, LA), 2227–2237.

PubMed Abstract | Google Scholar

Peters, M. E., Neumann, M., Logan, R., Schwartz, R., Joshi, V., Singh, S., et al. (2019). “Knowledge enhanced contextual word representations,” in Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP) (Hong Kong).

Google Scholar

Petit, S., Badcock, N. A., Grootswagers, T., and Woolgar, A. (2020). Unconstrained multivariate eeg decoding can help detect lexical-semantic processing in individual children. Sci. Rep. 10, 1–15. doi: 10.1038/s41598-020-67407-6

PubMed Abstract | CrossRef Full Text | Google Scholar

Petroni, F., Rocktäschel, T., Riedel, S., Lewis, P., Bakhtin, A., Wu, Y., et al. (2019). “Language models as knowledge bases?” in Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP) (Hong Kong), 2463–2473.

Google Scholar

Pires, T., Schlinger, E., and Garrette, D. (2019). “How multilingual is multilingual bert?” in Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (Florence), 4996–5001.

Google Scholar

Proverbio, A. M., Mariani, S., Zani, A., and Adorni, R. (2009). How are ‘barack obama' and ‘president elect' differentially stored in the brain? an erp investigation on the processing of proper and common noun pairs. PLoS ONE 4, e7126. doi: 10.1371/journal.pone.0007126

PubMed Abstract | CrossRef Full Text | Google Scholar

Rabovsky, M., Hansen, S. S., and McClelland, J. L. (2018). Modelling the n400 brain potential as change in a probabilistic representation of meaning. Nat. Hum. Behav. 2, 693–705. doi: 10.1038/s41562-018-0406-4

PubMed Abstract | CrossRef Full Text | Google Scholar

Ralph, M. A. L., Jefferies, E., Patterson, K., and Rogers, T. T. (2017). The neural and computational bases of semantic cognition. Nat. Rev. Neurosci. 18, 42–55. doi: 10.1038/nrn.2016.150

PubMed Abstract | CrossRef Full Text | Google Scholar

Rice, G. E., Hoffman, P., Binney, R. J., and Lambon Ralph, M. A. (2018). Concrete versus abstract forms of social concept: an fmri comparison of knowledge about people versus social terms. Philos. Trans. R. Soc. B Biol. Sci. 373, 20170136. doi: 10.1098/rstb.2017.0136

PubMed Abstract | CrossRef Full Text | Google Scholar

Rofes, A., Zakariás, L., Ceder, K., Lind, M., Johansson, M. B., De Aguiar, V., et al. (2018). Imageability ratings across languages. Behav. Res. Methods 50, 1187–1197. doi: 10.3758/s13428-017-0936-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Rogers, A., Kovaleva, O., and Rumshisky, A. (2020). A primer in bertology: What we know about how bert works. Trans. Assoc. Comput. Linguist. 8, 842–866. doi: 10.1162/tacl_a_00349

CrossRef Full Text | Google Scholar

Rosch, E. (1975). Cognitive representations of semantic categories. J. Exp. Psychol. Gen. 104, 192. doi: 10.1037/0096-3445.104.3.192

CrossRef Full Text | Google Scholar

Ross, L. A., and Olson, I. R. (2012). What's unique about unique entities? an fmri investigation of the semantics of famous faces and landmarks. Cereb. Cortex 22, 2005–2015. doi: 10.1093/cercor/bhr274

PubMed Abstract | CrossRef Full Text | Google Scholar

Rossion, B. (2014). Understanding face perception by means of human electrophysiology. Trends Cogn. Sci. 18, 310–318. doi: 10.1016/j.tics.2014.02.013

PubMed Abstract | CrossRef Full Text | Google Scholar

Rossion, B., and Caharel, S. (2011). Erp evidence for the speed of face categorization in the human brain: disentangling the contribution of low-level visual cues from face perception. Vision Res. 51, 1297–1311. doi: 10.1016/j.visres.2011.04.003

PubMed Abstract | CrossRef Full Text | Google Scholar

Sassenhagen, J., and Fiebach, C. J. (2020). Traces of meaning itself: Encoding distributional word vectors in brain activity. Neurobiol. Lang. 1, 54–76. doi: 10.1162/nol_a_00003

CrossRef Full Text | Google Scholar

Schneider, B., Heskje, J., Bruss, J., Tranel, D., and Belfi, A. M. (2018). The left temporal pole is a convergence region mediating the relation between names and semantic knowledge for unique entities: further evidence from a “recognition-from-name” study in neurological patients. Cortex 109, 14–24. doi: 10.1016/j.cortex.2018.08.026

PubMed Abstract | CrossRef Full Text | Google Scholar

Schütze, H. (1997). Ambiguity Resolution in Language Learning. Stanford, CA: CSLI.

Google Scholar

Schwartz, D., and Mitchell, T. (2019). “Understanding language-elicited eeg data by predicting it from a fine-tuned language model,” in Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers) (Minneapolis, MN), 43–57.

Google Scholar

Schweinberger, S. R. (1996). How gorbachev primed yeltsin: Analyses of associative priming in person recognition by means of reaction times and event-related brain potentials. J. Exp. Psychol. Learn. Mem. Cogn. 22, 1383. doi: 10.1037/0278-7393.22.6.1383

CrossRef Full Text | Google Scholar

Semenza, C. (2006). Retrieval pathways for common and proper names. Cortex 42, 884–891. doi: 10.1016/S0010-9452(08)70432-5

PubMed Abstract | CrossRef Full Text | Google Scholar

Semenza, C. (2009). The neuropsychology of proper names. Mind Lang. 24, 347–369. doi: 10.1111/j.1468-0017.2009.01366.x

CrossRef Full Text | Google Scholar

Semenza, C., and Zettin, M. (1989). Evidence from aphasia for the role of proper names as pure referring expressions. Nature 342, 678–679. doi: 10.1038/342678a0

PubMed Abstract | CrossRef Full Text | Google Scholar

Shatek, S. M., Grootswagers, T., Robinson, A. K., and Carlson, T. A. (2019). Decoding images in the mind's eye: the temporal dynamics of visual imagery. Vision 3, 53. doi: 10.3390/vision3040053

PubMed Abstract | CrossRef Full Text | Google Scholar

Simanova, I., Van Gerven, M., Oostenveld, R., and Hagoort, P. (2010). Identifying object categories from event-related eeg: toward decoding of conceptual representations. PLoS ONE 5, e14465. doi: 10.1371/journal.pone.0014465

PubMed Abstract | CrossRef Full Text | Google Scholar

Smith, S. M., and Nichols, T. E. (2009). Threshold-free cluster enhancement: addressing problems of smoothing, threshold dependence and localisation in cluster inference. Neuroimage 44, 83–98. doi: 10.1016/j.neuroimage.2008.03.061

PubMed Abstract | CrossRef Full Text | Google Scholar

Smith-Spark, J. H., Moore, V., Valentine, T., and Sherman, S. M. (2006). Stimulus generation, ratings, phoneme counts, and group classifications for 696 famous people by british adults over 40 years of age. Behav. Res. Methods 38, 590–597. doi: 10.3758/BF03193890

PubMed Abstract | CrossRef Full Text | Google Scholar

Sorodoc, I., Gulordava, K., and Boleda, G. (2020). “Probing for referential information in language models,” in Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 4177–4189.

Google Scholar

Stone, A. (2008). Categorical priming of famous person recognition: A hitherto overlooked methodological factor can resolve a long-standing debate. Cognition 108, 874–880. doi: 10.1016/j.cognition.2008.06.001

PubMed Abstract | CrossRef Full Text | Google Scholar

Stone, A., and Valentine, T. (2007). The categorical structure of knowledge for famous people (and a novel application of centre-surround theory). Cognition 104, 535–564. doi: 10.1016/j.cognition.2006.07.014

PubMed Abstract | CrossRef Full Text | Google Scholar

Strawson, P. F. (1950). On referring. Mind 59, 320–344. doi: 10.1093/mind/LIX.235.320

CrossRef Full Text | Google Scholar

Sudre, G., Pomerleau, D., Palatucci, M., Wehbe, L., Fyshe, A., Salmelin, R., et al. (2012). Tracking neural coding of perceptual and semantic features of concrete nouns. Neuroimage 62, 451–463. doi: 10.1016/j.neuroimage.2012.04.048

PubMed Abstract | CrossRef Full Text | Google Scholar

Sun, J., Wang, S., Zhang, J., and Zong, C. (2020a). Neural encoding and decoding with distributed sentence representations. IEEE Trans. Neural Netw. Learn. Syst. 32, 589–603. doi: 10.1109/TNNLS.2020.3027595

PubMed Abstract | CrossRef Full Text | Google Scholar

Sun, Y., Wang, S., Li, Y., Feng, S., Tian, H., Wu, H., et al. (2020b). Ernie 2.0: A continual pre-training framework for language understanding. Proc. AAAI Conf. Artif. Intell. 34, 8968–8975. doi: 10.1609/aaai.v34i05.6428

CrossRef Full Text | Google Scholar

Tanner, D., Norton, J. J., Morgan-Short, K., and Luck, S. J. (2016). On high-pass filter artifacts (they're real) and baseline correction (it'sa good idea) in erp/ermf analysis. J. Neurosci. Methods 266, 166–170. doi: 10.1016/j.jneumeth.2016.01.002

PubMed Abstract | CrossRef Full Text | Google Scholar

Tsantani, M., Kriegeskorte, N., McGettigan, C., and Garrido, L. (2019). Faces and voices in the brain: a modality-general person-identity representation in superior temporal sulcus. Neuroimage 201, 116004. doi: 10.1016/j.neuroimage.2019.07.017

PubMed Abstract | CrossRef Full Text | Google Scholar

Turk, D. J., Rosenblum, A. C., Gazzaniga, M. S., and Macrae, C. N. (2005). Seeing john malkovich: the neural substrates of person categorization. Neuroimage 24, 1147–1153. doi: 10.1016/j.neuroimage.2004.10.032

PubMed Abstract | CrossRef Full Text | Google Scholar

Turney, P. D. (2001). “Mining the web for synonyms: PMI-IR versus LSA on the TOEFL,” in Proceedings of 12th European Conference on Machine Learning (ECML) (Freiburg: ACM), 491–502.

Google Scholar

Urigüen, J. A., and Garcia-Zapirain, B. (2015). Eeg artifact removal—state-of-the-art and guidelines. J. Neural Eng. 12, 031001. doi: 10.1088/1741-2560/12/3/031001

PubMed Abstract | CrossRef Full Text | Google Scholar

Valentine, V. M. T. (1998). The effect of age of acquisition on speed and accuracy of naming famous faces. Q. J. Exp. Psychol. A 51, 485–513. doi: 10.1080/027249898391503

PubMed Abstract | CrossRef Full Text | Google Scholar

van Driel, J., Olivers, C. N., and Fahrenfort, J. J. (2021). High-pass filtering artifacts in multivariate classification of neural time series data. J. Neurosci. Methods 352, 109080. doi: 10.1016/j.jneumeth.2021.109080

PubMed Abstract | CrossRef Full Text | Google Scholar

Van Langendonck, W., and Van de Velde, M. (2016). “Names and gramma,”. in The Oxford Handbook of Names and Naming (Oxford).

Google Scholar

Varoquaux, G., Raamana, P. R., Engemann, D. A., Hoyos-Idrobo, A., Schwartz, Y., and Thirion, B. (2017). Assessing and tuning brain decoders: cross-validation, caveats, and guidelines. Neuroimage 145, 166–179. doi: 10.1016/j.neuroimage.2016.10.038

PubMed Abstract | CrossRef Full Text | Google Scholar

Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., et al. (2017). “Attention is all you need,” in Advances in Neural Information Processing Systems (Long Beach, CA), 5998–6008.

Google Scholar

Vladeanu, M., Lewis, M., and Ellis, H. (2006). Associative priming in faces: Semantic relatedness or simple co-occurrence? Mem. Cogn. 34, 1091–1101. doi: 10.3758/BF03193255

PubMed Abstract | CrossRef Full Text | Google Scholar

Vrandečić, D., and Krötzsch, M. (2014). Wikidata: a free collaborative knowledgebase. Commun. ACM. 57, 78–85. doi: 10.1145/2629489

CrossRef Full Text | Google Scholar

Warrington, E. K., and Shallice, T. (1984). Category specific semantic impairments. Brain 107, 829–853. doi: 10.1093/brain/107.3.829

PubMed Abstract | CrossRef Full Text | Google Scholar

Wehbe, L., Murphy, B., Talukdar, P., Fyshe, A., Ramdas, A., and Mitchell, T. (2014a). Simultaneously uncovering the patterns of brain regions involved in different story reading subprocesses. PLoS ONE 9, e112575. doi: 10.1371/journal.pone.0112575

PubMed Abstract | CrossRef Full Text | Google Scholar

Wehbe, L., Vaswani, A., Knight, K., and Mitchell, T. (2014b). “Aligning context-based statistical models of language with brain activity during reading,” in Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP) (Doha), 233–243.

Google Scholar

Westera, M., Gupta, A., Boleda, G., and Padó, S. (2021). Distributional models of category concepts based on names of category members. Cogn. Sci. 45, e13029. doi: 10.1111/cogs.13029

PubMed Abstract | CrossRef Full Text | Google Scholar

Wiese, H. (2011). The structure of semantic person memory: Evidence from semantic priming in person recognition. Br. J. Psychol. 102, 899–914. doi: 10.1111/j.2044-8295.2011.02042.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Wiese, H., and Schweinberger, S. R. (2008). Event-related potentials indicate different processes to mediate categorical and associative priming in person recognition. J. Exp. Psychol. Learn. Mem. Cogn. 34, 1246. doi: 10.1037/a0012937

PubMed Abstract | CrossRef Full Text | Google Scholar

Wiese, H., and Schweinberger, S. R. (2011). Accessing semantic person knowledge: Temporal dynamics of nonstrategic categorical and associative priming. J. Cogn. Neurosci. 23, 447–459. doi: 10.1162/jocn.2010.21432

PubMed Abstract | CrossRef Full Text | Google Scholar

Wittgenstein, L. (1953). Philosophical Investigations. Oxford: Basil Blackwell.

Google Scholar

Wolf, T., Chaumond, J., Debut, L., Sanh, V., Delangue, C., Moi, A., et al. (2020). “Transformers: state-of-the-art natural language processing,” in Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, 38–45.

Google Scholar

Yamada, I., Asai, A., Sakuma, J., Shindo, H., Takeda, H., Takefuji, Y., et al. (2020a). “Wikipedia2vec: an efficient toolkit for learning and visualizing the embeddings of words and entities from wikipedia,” in Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, 23–30.

Google Scholar

Yamada, I., Asai, A., Shindo, H., Takeda, H., and Matsumoto, Y. (2020b). “Luke: deep contextualized entity representations with entity-aware self-attention,” in Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), 6442–6454.

Google Scholar

Young, A. W. (1999). Simulating face recognition: Implications for modelling cognition. Cogn. Neuropsychol. 16, 1–48. doi: 10.1080/026432999380960

CrossRef Full Text | Google Scholar

Young, A. W., Flude, B. M., Hellawell, D. J., and Ellis, A. W. (1994). The nature of semantic priming effects in the recognition of familiar people. Br. J. Psychol. 85, 393–411. doi: 10.1111/j.2044-8295.1994.tb02531.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Young, A. W., Frühholz, S., and Schweinberger, S. R. (2020). Face and voice perception: Understanding commonalities and differences. Trends Cogn. Sci. 24, 398–410. doi: 10.1016/j.tics.2020.02.001

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhang, Y., Han, K., Worth, R., and Liu, Z. (2020). Connecting concepts in the brain by mapping cortical representations of semantic relations. Nat. Commun. 11, 1–13. doi: 10.1038/s41467-020-15804-w

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: brain decoding, proper names, individual entities, distributional semantics, language models, EEG, categories

Citation: Bruera A and Poesio M (2022) Exploring the Representations of Individual Entities in the Brain Combining EEG and Distributional Semantics. Front. Artif. Intell. 5:796793. doi: 10.3389/frai.2022.796793

Received: 17 October 2021; Accepted: 25 January 2022;
Published: 23 February 2022.

Edited by:

Sebastian Padó, University of Stuttgart, Germany

Reviewed by:

Constantin Orasan, University of Surrey, United Kingdom
Diego Frassinelli, University of Konstanz, Germany

Copyright © 2022 Bruera and Poesio. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Andrea Bruera, a.bruera@qmul.ac.uk

Download