<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Neurosci.</journal-id>
<journal-title>Frontiers in Neuroscience</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Neurosci.</abbrev-journal-title>
<issn pub-type="epub">1662-453X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fnins.2022.825746</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Neuroscience</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>When the Whole Is Less Than the Sum of Its Parts: Maximum Object Category Information and Behavioral Prediction in Multiscale Activation Patterns</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name><surname>Karimi-Rouzbahani</surname> <given-names>Hamid</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/1275232/overview"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Woolgar</surname> <given-names>Alexandra</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/10946/overview"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>Medical Research Council Cognition and Brain Sciences Unit, University of Cambridge</institution>, <addr-line>Cambridge</addr-line>, <country>United Kingdom</country></aff>
<aff id="aff2"><sup>2</sup><institution>Department of Cognitive Science, Perception in Action Research Centre, Macquarie University</institution>, <addr-line>Sydney, NSW</addr-line>, <country>Australia</country></aff>
<aff id="aff3"><sup>3</sup><institution>Department of Computing, Macquarie University</institution>, <addr-line>Sydney, NSW</addr-line>, <country>Australia</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Reza Lashgari, Shahid Beheshti University, Iran</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Abed Khorasani, Kerman Medical University, Iran; Mohsen Bafandehkar, Macquarie University, Australia</p></fn>
<corresp id="c001">&#x002A;Correspondence: Hamid Karimi-Rouzbahani, <email>hamid.karimi-rouzbahani@mrc-cbu.cam.ac.uk</email></corresp>
<fn fn-type="other" id="fn004"><p>This article was submitted to Brain Imaging Methods, a section of the journal Frontiers in Neuroscience</p></fn>
</author-notes>
<pub-date pub-type="epub">
<day>02</day>
<month>03</month>
<year>2022</year>
</pub-date>
<pub-date pub-type="collection">
<year>2022</year>
</pub-date>
<volume>16</volume>
<elocation-id>825746</elocation-id>
<history>
<date date-type="received">
<day>30</day>
<month>11</month>
<year>2021</year>
</date>
<date date-type="accepted">
<day>24</day>
<month>01</month>
<year>2022</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x00A9; 2022 Karimi-Rouzbahani and Woolgar.</copyright-statement>
<copyright-year>2022</copyright-year>
<copyright-holder>Karimi-Rouzbahani and Woolgar</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license>
</permissions>
<abstract>
<p>Neural codes are reflected in complex neural activation patterns. Conventional electroencephalography (EEG) decoding analyses summarize activations by averaging/down-sampling signals within the analysis window. This diminishes informative fine-grained patterns. While previous studies have proposed distinct statistical features capable of capturing variability-dependent neural codes, it has been suggested that the brain could use a combination of encoding protocols not reflected in any one mathematical feature alone. To check, we combined 30 features using state-of-the-art supervised and unsupervised feature selection procedures (<italic>n</italic> = 17). Across three datasets, we compared decoding of visual object category between these 17 sets of combined features, and between combined and individual features. Object category could be robustly decoded using the combined features from all of the 17 algorithms. However, the combination of features, which were equalized in dimension to the individual features, were outperformed across most of the time points by the multiscale feature of Wavelet coefficients. Moreover, the Wavelet coefficients also explained the behavioral performance more accurately than the combined features. These results suggest that a single but multiscale encoding protocol may capture the EEG neural codes better than any combination of protocols. Our findings put new constraints on the models of neural information encoding in EEG.</p>
</abstract>
<kwd-group>
<kwd>neural encoding</kwd>
<kwd>multivariate pattern decoding</kwd>
<kwd>EEG</kwd>
<kwd>feature extraction</kwd>
<kwd>feature selection</kwd>
</kwd-group>
<counts>
<fig-count count="4"/>
<table-count count="1"/>
<equation-count count="0"/>
<ref-count count="90"/>
<page-count count="18"/>
<word-count count="14628"/>
</counts>
</article-meta>
</front>
<body>
<sec id="S1" sec-type="intro">
<title>Introduction</title>
<p>How is information about the world encoded by the human brain? Researchers have tried to answer this question using variety of brain imaging techniques across all sensory modalities. In vision, people have used invasive (<xref ref-type="bibr" rid="B27">Hung et al., 2005</xref>; <xref ref-type="bibr" rid="B46">Liu et al., 2009</xref>; <xref ref-type="bibr" rid="B47">Majima et al., 2014</xref>; <xref ref-type="bibr" rid="B86">Watrous et al., 2015</xref>; <xref ref-type="bibr" rid="B66">Rupp et al., 2017</xref>; <xref ref-type="bibr" rid="B49">Miyakawa et al., 2018</xref>) and non-invasive (EEG and MEG; <xref ref-type="bibr" rid="B71">Simanova et al., 2010</xref>; <xref ref-type="bibr" rid="B7">Carlson et al., 2013</xref>; <xref ref-type="bibr" rid="B10">Cichy et al., 2014</xref>; <xref ref-type="bibr" rid="B34">Kaneshiro et al., 2015</xref>; <xref ref-type="bibr" rid="B11">Contini et al., 2017</xref>) brain imaging modalities to decode object category information from variety of features of the recorded neural activations. While majority of EEG and MEG decoding studies still rely on the within-trial &#x201C;mean&#x201D; of activity (average of activation level within the sliding analysis window) as the main source of information (<xref ref-type="bibr" rid="B22">Grootswagers et al., 2017</xref>; <xref ref-type="bibr" rid="B36">Karimi-Rouzbahani et al., 2017b</xref>), recent theoretical and experimental studies have shown evidence that temporal variabilities of neural activity (sample to sample changes in the level of activity) form an additional channel of information encoding (<xref ref-type="bibr" rid="B55">Orb&#x00E1;n et al., 2016</xref>). For example, these temporal variabilities have provided information about the &#x201C;complexity,&#x201D; &#x201C;uncertainty,&#x201D; and the &#x201C;variance&#x201D; of the visual stimulus, which correlated with the semantic category of the presented image (<xref ref-type="bibr" rid="B25">Hermundstad et al., 2014</xref>; <xref ref-type="bibr" rid="B55">Orb&#x00E1;n et al., 2016</xref>; <xref ref-type="bibr" rid="B17">Garrett et al., 2020</xref>). Specifically, object categories which show a wider variability in their exemplars (e.g., houses) evoke more variable neural activation than categories which have lower variability (e.g., faces; <xref ref-type="bibr" rid="B17">Garrett et al., 2020</xref>). Accordingly, it is now clear that neural variabilities carry significant amounts of information about different aspects of sensory processing and may also play a major role in determining behavior (<xref ref-type="bibr" rid="B85">Waschke et al., 2021</xref>).</p>
<p>Despite the richness of information in neural variabilities, there is no consensus yet about how to quantify informative neural variabilities. Specifically, neural variabilities have been quantified using three classes of mathematical features: variance-, frequency-, and information theory-based features, each detecting specific, but potentially overlapping aspects of the neural variabilities (<xref ref-type="bibr" rid="B85">Waschke et al., 2021</xref>). Accordingly, previous studies have decoded object category information from EEG using variance-based (<xref ref-type="bibr" rid="B88">Wong et al., 2006</xref>; <xref ref-type="bibr" rid="B48">Mazaheri and Jensen, 2008</xref>; <xref ref-type="bibr" rid="B3">Alimardani et al., 2018</xref>; <xref ref-type="bibr" rid="B33">Joshi et al., 2018</xref>), frequency-based (<xref ref-type="bibr" rid="B75">Taghizadeh-Sarabi et al., 2015</xref>; <xref ref-type="bibr" rid="B86">Watrous et al., 2015</xref>; <xref ref-type="bibr" rid="B31">Jadidi et al., 2016</xref>; <xref ref-type="bibr" rid="B83">Wang et al., 2018</xref>; <xref ref-type="bibr" rid="B80">Voloh et al., 2020</xref>) and information theory-based (<xref ref-type="bibr" rid="B61">Richman and Moorman, 2000</xref>; <xref ref-type="bibr" rid="B70">Shourie et al., 2014</xref>; <xref ref-type="bibr" rid="B76">Torabi et al., 2017</xref>; <xref ref-type="bibr" rid="B2">Ahmadi-Pajouh et al., 2018</xref>) features. However, these previous studies remained silent about the temporal dynamics of category encoding as they performed the analyses (i.e., feature extraction and decoding) on the whole-trial data to maximize the decoding accuracy. On the other hand, time-resolved decoding analyses studied the temporal dynamics of category information encoding (<xref ref-type="bibr" rid="B34">Kaneshiro et al., 2015</xref>; <xref ref-type="bibr" rid="B22">Grootswagers et al., 2017</xref>; <xref ref-type="bibr" rid="B35">Karimi-Rouzbahani, 2018</xref>). However, few time-resolved studies have extracted any features other than the instantaneous activity at each time point, or the mean of activity across a short sliding window (e.g., by down-sampling the data), to incorporate the information contained in neural variabilities (<xref ref-type="bibr" rid="B47">Majima et al., 2014</xref>; <xref ref-type="bibr" rid="B37">Karimi-Rouzbahani et al., 2017a</xref>). Therefore, previous studies either did not focus on the temporal dynamics of information processing or did not include the contents of neural variabilities in time-resolved decoding.</p>
<p>Critically, as opposed to the Brain-Computer Interface (BCI) community, where the goal of feature extraction is to maximize the decoding accuracy, in cognitive neuroscience the goal is to find better neural correlates for the behavioral effect under study (<xref ref-type="bibr" rid="B87">Williams et al., 2007</xref>; <xref ref-type="bibr" rid="B30">Jacobs et al., 2009</xref>; <xref ref-type="bibr" rid="B24">Hebart and Baker, 2018</xref>; <xref ref-type="bibr" rid="B89">Woolgar et al., 2019</xref>; <xref ref-type="bibr" rid="B38">Karimi-Rouzbahani et al., 2021a</xref>,<xref ref-type="bibr" rid="B39">b</xref>). Specifically, a given feature is arguably only informative if it predicts behavior. Therefore, behavior is a key benchmark for evaluating the information content of any features including those which quantify neural variabilities. Interestingly, almost none of the above-mentioned decoding studies focused on evaluating the predictive power of their suggested informative features about behavior. Therefore, it remains unclear if the additional information they obtained from features of neural variabilities was task-relevant or epiphenomenal to the experimental conditions.</p>
<p>To overcome these issues, we proposed a new approach using medium-sized (50 ms) sliding windows at each time step (5 ms apart). The 50 ms time window makes a compromise between concatenating the whole time window, which in theory allows any feature to be used at the expense of temporal resolution, and decoding in a time resolved fashion at each time point separately, which might lose temporal patterns of activity (<xref ref-type="bibr" rid="B39">Karimi-Rouzbahani et al., 2021b</xref>). Within each window, we quantify multiple different mathematical features of the continuous data. This allows us to be sensitive to any information carried in local temporal variability in the EEG response, while also maintaining reasonable temporal resolution in the analysis. In a recent study, we extracted a large set of such features and quantified the information contained in each using multivariate classification (<xref ref-type="bibr" rid="B39">Karimi-Rouzbahani et al., 2021b</xref>). We balanced the number of extracted values across features using Principal Component Analysis (PCA). Across three datasets, we found that that the incorporation of temporal patterns of activity in decoding, through the extraction of spatiotemporal &#x201C;Wavelet coefficients&#x201D; or even using the informative &#x201C;original magnitude data (i.e., no feature extraction),&#x201D; provided higher decoding performance than the more conventional average of activity within each window (&#x201C;mean&#x201D;). Importantly, we also observed that for our Active dataset where participants categorized objects, the decoding results obtained from the same two features (i.e., Wavelet coefficients and original magnitude data) could predict/explain the participants&#x2019; reaction time in categorization significantly better than the &#x201C;mean&#x201D; of activity in each window (Wavelet outperformed original magnitude data). We further observed that more effective decoding of the neural codes, through the extraction of more informative features, corresponded to better prediction of behavioral performance. We concluded that the incorporation of temporal variabilities in decoding can provide additional category information and improved prediction of behavior compared to the conventional &#x201C;mean&#x201D; of activity.</p>
<p>One critical open question, however, is whether we should expect the brain to encode the information via each of these features individually, or whether it may instead use combinations of these features. In other words, while each of feature may potentially capture a specific and limited aspect of the generated neural codes, the brain may recruit multiple neural encoding protocols at the same time point or in succession within the same trial. Specifically, an encoding protocol might be active only for a limited time window or for specific aspects of the visual input (<xref ref-type="bibr" rid="B18">Gawne et al., 1996</xref>; <xref ref-type="bibr" rid="B84">Wark et al., 2009</xref>). For example, it has been shown in auditory cortex that two distinct encoding protocols (millisecond-order codes and phase coding) are simultaneously informative (<xref ref-type="bibr" rid="B42">Kayser et al., 2009</xref>). Another study showed that spike <italic>rates</italic> on 5&#x2013;10 ms timescales carried complementary information to that in the <italic>phase</italic> of firing relative to low-frequency (1&#x2013;8 Hz) local field potentials (LFPs) about which epoch of naturalistic movie was being shown (<xref ref-type="bibr" rid="B50">Montemurro et al., 2008</xref>). These examples suggest that two very distinct encoding protocols (rate vs. phase coding) might be at work simultaneously to provide information about distinct aspects of the same sensory input. Therefore, it might be the case that multiple neural encoding protocols contribute to the encoding of information. Alternatively, the brain may implement one general multiscale encoding protocol [e.g., multiplexing strategy which combines same-structure neural codes at different time scales (<xref ref-type="bibr" rid="B56">Panzeri et al., 2010</xref>)], which allows different aspects of information to be represented within a more flexible encoding protocol. More specifically, the brain might implement a general platform, which allows the representation of information at different temporal and spatial scales. For example, in visual stimulus processing, one study found that stimulus contrast was represented by latency coding at a temporal precision of &#x223C;10 ms, whereas stimulus orientation and its spatial frequency were encoded at a coarser temporal precision (30 and 100 ms, respectively; <xref ref-type="bibr" rid="B78">Victor, 2000</xref>). This multiplexed encoding protocol has been suggested to provide several computational benefits to fixed encoding protocol including enhancing the coding capacity of the system (<xref ref-type="bibr" rid="B67">Schaefer et al., 2006</xref>; <xref ref-type="bibr" rid="B42">Kayser et al., 2009</xref>), reducing the ambiguity inherent to single-scale codes (<xref ref-type="bibr" rid="B67">Schaefer et al., 2006</xref>; <xref ref-type="bibr" rid="B68">Schroeder and Lakatos, 2009</xref>) and improving the robustness of neural representations to environmental noise (<xref ref-type="bibr" rid="B42">Kayser et al., 2009</xref>).</p>
<p>To see if EEG activations reflect the neural codes using several encoding protocols simultaneously, we created combinations from the large set of distinct mathematical features in our previous study (<xref ref-type="bibr" rid="B39">Karimi-Rouzbahani et al., 2021b</xref>). We asked whether their combination recovers more of the underlying neural code, leading to additional object category information and increased accuracy in predicting behavior, compared to the best performing individual feature from the previous study (i.e., Wavelet). Specifically, we used the same three datasets, extracted the same features from neural activity, selected the most informative features at each sliding time window and evaluated their information about object categories. We also evaluated how well each combined feature set explained behavioral recognition performance. Our prediction was that as targeted combinations of informative features provide more flexibility in detecting subtle differences, which might be ignored when using each individual feature, we should see both a higher decoding accuracy and predictive power for behavior compared to when using individual features. However, our results show that, the most informative individual feature (the Wavelet transform) outperformed all of the feature combinations (combined using 17 different feature selection algorithms). Similarly, Wavelet coefficients outperformed all combinations of features in predicting behavioral performance. Therefore, while the relationship between neuron-level encoding of information and EEG signals remains to be investigated in the future, these results provide evidence for a general multiscale encoding protocol (i.e., captured by Wavelet coefficients) rather than a combination of several protocols for category encoding in the EEG data.</p>
</sec>
<sec id="S2" sec-type="materials|methods">
<title>Materials and Methods</title>
<p>As this study uses the same set of datasets and features used in our previous study, we only briefly explain the datasets and the features. The readers are referred to our previous manuscript (<xref ref-type="bibr" rid="B39">Karimi-Rouzbahani et al., 2021b</xref>) as well as the original manuscripts (cited below) for more detailed explanation of the datasets and features. The datasets used in this study and the code are available online at <ext-link ext-link-type="uri" xlink:href="https://osf.io/wbvpn/">https://osf.io/wbvpn/</ext-link>. The EEG and behavioral data are available in Matlab &#x201C;.mat&#x201D; format and the code in Matlab &#x201C;.m&#x201D; format.</p>
<p>All the open-source scripts used in this study for feature extraction were compared/validated against other implementations of identical algorithms in simulations and used only if they produced identical results. All open-source scripts of similar algorithms produced identical results in our validations. To validate the scripts, we used 1,000 random (normally distributed with unit variance and zero mean) time series each including 1,000 samples.</p>
<sec id="S2.SS1">
<title>Overview of Datasets</title>
<p>We selected three highly varied previously published EEG datasets (<xref ref-type="table" rid="T1">Table 1</xref>) for this study to be able to evaluate the generalizability of our results and conclusions. Specifically, the datasets differed in a wide range of aspects including the recording set-up (e.g., amplifier, number of electrodes, preprocessing steps, etc.), properties of the image-set (e.g., number of categories and exemplars within each category, colorfulness of images, etc.), paradigm and task (e.g., presentation length, order and the participants&#x2019; task). The EEG datasets were collected while the participants were presented with images of objects, animals, face, etc. Participants&#x2019; task in Dataset 1 was irrelevant to the identity of the presented objects; they reported if the color of fixation changed from the first stimulus to the second in pairs of stimuli. Participants&#x2019; task for Dataset 2 was to respond/withhold response to indicate if the presented object belonged to the category (e.g., animal) cued at the beginning of the block. Participants had no explicit active task except for keeping fixation on the center of the screen for Dataset 3. To obtain relatively high signal to noise ratios for the analyses, each unique stimulus was presented to the participants 3, 6, and 12 times in datasets 1&#x2013;3, respectively. The three datasets previously successfully provided object category information using multivariate decoding methods. For more details about the datasets see the original manuscripts cited in <xref ref-type="table" rid="T1">Table 1</xref>.</p>
<table-wrap position="float" id="T1">
<label>TABLE 1</label>
<caption><p>Details of the three datasets used in the study.</p></caption>
<table cellspacing="5" cellpadding="5" frame="hsides" rules="groups">
<thead>
<tr>
<td valign="top" align="left" colspan="2">Dataset</td>
<td valign="top" align="center"># and type of electrodes</td>
<td valign="top" align="center">Band-pass filtering</td>
<td valign="top" align="center">Notch filtering</td>
<td valign="top" align="center">#<break/> object categories</td>
<td valign="top" align="center"># stimulus repetition</td>
<td valign="top" align="center">Stimulus presentation time</td>
<td valign="top" align="center">Stimulus size (periphery)</td>
<td valign="top" align="center">Task</td>
<td valign="top" align="center">Participants&#x2019; accuracy</td>
<td valign="top" align="center">Participants&#x2019; Age (median)</td>
<td valign="top" align="center">Participants&#x2019; gender</td>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">1</td>
<td valign="top" align="center"><xref ref-type="bibr" rid="B37">Karimi-Rouzbahani et al., 2017a</xref></td>
<td valign="top" align="center">31 (Passive- 10&#x2013;20 system)</td>
<td valign="top" align="center">0.03&#x2013;200 Hz</td>
<td valign="top" align="center">50 Hz</td>
<td valign="top" align="center">4</td>
<td valign="top" align="center">3</td>
<td valign="top" align="center">50 ms</td>
<td valign="top" align="center">2&#x00B0;&#x2013;13.5&#x00B0; (0.7&#x00B0;&#x2013;8.8&#x00B0;)</td>
<td valign="top" align="center">Color matching (passive)</td>
<td valign="top" align="center">%94.68</td>
<td valign="top" align="center">22.1</td>
<td valign="top" align="center">Seven male<break/> Three female</td>
</tr>
<tr>
<td valign="top" align="left">2</td>
<td valign="top" align="center"><xref ref-type="bibr" rid="B41">Karimi-Rouzbahani et al., 2019</xref></td>
<td valign="top" align="center">31 (Passive- 10&#x2013;20 system)</td>
<td valign="top" align="center">0.03&#x2013;200 Hz</td>
<td valign="top" align="center">50 Hz</td>
<td valign="top" align="center">4</td>
<td valign="top" align="center">6</td>
<td valign="top" align="center">900 ms</td>
<td valign="top" align="center">8&#x00B0; &#x00D7; 8&#x00B0; (0)</td>
<td valign="top" align="center">Object category detection (active)</td>
<td valign="top" align="center">%94.65</td>
<td valign="top" align="center">26.4</td>
<td valign="top" align="center">Six male<break/> Four female</td>
</tr>
<tr>
<td valign="top" align="left">3</td>
<td valign="top" align="center"><xref ref-type="bibr" rid="B34">Kaneshiro et al., 2015</xref></td>
<td valign="top" align="center">128 (Passive high-density HCGSN 128)</td>
<td valign="top" align="center">0.03&#x2013;50 Hz</td>
<td valign="top" align="center">No</td>
<td valign="top" align="center">6</td>
<td valign="top" align="center">12</td>
<td valign="top" align="center">500 ms</td>
<td valign="top" align="center">7.0&#x00B0; &#x00D7; 6.5&#x00B0; (0)</td>
<td valign="top" align="center">No task (fixation)</td>
<td valign="top" align="center">N/A</td>
<td valign="top" align="center">30.5</td>
<td valign="top" align="center">Seven male<break/> Three female</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="S2.SS2">
<title>Preprocessing</title>
<p>The datasets were collected at a sampling rate of 1,000 Hz. Each dataset consisted of data from 10 participants. Each object category in each dataset included 12 exemplars. For datasets 1 and 2, only the trials with correct responses were used in the analyses (dataset 3 had no task). To make the three datasets as consistent as possible, we pre-processed them differently from their original manuscripts. We performed notch-filtering on datasets 1 and 2 at 50 Hz. Datasets 1 and 2 were also band-pass-filtered in the range from 0.03 to 200 Hz. The band-pass filtering range of dataset 3 was 0.03&#x2013;50 Hz, as we did not have access to the raw data to increase the upper bound to 200 Hz. Despite potential muscle artifacts in higher frequency bands of EEG (e.g., &#x003E;30 Hz; <xref ref-type="bibr" rid="B12">da Silva, 2013</xref>; <xref ref-type="bibr" rid="B52">Muthukumaraswamy, 2013</xref>), the upper limit of the frequency band was selected liberally (200 Hz) to avoid missing any potential information which might be contained in high-frequency components (gamma band) of the signals (<xref ref-type="bibr" rid="B86">Watrous et al., 2015</xref>). As sporadic artefacts (including muscle activity, eye and movement artifacts) do not generally consistently differ across conditions (here categories), they will only minimally affect multivariate decoding analyses (<xref ref-type="bibr" rid="B22">Grootswagers et al., 2017</xref>; <xref ref-type="bibr" rid="B40">Karimi-Rouzbahani et al., 2021c</xref>). For the same reason, we did not remove the artifacts. We used finite-impulse-response filters with 12 dB roll-off per octave for band-pass filtering of datasets 1 and 2. The filtering was applied on the data before they were epoched relative to the trial onset times. Data were epoched from 200 ms before to 1,000 ms after the stimulus onset to cover most of the range of event-related neural activations. The average pre-stimulus (&#x2212;200 to 0 ms relative to the stimulus onset) signal amplitude was removed from each trial of the data. For more information about each dataset see the references cited in <xref ref-type="table" rid="T1">Table 1</xref>.</p>
</sec>
<sec id="S2.SS3">
<title>Features</title>
<p>We briefly explain the 26 mathematically distinct features used in this study below. Note that 4 of the features, which were event-related potentials, were excluded from this study as they could not be defined across time. For more details about their algorithms, their plausibility and possible neural underpinnings please see <xref ref-type="bibr" rid="B39">Karimi-Rouzbahani et al. (2021b)</xref>. Each feature was calculated for each EEG electrode and each participant separately. The following features were extracted after the raw data was filtered, epoched and baselined as explained. Each of the features was extracted from the 50 samples contained in 50 ms sliding time windows at a step size of 5 ms along each trial. The sampling rate of the data remained at 1,000 Hz and the features were extracted from the 1,000-Hz data but only calculated every 5 ms to decrease the computational load. Note that the width of the sliding analysis window needs special attention as it involves a trade-off between noise and potential information (about conditions and behavior) in EEG signals. Specifically, very short windows may lose potentially informative longer patterns, whereas very long windows might lose shorter patterns as they might be dominated by slow fluctuations. In the original work (<xref ref-type="bibr" rid="B39">Karimi-Rouzbahani et al., 2021b</xref>) we tested window widths between 5 and 100 ms and found 50 ms to be the most informative range for decoding, so that is the value we use here.</p>
<sec id="S2.SS3.SSS1">
<title>Mean, Variance, Skewness, and Kurtosis</title>
<p>These are the standard 1st to 4th moments of EEG time series. To calculate these features, we simply calculated the mean, variance, skewness and variance of EEG signals <italic>over the samples within each sliding analysis window within each trial</italic> (50 samples). Please note that this differs from averaging over trials, which is sometimes used to increase signal to noise ratio (<xref ref-type="bibr" rid="B24">Hebart and Baker, 2018</xref>). &#x201C;Mean&#x201D; of activity is by far the most common feature of EEG signal used in time-resolved decoding (<xref ref-type="bibr" rid="B22">Grootswagers et al., 2017</xref>). Specifically, in time-resolved decoding, generally the samples within each sliding time window are averaged and used as the input for the classification algorithm. People sometimes perform down-sampling of EEG time series, which either performs simple averaging or retains the selected samples every few samples. Variance (<xref ref-type="bibr" rid="B88">Wong et al., 2006</xref>), Skewness (<xref ref-type="bibr" rid="B48">Mazaheri and Jensen, 2008</xref>), and Kurtosis (<xref ref-type="bibr" rid="B57">Pouryazdian and Erfanian, 2009</xref>; <xref ref-type="bibr" rid="B3">Alimardani et al., 2018</xref>) have shown success in providing information about different conditions of visually evoked potentials.</p>
</sec>
<sec id="S2.SS3.SSS2">
<title>Median</title>
<p>We also calculated signal&#x2019;s median as it is less affected by spurious values compared to the signal mean providing less noisy representations of the neural processes.</p>
<p>While the moment features above provide valuable information about the content of EEG evoked potentials, many distinct time series could lead to similar moment features. In order to be sensitive to this potentially informative differences nonlinear features can be used which, roughly speaking, are sensitive to nonlinear and complex patterns in time series. Below we define the most common nonlinear features of EEG time series analysis, which we used in this study.</p>
</sec>
<sec id="S2.SS3.SSS3">
<title>Lempel-Ziv Complexity</title>
<p>We calculated the Lempel-Ziv (LZ) complexity as an index of signal complexity. This measure counts the number of unique sub-sequences within the analysis window (50 time samples), after turning the time samples into a binary sequence. To generate the binary sequence, we used the signal median, within the same analysis window, as the threshold. Accordingly, the LZ complexity of a time series grows with the length of the signal and its irregularity over time. See <xref ref-type="bibr" rid="B45">Lempel and Ziv (1976)</xref> for more details. This measure has previously provided information about neural responses in primary visual cortices (<xref ref-type="bibr" rid="B74">Szczepa&#x0144;ski et al., 2003</xref>). We used the script by Quang Thai<sup><xref ref-type="fn" rid="footnote1">1</xref></sup> implemented based on &#x201C;exhaustive complexity&#x201D; which is considered to provide the lower limit of the complexity as explained by <xref ref-type="bibr" rid="B45">Lempel and Ziv (1976)</xref>.</p>
</sec>
<sec id="S2.SS3.SSS4">
<title>Higuchi and Katz Fractal Dimensions</title>
<p>Fractal is an indexing technique which provides statistical information determining the complexity of how data are organized within time series. Accordingly, higher fractal values, suggest more complexity and vice versa. In this study, we calculated the complexity of the signals using two methods of Higuchi and Katz, as used previously for categorizing object categories (<xref ref-type="bibr" rid="B76">Torabi et al., 2017</xref>; <xref ref-type="bibr" rid="B2">Ahmadi-Pajouh et al., 2018</xref>; <xref ref-type="bibr" rid="B53">Namazi et al., 2018</xref>). We used the implementations by Jes&#x00FA;s Monge &#x00C1;lvarez<sup><xref ref-type="fn" rid="footnote2">2</xref></sup> after verifying it against other implementations.</p>
</sec>
<sec id="S2.SS3.SSS5">
<title>Hurst Exponent</title>
<p>This measure quantifies the long-term &#x201C;memory&#x201D; in a time series. Basically, it calculates the degree of dependence among consecutive samples of time series and functions similarly to the autocorrelation function (<xref ref-type="bibr" rid="B60">Racine, 2011</xref>; <xref ref-type="bibr" rid="B76">Torabi et al., 2017</xref>). Hurst values between 0.5 and 1 suggest consecutive appearance of high signal values on large time scales while values between 0 and 0.5 suggest frequent switching between high and low signal values. Values around 0.5 suggest no specific patterns among samples of a time series.</p>
</sec>
<sec id="S2.SS3.SSS6">
<title>Sample and Approximate Entropy</title>
<p>Entropy measures the level of perturbation in time series. As the precise calculation of entropy needs large sample sizes and is also noise-sensitive, we calculated it using two of the most common approaches: sample entropy and approximate entropy. Sample entropy is not as sensitive to the sample size and simpler to implement compared to approximate entropy. Sample entropy, however, does not take into account self-similar patterns in the time series (<xref ref-type="bibr" rid="B61">Richman and Moorman, 2000</xref>). We used an open-source code<sup><xref ref-type="fn" rid="footnote3">3</xref></sup> for calculating approximate entropy.</p>
</sec>
<sec id="S2.SS3.SSS7">
<title>Autocorrelation</title>
<p>This index quantifies the self-similarity of a time series at specific time lags. Accordingly, if a time series has a repeating pattern at the rate of F hertz, an autocorrelation measure with a lag of 1/F will provide a value of 1. However, it would return &#x2212;1 at the lag of 1/2F. It would provide values between &#x2212;1 and 1 for other lags. More complex signals would provide values close to 0. A previous study has been able to decode neural information about motor imagery using the autocorrelation function from EEG signals (<xref ref-type="bibr" rid="B81">Wairagkar et al., 2016</xref>).</p>
</sec>
<sec id="S2.SS3.SSS8">
<title>Hjorth Complexity and Mobility</title>
<p>These parameters measure the variation in the signals&#x2019; characteristics. The complexity measure calculates the variation in a signal&#x2019;s dominant frequency, and the mobility measures the width of the signal&#x2019;s power spectrum [how widely the frequencies are scattered in the power spectrum of the signal (<xref ref-type="bibr" rid="B33">Joshi et al., 2018</xref>)].</p>
</sec>
<sec id="S2.SS3.SSS9">
<title>Mean, Median, and Average Frequency</title>
<p>These measures calculate the central frequency of the signal in different ways. Mean frequency is the average of all frequency components available in a signal. Median frequency is the median normalized frequency of the power spectrum of the signal and the average frequency is the number of times the signal time series crosses zero. They have shown information about visual categories in previous studies (<xref ref-type="bibr" rid="B31">Jadidi et al., 2016</xref>; <xref ref-type="bibr" rid="B28">Iranmanesh and Rodriguez-Villegas, 2017</xref>; <xref ref-type="bibr" rid="B33">Joshi et al., 2018</xref>).</p>
</sec>
<sec id="S2.SS3.SSS10">
<title>Spectral Edge Frequency (95%)</title>
<p>Spectral edge frequency (SEF) indicates the high frequency below which x percent of the signal&#x2019;s power spectrum exists. X was set to 95% in this study. Therefore, SEF reflects the upper-bound of frequency in the power spectrum.</p>
</sec>
<sec id="S2.SS3.SSS11">
<title>Signal Power, Power, and Phase at Median Frequency</title>
<p>Power spectrum density (PSD) represents the intensity or the distribution of the signal power into its constituent frequency components. Signal power was used as a feature here as in previous studies (<xref ref-type="bibr" rid="B47">Majima et al., 2014</xref>; <xref ref-type="bibr" rid="B66">Rupp et al., 2017</xref>), where it showed associations between aspects of visual perception and power in certain frequency bands. Signal power is the frequency-domain representation of temporal neural variability (<xref ref-type="bibr" rid="B85">Waschke et al., 2021</xref>). We also extracted signal power and phase at median frequency which have previously shown to be informative about object categories (<xref ref-type="bibr" rid="B31">Jadidi et al., 2016</xref>; <xref ref-type="bibr" rid="B66">Rupp et al., 2017</xref>).</p>
<p>For the following features we had more than one value per trial and sliding time window. We extracted all these features but later down-sampled the values to <italic>one</italic> per trial using the (first) PCA procedure explained below (<xref ref-type="fig" rid="F1">Figure 1</xref>) before using them in the feature combination procedure.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption><p>Decoding pipeline. From left to right: successive stages shown for a sample dataset comprising 100 trials of data from two categories recorded using a 31-electrode EEG amplifier. (1) Features are extracted from each trial and time window of the data. The features can be single- or multi-valued resulting in different number of values per trial and analysis time window. (2) We split the trials into training and testing sets and use the training sets in PCA and training the classifiers throughout the pipeline. (3) We used a PCA-based dimension reduction to reduce the number of values of only the multi-valued features to one equalizing them with single-valued features. (4) We used a second PCA to project all values of each feature to one dimension to be able to feed to the feature selection (FS) algorithms. (5) We selected the five most informative features using the FS algorithms. (6) We combined these features using concatenation of the selected features in their original size received from stage 4. (7) We reduced the dimension of the concatenated feature set to equalize it with the single-valued individual features from the previous study so that they could be compared. (8) We decoded/classified all pair-wise categories using the final dataset in each fold. This figure shows the procedure for a single cross-validation fold at one time point and was repeated for all the folds and time points. To avoid circularity, PCA was only ever applied on the training set and the parameters (mean and eigen vectors) used to derive the principal component of both the training and testing sets. The green arrows indicate example selected feature sets sent for combination.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-16-825746-g001.tif"/>
</fig>
</sec>
<sec id="S2.SS3.SSS12">
<title>Cross-Correlation</title>
<p>This refers to the inter-electrode correlation of EEG time series. It simply quantifies the similarity of activations between pairs of EEG electrodes. Therefore, for each electrode, we had e-1 cross-correlation values with e referring to the number of electrodes. This measure has been shown to contain information about visual object categories before (<xref ref-type="bibr" rid="B47">Majima et al., 2014</xref>; <xref ref-type="bibr" rid="B37">Karimi-Rouzbahani et al., 2017a</xref>).</p>
</sec>
<sec id="S2.SS3.SSS13">
<title>Wavelet Coefficients</title>
<p>Considering the time- and frequency-dependent nature of ERPs, Wavelet transform seems to be a very reasonable choice as it provides a time-frequency representation of signal components. It determines the primary frequency components and their temporal position in time series. The transformation passes the signal time series through digital filters (<xref ref-type="bibr" rid="B23">Guo et al., 2009</xref>), each of which adjusted to extract a specific frequency (scale) at a specific time. This filtering procedure is repeated for several rounds (levels) filtering low- (approximations) and high-frequency (details) components of the signal to provide more fine-grained information about the constituent components of the signal. This can lead to coefficients which can potentially discriminate signals evoked by different conditions. Following up on a previous study (<xref ref-type="bibr" rid="B75">Taghizadeh-Sarabi et al., 2015</xref>), and to make the number of Wavelet features comparable in number to signal samples, we used detail coefficients at five levels D1,&#x2026;,D5 as well as the approximate coefficients at level 5, A5. This led to 57 features in the 50 ms sliding time windows. We used the &#x201C;Symlet2&#x201D; basis function for our Wavelet transformations as implemented in Matlab. The multistage, variable-sized filtering procedure implemented in Wavelet coefficients, make them ideal for detecting multiscale patterns of neural activity, which has been suggested to be produced by the brain for information encoding (<xref ref-type="bibr" rid="B56">Panzeri et al., 2010</xref>).</p>
</sec>
<sec id="S2.SS3.SSS14">
<title>Hilbert Amplitude and Phase</title>
<p>This transformation is a mapping function that takes a function x(t) of a real variable, and using convolution with the function, 1/&#x03C0;t, produces another function of a real variable H(u) (t). This technique provides amplitude and phase information of the signal in the transformed space allowing us to tease them apart and evaluate their information content about visual categories (<xref ref-type="bibr" rid="B83">Wang et al., 2018</xref>).</p>
</sec>
<sec id="S2.SS3.SSS15">
<title>Original Magnitude Data (Samples)</title>
<p>We also used the post-stimulus signal samples (i.e., 50 samples in each sliding analysis window) to decode object category information without any feature extraction. This allowed us to compare the information content of the extracted features with the original signal samples to see if the former provided any extra information. Note that, this is different from averaging/down-sampling of magnitude data within the analysis windows conventionally used in multivariate decoding (<xref ref-type="bibr" rid="B37">Karimi-Rouzbahani et al., 2017a</xref>).</p>
</sec>
</sec>
<sec id="S2.SS4">
<title>Feature Selection Algorithms</title>
<p>We set out to test whether neural information about object categories might be captured by combinations of the above features, better than by any one feature individually. For this, we combined the 26 extracted features using Feature Selection Library (FSLib, version 6.2.1; <xref ref-type="bibr" rid="B63">Roffo, 2016</xref>). Feature selection (FS), which refers to selecting a subset of features from a larger set, is generally used (for example, in machine learning) to reduce the dimensionality of the data by removing the less informative features from the dataset. FS algorithms can be categorized as supervised or unsupervised (<xref ref-type="bibr" rid="B13">Dash and Liu, 1997</xref>). The supervised methods receive, as input, the labels of trials for each condition (i.e., object categories here), and try to maximize the distance between conditions. We used eight different supervised FS algorithms. The unsupervised methods, on the other hand, incorporate different criteria for FS such as selecting features that provide maximum distance (i.e., <italic>unfol</italic>) or minimum correlation (i.e., <italic>cfs</italic>). The FSLib implements 19 different feature selection algorithms. As it is not yet known how the brain might recruit different encoding protocols or a potential combination of them, we used all the FS algorithms available by the FSLib to combine the features in this study, except two (rfe-SVM and L0) which we were not able to implement. Although there are other feature selection algorithms in the literature, we believe that using these 17 methods, we capture a decent range of different approaches. We set the number of selected features to 5, which was chosen to balance between including too many features, which could obscure interpretability, and including too few, which risks missing informative but lower-ranked features. Below we briefly explain the eight supervised and nine unsupervised feature selection algorithms. Readers are referred to the original manuscripts for more detail about each feature selection method as reviewed (<xref ref-type="bibr" rid="B63">Roffo, 2016</xref>).</p>
<p>Among supervised algorithms, <italic>Relief</italic> is a randomized and iterative algorithm that evaluates the quality of the features based on how well their values discriminate data samples from opposing conditions. This algorithm can be sensitive when used on small data samples. <italic>Fisher</italic> evaluates the information of features as the ratio of inter-class to intra-class distances. <italic>Mutual Information</italic> (mutinffs) measures the association between the data samples (observations) within each feature and their class labels. <italic>Max-Relevance, Min-Redundancy</italic> (mrmr) method, which is an extension of the mutual information method, is designed to follow two basic rules when selecting the features: to select the features which are mutually far away from each other while still having &#x201C;high&#x201D; correlation to the classification labels. As opposed to the above methods, which rank and select the features according to their specific criteria, the <italic>Infinite latent</italic> (ILFS) method, selects the most informative features based on the importance of their neighboring features in a graph-based algorithm. It is a supervised probabilistic approach that models the features &#x201C;relevancy&#x201D; in a generative process and derives the graph of features which allows the evaluation of each feature based on its neighbors. Similarly, the method of <italic>Eigenvector Centrality</italic> (ECFS), generates a graph of features with features as nodes and evaluates the importance of each node through an indicator of centrality, i.e., eigen vector centrality. The ranking of central nodes determines the most informative features. <italic>LASSO</italic> algorithm works based on error minimization in predicting the class labels using the features as regression variables. The algorithm penalizes the coefficients of the regression variables while setting the less relevant to zero to follow the minimal sum constraint. The selected features are those which have non-zero coefficients in this process. <italic>Concave Minimization</italic> (fsv) uses a linear programming technique to inject the feature selection process into the training of a support vector machine (SVM).</p>
<p>Among unsupervised FS algorithms, <italic>Infinite FS</italic> (InfFS), is similar to the graph-based supervised methods in which each feature is a node in a graph. Here, however, a path on a graph is a subset of features and the importance of each feature is measured by evaluating all possible paths on the graph as feature subsets in a cross-validation procedure. <italic>Laplacian Score</italic> (laplacian), evaluates the information content of each feature by its ability of locality preserving. To model the local geometry of the features space, this method generates a graph based on nearest neighbor and selects the features which respect this graph structure. <italic>Dependence Guided</italic> (dgufs) method evaluates the relationship between the original data, cluster labels and selected features. This algorithm tries to achieve two goals: to increase the dependence on the original data, and to maximize the dependence of the selected features on cluster labels. <italic>Adaptive Structure Learning</italic> (fsasl), which learns the structure of the data and FS at the same time is based on linear regression. <italic>Ordinal Locality</italic> (ufsol) is a clustering-based method which achieves distance-based clustering by preserving the relative neighborhood proximities. <italic>Multi-Cluster</italic> (mcfs) method is based on manifold learning and L1-regularized models for subset selection. This method selects the features such that the multi-cluster structure of the data can be best preserved. As opposed to most of the unsupervised methods which try to select the features which preserve the structure of the data, e.g., manifold learning, <italic>L2,1-norm Regularized</italic> (UDFS) method assumes that the class label of data can be predicted using a linear classifier and incorporates discriminative analysis and L2,1-norm minimization into a joint framework for feature selection. <italic>Local Learning-Based</italic> (llcfs) method is designed to work with high-dimensional manifold data. This method associates weights to features which are incorporated into the regularization procedure to evaluate their relevance for the clustering. The weights are optimized iteratively during clustering which leads to the selection of the most informative features in an unsupervised fashion. <italic>Correlation-Based</italic> (cfs) method simply ranks the features based on how uncorrelated they are to the other features in the feature set. Therefore, the selected features are those which are most distinct from others.</p>
</sec>
<sec id="S2.SS5">
<title>Decoding Pipeline</title>
<p>The pipeline used in this study for feature extraction, dimensionality reduction, feature selection, feature combination and decoding had eight stages and is summarized in <xref ref-type="fig" rid="F1">Figure 1</xref>. Below we explain each stage of the pipeline for a simple sample dataset with 100 trials collected using a 31-electrode EEG setup. Our actual datasets, however, had varied number of trials and electrodes as explained in <xref ref-type="table" rid="T1">Table 1</xref>. Note that the data from all electrodes were included in the analysis and could have affected the final decoding results equally.</p>
<sec id="S2.SS5.SSS1">
<title>Feature Extraction</title>
<p>We extracted the set of 26 above-mentioned features from the dataset. This included features which provided one value for each sliding time window per trial (single-valued) and more than one value (multi-valued). For the sample dataset, this resulted in data matrices with 100 rows (trials) and 31 columns (electrodes) for the single-valued datasets and 31 &#x00D7; <italic>e</italic> columns for multi-valued features, where <italic>e</italic> refers to the number of values extracted for each trial and time window.</p>
</sec>
<sec id="S2.SS5.SSS2">
<title>Cross Validation</title>
<p>After extracting the features, we split the data into 10 folds, used 9 folds for dimension reductions and training the classifiers and the left-out fold for testing the classifiers. Therefore, we used a 10-fold cross-validation procedure in which we trained the classifier on 90% of the data and tested it on the left-out 10% of the data, repeating the procedure 10 times until all trials from the pair of categories participate once in the training and once in the testing of the classifiers. The same trials were chosen for all features in each cross-validation fold.</p>
</sec>
<sec id="S2.SS5.SSS3">
<title>Dimensionality Reduction 1: Only for Multi-Valued Features</title>
<p>The multi-valued features explained above resulted in more than a single feature value per trial per sliding time window (e.g., cross-correlation, wavelet, Hilbert amplitude, and phase and signal samples). This could lead to the domination of the multi-valued over single-valued features in feature selection and combination. To avoid that, we used principle component analysis (PCA) to reduce the number of values in the multi-valued features to one per electrode per time window, which was the number of values for all single-valued features. Specifically, the data matrix before dimension reduction, had a dimension of <italic>n</italic> rows by <italic>e</italic> &#x00D7; <italic>f</italic> columns where <italic>n</italic>, <italic>e</italic>, and <italic>f</italic> were the number of trials in the dataset (consisting of all trials from all categories), the number of electrodes and the number of values obtained from a given feature (concatenated in columns), respectively. <italic>Therefore, the columns of multi-valued features included both the spatial (electrodes) and temporal (elements of each feature) patterns of activity from which the information was obtained</italic>. This is different from single-valued features where the columns of their data matrix only included spatial patterns of activity. As <italic>f</italic> = 1 for the single-valued features, for the multi-valued features, we only retained the <italic>e</italic> most informative columns that corresponded to the <italic>e</italic> eigen values with highest variance and removed the other columns using PCA. Therefore, we reduced the dimension of the data matrix to <italic>n</italic> &#x00D7; <italic>e</italic> which was the same for single- and multi-valued features and used the resulting data matrix for decoding. This means that, for the multi-valued features, in every analysis window, we only retained the most informative value of the extracted feature elements and electrodes (i.e., the one with the most variance in PCA). Accordingly, multi-valued features had the advantage over single-valued features as the former utilized both the <italic>spatial</italic> and <italic>temporal</italic> patterns of activity in each sliding time window, while the latter only had access to the <italic>spatial</italic> patterns.</p>
</sec>
<sec id="S2.SS5.SSS4">
<title>Dimensionality Reduction 2: For Feature Selection</title>
<p>For feature selection, each feature should have a dimension of 1 to go into the FS algorithm. However, our features had as many dimensions as the number of electrodes (i.e., <italic>e</italic>). Therefore, we further reduced the dimension of each feature from <italic>e</italic> to 1 to be able to feed them to the FS algorithms, compare them and select the most informative features. This allowed us to know the general amount of information that each feature rather than each of its elements/dimensions (e.g., electrodes in single-valued features) had about object categories. Please note that, however, after finding the most informative features, we used the selected features in their original size which was <italic>e</italic> (output of step 3 goes to stage 6).</p>
</sec>
<sec id="S2.SS5.SSS5">
<title>Feature Selection</title>
<p>Feature selection was done using 17 distinct algorithms (above) to find the five most informative features in every sliding time window. This stage only provided indices of the selected features for combination in the next stage. To avoid any circularity (<xref ref-type="bibr" rid="B58">Pulini et al., 2019</xref>), we applied the FS algorithms only on the training data (folds) and used the selected features in both training and testing in each cross-validation run. Please note that feature selection was performed in every analysis window across the trial. In other words, different sets of five features could be selected for each individual analysis window. This allowed multiple features to contribute at each time point (multiple codes to be in use at the same time) and for different features to be selected at different time points (different codes used at different points in the trial).</p>
</sec>
<sec id="S2.SS5.SSS6">
<title>Feature Combination</title>
<p>We only concatenated the five selected features into a new data matrix. At this stage, we received five feature data matrices which had a dimension of <italic>n</italic> &#x00D7; <italic>e</italic> with <italic>n</italic> referring to the number of trials and <italic>e</italic> referring to the number of values per trial, which were 100 &#x00D7; 31 for the sample dataset explained in <xref ref-type="fig" rid="F1">Figure 1</xref>. The combination procedure led to a concatenated data matrix of&#x2005;100 &#x00D7; 155 (<italic>n</italic> &#x00D7; 5<italic>e</italic>).</p>
</sec>
<sec id="S2.SS5.SSS7">
<title>Dimensionality Reduction 3: Equalizing the Dimensions of Combined and Individual Feature Spaces</title>
<p>We used another round of PCA to simultaneously combine and reduce the dimensionality of each data matrix (feature space) to equalize it with the feature space of the individual features. This made the combined and individual features directly comparable, so that we could test whether a combination of the most informative features could provide additional category-related information, over and above the information decodable from individual features. Had we not controlled for the dimension of the data matrix, superior decoding for the combined features could arise trivially (due to having more predictors). Note that, whereas we knew the features which were selected on stage 5, as a result of this PCA transformation, we did not know which features contributed to the final decoding result. Therefore, in the worst case scenario, the final feature set might have only contained one of the five selected features. However, this seems unlikely to be the case as generally all inputs contribute to the distributions of the data in the PCA space. To avoid circularity (<xref ref-type="bibr" rid="B58">Pulini et al., 2019</xref>), we again applied the PCA algorithms on the training data (folds) only and used the training PCA parameters (i.e., eigen values and means) for both training and testing (fold) sets for dimension reduction, carrying this out in each cross-validation run separately.</p>
</sec>
<sec id="S2.SS5.SSS8">
<title>Multivariate Decoding</title>
<p>Finally we used time-resolved multivariate decoding to test for information about object categories in the features and combinations of features. We used linear discriminant analysis (LDA) classifiers to measure the information content across all possible pairs of conditions (i.e., object categories) in each dataset. We repeated the decoding across all possible pairs of categories within each dataset, which were 6, 6 and 15 pairs for datasets 1&#x2013;3, which consisted of 4, 4 and 6 object categories, respectively. Finally, we averaged the results across all combinations and reported them as the average decoding for each participant. We extracted the features from 50 ms sliding time windows in steps of 5 ms across the time course of the trial (&#x2212;200 to 1,000 ms relative to the stimulus onset time). Therefore, the decoding results at each time point reflect the data for the 50 ms window around the time point, from &#x2212;25 to +24 ms relative to the time point.</p>
</sec>
</sec>
<sec id="S2.SS6">
<title>Decoding-Behavior Correlation</title>
<p>We evaluated the correlation between neural representations of object categories and the reaction time of participants in discriminating them. To that end, we generated a 10-dimensional vector of neural decoding accuracies (averaged over all pairwise category decoding accuracies obtained from each participant) at every time point and a 10-dimensional vector which contained the behavioral reaction times (averaged over all categories obtained from each participant) for the same group of 10 participants. Then we correlated the two vectors at each time point using Spearman&#x2019;s rank-order correlation (<xref ref-type="bibr" rid="B10">Cichy et al., 2014</xref>; <xref ref-type="bibr" rid="B62">Ritchie et al., 2015</xref>). This resulted in a single correlation value for each time point for the group of 10 participants.</p>
</sec>
<sec id="S2.SS7">
<title>Parameters of Decoding Curves</title>
<p>To quantitatively evaluate the patterns of decoding curves and decoding-behavior correlations, we extracted four distinct parameters from the decoding curves and one parameter from the correlation to behavior curves. All parameters were calculated in the post-stimulus time span. The &#x201C;average correlation to behavior&#x201D; was calculated by averaging the level of across-subject correlation to behavior. The parameters of &#x201C;average decoding&#x201D; and &#x201C;maximum decoding&#x201D; were calculated for each participant simply by calculating the average and maximum of the decoding curves. The &#x201C;time of maximum decoding&#x201D; and &#x201C;time of first above-chance decoding&#x201D; were also calculated for each participant relative to the time of the stimulus onset.</p>
</sec>
<sec id="S2.SS8">
<title>Statistical Analyses</title>
<sec id="S2.SS8.SSS1">
<title>Bayes Factor Analysis</title>
<p>First we asked whether we could decode object category from the combined features returned by each of the 17 FS methods. To determine the evidence for the null and the alternative hypotheses, we used Bayes analyses as implemented by Bart Krekelberg<sup><xref ref-type="fn" rid="footnote4">4</xref></sup> based on <xref ref-type="bibr" rid="B65">Rouder et al. (2012)</xref>. We used standard rules of thumb for interpreting levels of evidence (<xref ref-type="bibr" rid="B44">Lee and Wagenmakers, 2005</xref>; <xref ref-type="bibr" rid="B15">Dienes, 2014</xref>): Bayes factors of &#x003E;10 and &#x003C;1/10 were interpreted as strong evidence for the alternative and null hypotheses, respectively, and &#x003E;3 and &#x003C;1/3 were interpreted as moderate evidence for the alternative and null hypotheses, respectively. We considered the Bayes factors which fell between 3 and 1/3 as suggesting insufficient evidence either way.</p>
<p>To evaluate the evidence for the null and alternative hypotheses of at-chance and above-chance decoding, respectively, we compared the decoding accuracies obtained from all participants in the post-stimulus onset time against the decoding accuracies obtained from the same participants averaged in the pre-stimulus onset time (&#x2212;200 to 0 ms). We also asked whether there was a difference between the decoding values obtained from all possible pairs of FS methods. Accordingly, we performed the Bayes factor unpaired <italic>t-test</italic> and calculated the Bayes factor as the probability of the data under alternative (i.e., difference; H1) relative to the null (i.e., no difference; H0) hypothesis between all possible pairs of FS methods for each dataset separately. The same procedure was used to evaluate evidence for difference (i.e., alternative hypothesis) or no difference (i.e., null hypothesis) in the maximum and average decoding accuracies, the time of maximum and above-chance decoding accuracies across FS methods for each dataset separately. To evaluate the evidence for the null or alternative hypotheses of lack of or the existence of difference between the decoding accuracies obtained from FS algorithm and the Wavelet feature, we calculated the Bayes factor between the distribution of the two distributions of decoding accuracies on every time point and for dataset separately.</p>
<p>Priors for the Bayes analysis can be selected based on previous work or can be estimated based on predetermined Cauchy distribution according to common effect sizes. We opted to use default priors. This choice was motivated by the absence of identical studies to ours available from which we could accurately estimate priors and the awareness that publication biases in any case will tend to exaggerate effect sizes. The priors for all Bayes factor analyses were determined based on Jeffrey-Zellner-Siow priors (<xref ref-type="bibr" rid="B90">Zellner and Siow, 1980</xref>; <xref ref-type="bibr" rid="B32">Jeffreys, 1998</xref>) which are from the Cauchy distribution based on the effect size that is initially calculated in the algorithm using t-test (<xref ref-type="bibr" rid="B65">Rouder et al., 2012</xref>). The priors are data-driven and have been shown to be invariant with respect to linear transformations of measurement units (<xref ref-type="bibr" rid="B65">Rouder et al., 2012</xref>), which reduces the chance of being biased toward the null or alternative hypotheses. We did not perform correction for multiple comparisons when using Bayes factors as they are much more conservative than frequentist analysis in providing false claims with confidence (<xref ref-type="bibr" rid="B19">Gelman and Tuerlinckx, 2000</xref>; <xref ref-type="bibr" rid="B20">Gelman et al., 2012</xref>). The reason for this is that properly chosen priors [here using the data-driven approach developed by <xref ref-type="bibr" rid="B65">Rouder et al. (2012)</xref>], reduce the chance of making type I (false positive) errors (<xref ref-type="bibr" rid="B19">Gelman and Tuerlinckx, 2000</xref>; <xref ref-type="bibr" rid="B20">Gelman et al., 2012</xref>).</p>
</sec>
<sec id="S2.SS8.SSS2">
<title>Random Permutation Testing</title>
<p>To evaluate the significance of correlations between decoding accuracies and behavioral reaction times, we calculated the percentage of the actual correlations that were higher (if positive) or lower (if negative) than a set of 1,000 randomly generated correlations. These random correlations were obtained by randomizing the order of participants&#x2019; data in the behavioral reaction time vector (null distribution) on every time point, for each feature separately. The correlation was considered significant if surpassed 95% of the randomly generated correlations in the null distribution in either positive or negative directions (<italic>p</italic> &#x003C; 0.05) and the <italic>p</italic>-values were corrected for multiple comparisons across time using Matlab mafdr function, where the algorithm fixes the rejection region and then estimates its corresponding error rate resulting in increased accuracy and power (<xref ref-type="bibr" rid="B73">Storey, 2002</xref>).</p>
</sec>
</sec>
</sec>
<sec id="S3" sec-type="results">
<title>Results</title>
<sec id="S3.SS1">
<title>Do Different Ways of Combining Individual Features Affect the Level and Temporal Dynamics of Information Decoding?</title>
<p>As an initial step, we evaluated the level of information which can be obtained from the combination of features, each potentially capturing different aspects of the neural codes. To be as confident as possible, we used a large set of 17 distinct supervised and unsupervised FS methods to select and combine the top 5 most informative features at every time point in the time-resolved decoding procedure. The information content of features were determined based on either how much they could contribute to discriminating the target object categories (supervised) or some predefined criteria which could implicitly suggest more separation between object categories (unsupervised). We split the FS algorithms into three arbitrary groups for the sake of clearer presentation of the results (<xref ref-type="fig" rid="F2">Figure 2</xref>).</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption><p>Time-resolved decoding of object categories from the three datasets using the 17 FS methods. We split the FS algorithms into three arbitrary groups (rows) for each dataset for the sake of clearer presentation. Each column shows the results for one dataset. The top section in each of the nine panels shows the decoding accuracies across time and the bottom panels show the Bayes factor evidence for decoding to be different (H1) or not different (H0) from chance-level. The horizontal dashed lines refer to chance-level decoding, the vertical dashed lines indicates time of stimulus onset. Non-black colored filled circles in the Bayes Factors show moderate (BF &#x003E; 3) or strong (BF &#x003E; 10) evidence for difference from chance-level decoding, black filled circles show moderate (BF &#x003E; 3) or strong (BF &#x003E; 10) evidence for no difference from chance-level decoding and empty circles indicate insufficient evidence (1/3 &#x003C; BF &#x003C; 3) for either hypotheses.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-16-825746-g002.tif"/>
</fig>
<p>All FS algorithms for the three datasets showed strong (BF &#x003E; 10) evidence for difference from chance-level decoding at some time points/windows after the stimulus onset (<xref ref-type="fig" rid="F2">Figure 2</xref>). This means that, any of the FS algorithms could combine the features in a way that they could decode object category information from brain signals. As expected from the difference in their mathematical formulations, however, no pairs of FS algorithms provided identical patterns of decoding in any of the three datasets. Consistently across the three datasets there was moderate (3 &#x003C; BF &#x003C; 10) or strong (BF &#x003E; 10) evidence for continuous above-chance decoding from around 80 ms post stimulus onset for all FS algorithms. While the decoding showed evidence for above-chance accuracy (BF &#x003E; 3) up until 550 ms (dataset 2) or even later than 800 ms (dataset 3) for the best FS algorithms such as UDFS, lasso and ufsol, all curves converged back to the chance-level earlier than 500 ms for dataset 1. This difference may reflect the longer stimulus presentation time for datasets 2 and 3 vs. dataset 1, which may have provided stronger sensory input for neural processing of category information, as we saw previously when evaluating individual features alone (<xref ref-type="bibr" rid="B39">Karimi-Rouzbahani et al., 2021b</xref>).</p>
<p>In order to quantitatively compare the decoding curves for the different FS algorithms, we extracted four different amplitude and timing parameters from their decoding curves as in previous studies (<xref ref-type="bibr" rid="B29">Isik et al., 2014</xref>): maximum and average decoding accuracies (in the post-stimulus time window), time of maximum decoding, and time of first above-chance decoding relative to stimulus onset (<xref ref-type="supplementary-material" rid="PS1">Supplementary Figure 1</xref>). Results showed that ILFS, relief and llcfs were the worst performing FS algorithms with the lowest maximum and average decoding accuracy (<xref ref-type="supplementary-material" rid="PS1">Supplementary Figures 1A,B</xref>; red boxes). UDFS, lasso and ufsol were the best performing FS algorithms leading to the highest maximum and average decoding accuracies (<xref ref-type="supplementary-material" rid="PS1">Supplementary Figures 1A,B</xref>; black boxes). Dataset 2 tended to yield higher decoding accuracies compared to the other datasets, which might be attributed to the longer presentation time of the stimuli and the active task of the participants (<xref ref-type="bibr" rid="B64">Roth et al., 2020</xref>; <xref ref-type="bibr" rid="B38">Karimi-Rouzbahani et al., 2021a</xref>,<xref ref-type="bibr" rid="B40">c</xref>). UDFS, ufsol and relief were among the earliest FS algorithms to reach their first above-chance and maximum decoding accuracies (<xref ref-type="supplementary-material" rid="PS1">Supplementary Figures 1C,D</xref>). However, there was not a consistent pattern of temporal precedence for any FS algorithms across the datasets.</p>
</sec>
<sec id="S3.SS2">
<title>Which Individual Features Are Selected by the Most Successful Algorithms?</title>
<p>The difference in the decoding patterns for different FS algorithms suggest that they used different sets of features in decoding. To see what features were selected by different FS algorithms, and whether the informative individual features were selected, we calculated the merit of each of the individual features in each FS algorithm across the time course of the trial (<xref ref-type="supplementary-material" rid="PS2">Supplementary Figure 2</xref>). Here, merit refers to the frequency of a feature being selected by the FS algorithm for decoding. We calculated the merit as the ratio of the number of times the feature was among the top selected five features to the number of times the decoding was performed on every time point (i.e., all possible combination of category pairs).</p>
<p>Visual inspection of the results suggests that each FS algorithm seemed to rely on consistent sets of features across the three datasets, which are generally different between FS algorithms. This reflects that different FS algorithms have different levels of sensitivity and distinct selection criteria. Results also showed that the merit of different features varied across the time course of trials based on their information content about object categories relative to other features (<xref ref-type="supplementary-material" rid="PS2">Supplementary Figure 2</xref>). Therefore, the recruitment of features varied across the time course of the trial: while some features were only temporarily selected (e.g., Average and Mean frequency in the laplacian method from &#x223C;200 to 600 post-stimulus onset), there were features which were constantly used for decoding even before the stimulus onset (e.g., Cros Cor in the fsasl method), although they did not lead to any information decoding in the pre-stimulus time span (<xref ref-type="fig" rid="F2">Figure 2</xref>). This might again be explained by the different levels of sensitivity and distinct selection criteria implemented by different FS algorithms. Importantly, the FS algorithms that provided the highest level of decoding (i.e., ufsol, lasso, and UDFS) showed the highest merits for the features of Mean, Median, Samples, and Wavelet which were among the most informative features when evaluated individually across the three datasets (<xref ref-type="bibr" rid="B39">Karimi-Rouzbahani et al., 2021b</xref>). On the other hand, the FS algorithms that performed most poorly (ILFS, relief, and llcfs) either used scattered sets of features (ILFS) or did not use the informative features of Mean, Median, Samples and Wavelet (llcfs and relief). Therefore, the FS algorithms that used the informative individual features outperformed other FS algorithms which did not.</p>
</sec>
<sec id="S3.SS3">
<title>Are the Neural Codes Better Captured by a Combinatorial Encoding Protocol or by a General Multiscale Encoding Protocol?</title>
<p>The main question of this study was to see whether the flexibility obtained by the combination of features provides any additional information about object categories compared to the best-performing individual features by detecting the neural codes more completely. In other words, we wanted to test the hypothesis that the brain uses a combination of different neural encoding protocols simultaneously as opposed to using a general multiscale encoding protocol (such as reflected in the Wavelet transform). To test this hypothesis, we directly compared the decoding accuracy obtained from the top performing individual feature from the original study (Wavelet; <xref ref-type="bibr" rid="B39">Karimi-Rouzbahani et al., 2021b</xref>), which is able to detect multiscale spatiotemporal patterns of information, with the decoding accuracy obtained from the top performing FS algorithm, which used a set of combined features (ufsol; <xref ref-type="fig" rid="F3">Figure 3</xref>). Results showed consistent patterns across the three datasets with the Wavelet feature outperforming the decoding accuracies obtained by the ufsol FS algorithm across most time points. Maximum continuous evidence for difference (BF &#x003E; 10) occurred between 80 and 320, 75&#x2013;180, and 85&#x2013;325 ms for datasets 1&#x2013;3, respectively. Therefore, it seems that, at least for object categories, the coding scheme in the brain is best captured by a general multiscale encoding protocol (implemented here by the Wavelet coefficients), rather than a combination of distinct encoding protocols (captured here by different features).</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption><p>Comparison of decoding accuracies between the most informative individual feature [Wavelet; from <xref ref-type="bibr" rid="B39">Karimi-Rouzbahani et al. (2021b)</xref>] and combined feature set (obtained using ufsol algorithm) from the three datasets and Bayesian evidence for a difference between them. Each column shows the results for one dataset. Thick lines show the average decoding accuracy across participants (error bars show Standard Error across participants). Top section in each panel shows the decoding accuracies across time and the bottom section shows the Bayes factor evidence for the difference between the decoding curves. The horizontal dashed lines on the top panels refer to chance-level decoding. Red filled circles in the Bayes Factors show moderate (BF &#x003E; 3) or strong (BF &#x003E; 10) evidence for difference between decoding curves, black filled circles show moderate (BF &#x003E; 3) or strong (BF &#x003E; 10) evidence for no difference and empty circles indicate insufficient evidence (1/3 &#x003C; BF &#x003C; 3) for either hypotheses.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-16-825746-g003.tif"/>
</fig>
</sec>
<sec id="S3.SS4">
<title>Can a Combinatorial Encoding Protocol Predict Behavioral Accuracy Better Than a General Multiscale Encoding Protocol?</title>
<p>Our final hypothesis was that a combinatorial encoding protocol might predict the behavioral performance more accurately than a general multiscale encoding protocol as the former can potentially detect more distinctly encoded neural codes from brain activation. We could test this hypothesis only for Dataset 2 where the task was active and we had the participants&#x2019; reaction times (i.e., time to categorize objects) to work with. We calculated the (Spearman&#x2019;s rank) correlation between the decoding accuracies and the behavioral reaction time across participants, to see whether, at each time point, participants with higher decoding values were those with the fastest reaction times. We expected to observe negative correlations between the decoding accuracies and the participants&#x2019; reaction times in the post-stimulus span (<xref ref-type="bibr" rid="B62">Ritchie et al., 2015</xref>). Note that since correlation normalizes the absolute level of the input variables, the higher level of decoding for the individual (Wavelet) feature vs. the combined features (ufsol; <xref ref-type="fig" rid="F3">Figure 3</xref>) does not necessarily predict a higher correlation for the individual feature of Wavelet.</p>
<p>Results showed significant negative correlations appearing after the stimulus onset for most FS algorithms (except dgufs) especially the laplacian algorithm which showed the most negative peak (<xref ref-type="fig" rid="F4">Figure 4A</xref>). This confirms that the distances between object categories in neural representations have inverse relationship to behavioral reaction times (<xref ref-type="bibr" rid="B62">Ritchie et al., 2015</xref>). We previously observed that the individual features which provided the highest decoding accuracies could also predict the behavior most accurately (<xref ref-type="bibr" rid="B39">Karimi-Rouzbahani et al., 2021b</xref>). Therefore, we asked if the FS algorithms which provided the highest levels of decoding could also predict the behavior more accurately than the less informative algorithms. The rationale behind this hypothesis was that, more effective decoding of neural codes, as measured by higher &#x201C;average decoding&#x201D; and &#x201C;maximum decoding&#x201D; accuracies (<xref ref-type="fig" rid="F2">Figure 2</xref>), should facilitate the prediction of behavior by detecting subtle but overlooked behavior-related neural codes. To test this hypothesis, we evaluated the correlation between the parameters of &#x201C;maximum decoding&#x201D; and &#x201C;average decoding&#x201D; accuracies (extracted from the decoding curve of each feature in <xref ref-type="fig" rid="F4">Figure 4A</xref>) and the &#x201C;average correlation to behavior&#x201D; (calculated simply by averaging the correlation to behavior in the post-stimulus time span for each FS algorithm in <xref ref-type="fig" rid="F4">Figure 4A</xref>). We also calculated the correlation between the &#x201C;time of maximum decoding&#x201D; and &#x201C;time of first above-chance decoding&#x201D; as control variables, which we did not expect to correlate with behavior (as in <xref ref-type="bibr" rid="B39">Karimi-Rouzbahani et al., 2021b</xref>). Results showed no significant correlations between any of the four parameters of decoding curves and the level of prediction of behavior (<xref ref-type="fig" rid="F4">Figure 4B</xref>). Therefore, more efficient combinations of features (as measured by higher decoding accuracies) did not correspond to more accurate prediction of behavior.</p>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption><p>Correlation between the decoding accuracies obtained using 17 FS algorithms and behavioral reaction time of Dataset 2. <bold>(A)</bold> Top section in each panel shows the (Spearman&#x2019;s) correlation coefficient obtained from correlating the decoding values and the reaction times for each feature separately. Thickened time points on the curves indicate time points of positively or negatively significant (<italic>P</italic> &#x003C; 0.05; corrected for multiple comparisons) correlations as evaluated by random permutation of the variables in correlation. <bold>(B)</bold> Correlation between each of the amplitude and timing parameters of time-resolved decoding (i.e., maximum and average decoding accuracy and time of first and maximum decoding) with the average time-resolved correlations calculated from panel <bold>(A)</bold> for the set of <italic>N</italic> = 17 FS algorithms. The slant line shows the best linear fit to the distribution of the correlation data. <bold>(C)</bold> Correlation between the decoding accuracies obtained from the feature which showed the highest maximum correlation from individual features (Wavelet) and from the combined features (laplacian).</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-16-825746-g004.tif"/>
</fig>
<p>To visually compare the behavioral prediction power of the top-performing individual and combined features we plotted their correlation-to-behavior results on the same figure (<xref ref-type="fig" rid="F4">Figure 4C</xref>). For this we selected Wavelet and laplacian FS, based on them being the single feature and FS algorithm with the largest negative peak. We used this, rather than selecting based on average correlation with behavior because the temporal position of the peak can also provide some temporal indication about the timing of the decision, which if reasonable [e.g., after 200 ms post-stimulus and before the median reaction times of participants: 1,146 ms (<xref ref-type="bibr" rid="B41">Karimi-Rouzbahani et al., 2019</xref>)], can be more assuring about the existence of true correlation to behavior. The combined features (laplacian) did not provide a negative peak as large as the Wavelet feature, and tended to underperform Wavelet throughout the time course (<xref ref-type="fig" rid="F4">Figure 4C</xref>). Therefore, in contradiction to our hypothesis, the combined features did not provide additional prediction of behavior compared to the individual feature of Wavelet.</p>
</sec>
</sec>
<sec id="S4" sec-type="discussion">
<title>Discussion</title>
<p>Abstract models of feed-forward visual processing suggest that visual sensory information enters the brain through retina, reaches the lateral geniculate nucleus in thalamus and continues to early visual cortices before moving forward (along the ventral visual stream) to reach the anterior parts of the inferior temporal cortices where semantic information (e.g., about the category of the presented object) is extracted from the visual inputs (<xref ref-type="bibr" rid="B14">DiCarlo et al., 2012</xref>). However, two outstanding questions are how neurons along the way encode the information and how this information is reflected in invasively (e.g., LFPs) and non-invasively collected (e.g., EEG) neural data. While in invasively recorded data, researchers have found significant information about visual information in low-frequency power of LFPs (<xref ref-type="bibr" rid="B5">Belitski et al., 2008</xref>) or phase-amplitude coupling of electrocorticography (ECoG), there is no reason for these to directly imprint on EEG. In fact, there is evidence that EEG activations represent the information in a feature different [e.g., phase rather than the amplitude of slow (theta band) oscillations] from the invasive neural data such as spiking activity (<xref ref-type="bibr" rid="B54">Ng et al., 2013</xref>). Therefore, more detailed investigation of neural coding in EEG seems necessary.</p>
<p>To gain a better understanding of EEG, previous studies have extracted a wide variety of features of neural activations to extract information about visual object categories. However, they have generally used whole-trial analyses, which hide the temporal dynamics of information processing, or time-resolved decoding analyses, or considered the response at each time point separately, ignoring potentially informative temporal features of the time series data. To fill this gap, our previous study extracted and compared a large set of features from EEG in time-resolved analysis (<xref ref-type="bibr" rid="B39">Karimi-Rouzbahani et al., 2021b</xref>). However, an outstanding question in the literature was whether the neural code might be best captured by combinations of these features, i.e., if the brain uses a combinatorial encoding protocol to encode different aspects of the sensory input using distinct encoding protocols on the same trial (<xref ref-type="bibr" rid="B18">Gawne et al., 1996</xref>; <xref ref-type="bibr" rid="B50">Montemurro et al., 2008</xref>). Alternatively, previous invasive neural recording studies have suggested a general multiscale encoding procedure that allows the generation of all the information within the same platform (<xref ref-type="bibr" rid="B78">Victor, 2000</xref>; <xref ref-type="bibr" rid="B42">Kayser et al., 2009</xref>; <xref ref-type="bibr" rid="B56">Panzeri et al., 2010</xref>). To address this question we combined a large set of distinct mathematical features (<italic>n</italic> = 26) of the EEG time series data from three datasets, and combined them using a large set of FS algorithms (<italic>n</italic> = 17), each having different criteria for selection. We compared the performance of different FS algorithms using multivariate decoding of category information. Our results showed that, no matter how we combined the informative features, their combined decodable information about object categories, and their power in predicting behavioral performance, was outperformed by the most informative individual feature (i.e., Wavelet), which was sensitive to multi-scale codes from the analysis time window and across electrodes (i.e., spatiotemporal specificity).</p>
<p>The main question of this study was whether the brain recruits and combines a number of different protocols to encode different aspects of cognitive processes involved in object category recognition ranging from sensory information to behavioral response. For example, the brain may use one encoding protocol for the encoding of feed-forward visual information processing, e.g., theta-band power, which would later in the trial be dominated by alpha/beta-band feedback information flow involved in semantic object categorization (<xref ref-type="bibr" rid="B4">Bastos et al., 2015</xref>). The brain may also use different encoding protocols to process different aspects of the same stimulus [e.g., contrast or the orientation of visual stimulus (<xref ref-type="bibr" rid="B18">Gawne et al., 1996</xref>)]. Alternatively, the brain may implement a single but multiscale protocol [e.g., multiplexing strategy which combines the codes at different time scales (<xref ref-type="bibr" rid="B56">Panzeri et al., 2010</xref>)] which allows different aspects of information to be represented within the same encoding protocol. Our results provide support for the latter by showing that spatiotemporally sensitive features, which can detect patterns across multiple scales (e.g., Wavelet coefficients) best capture variance in the EEG responses evoked by different categories of visual objects. Therefore, rather than a combinatorial and switching encoding protocol, the brain may instead encode object category information through a single but multiscale encoding protocol.</p>
<p>This study does not provide the first evidence showing that temporal patterns of activity provide information about different aspects of visual sensory input. The richness of information in the temporal patterns of activity has been previously observed in light encoding (<xref ref-type="bibr" rid="B21">Gollisch and Meister, 2008</xref>), co-occurrences of visual edges (<xref ref-type="bibr" rid="B16">Eckhorn et al., 1988</xref>), orientations in primary visual cortex (<xref ref-type="bibr" rid="B8">Celebrini et al., 1993</xref>) as well as object category information in the temporal cortex (<xref ref-type="bibr" rid="B47">Majima et al., 2014</xref>). While we do not claim that this EEG study provides direct evidence about processing of information at the level of single neurons, our findings are consistent with the above invasively-recorded neural data and provide evidence for information content in neural variability of EEG data. Our study also aligns with the recent move toward incorporating within- and across-trial temporal variability in the decoding of information from neural time series such as MEG (<xref ref-type="bibr" rid="B79">Vidaurre et al., 2019</xref>), EEG (<xref ref-type="bibr" rid="B47">Majima et al., 2014</xref>), invasive electrophysiological (<xref ref-type="bibr" rid="B55">Orb&#x00E1;n et al., 2016</xref>) and even fMRI (<xref ref-type="bibr" rid="B17">Garrett et al., 2020</xref>) data. On the other hand, this current study contrasts with the conventional time-resolved decoding analyses which merely consider amplitude at each time point (<xref ref-type="bibr" rid="B22">Grootswagers et al., 2017</xref>), overlooking informative multi-scale temporal codes.</p>
<p>The field of Brain-Computer Interface (BCI) has already achieved great success in decoding visually evoked information from EEG representations in the past two decades, mainly through the use of rigorous supervised learning algorithms [e.g., Voltage Topographies (<xref ref-type="bibr" rid="B77">Tzovara et al., 2012</xref>), Independent Component Analysis (<xref ref-type="bibr" rid="B72">Stewart et al., 2014</xref>), Common Spatial Patterns (<xref ref-type="bibr" rid="B51">Murphy et al., 2011</xref>), and Convolutional Neural Networks (<xref ref-type="bibr" rid="B69">Seeliger et al., 2018</xref>)] or by combining multiple features (<xref ref-type="bibr" rid="B9">Chan et al., 2011</xref>; <xref ref-type="bibr" rid="B82">Wang et al., 2012</xref>; <xref ref-type="bibr" rid="B59">Qin et al., 2016</xref>; <xref ref-type="bibr" rid="B76">Torabi et al., 2017</xref>). However, the predictive power of a feature about behavior might not be as important for BCI where the goal is to maximize the accuracy of the commands sent to a computer or an actuator. In contrast, one of the most critical questions in cognitive neuroscience to understand whether the neural signatures that we observe are meaningful in bringing about behavior, as opposed to being epiphenomenal to our experimental setup (e.g., <xref ref-type="bibr" rid="B87">Williams et al., 2007</xref>; <xref ref-type="bibr" rid="B30">Jacobs et al., 2009</xref>; <xref ref-type="bibr" rid="B62">Ritchie et al., 2015</xref>; <xref ref-type="bibr" rid="B24">Hebart and Baker, 2018</xref>; <xref ref-type="bibr" rid="B89">Woolgar et al., 2019</xref>; <xref ref-type="bibr" rid="B38">Karimi-Rouzbahani et al., 2021a</xref>,<xref ref-type="bibr" rid="B39">b</xref>). To address this point, we evaluated whether our extracted features and their combinations were behaviorally relevant, by correlating our decoding patterns with the behavioral object recognition performance (reaction times in Dataset 2). Moreover, to directly compare the information content of the combined feature sets with the individual features, we equalized the dimensions of the data matrix for the FS algorithm to that obtained for individual features. This avoided artefactualy improving behavioral predictive power with higher dimensionality. Contrary to what we predicted, however, we observed that even the laplacian FS algorithm, which provided the best peak prediction for the behavioral performance, was outperformed by the individual Wavelet feature at most time points. Therefore, the multiscale feature of Wavelet not only provides the most decodable information, but seems to most closely reflect the neural processes involved in generating participant behavior.</p>
<p>One unique property of our decoding pipeline, which we believe led to the enhanced information encoding for the Wavelet feature relative to other individual features (<xref ref-type="bibr" rid="B39">Karimi-Rouzbahani et al., 2021b</xref>), is the incorporation of <italic>spatiotemporal</italic> codes in decoding in each 50 ms analysis window. The neural code can be represented in either time (across the analysis time window), space (across electrodes in EEG) or a combination of both (<xref ref-type="bibr" rid="B56">Panzeri et al., 2010</xref>). Specifically, most of the previous studies have evaluated the neural codes in either time, being limited by the nature of their invasive recording modality (<xref ref-type="bibr" rid="B26">Houweling and Brecht, 2008</xref>; <xref ref-type="bibr" rid="B6">Benucci et al., 2009</xref>), or space by averaging/down-sampling of data within the analysis window. However, our spatiotemporal concatenation of EEG activity across both time and electrodes (i.e., performed at the first PCA stage for individual features and at the third PCA stage for the combined features in <xref ref-type="fig" rid="F1">Figure 1</xref>), allows the neural codes to be detected from both spatially and temporally informative patterns. The 50 ms time window chosen here makes a compromise between concatenating and decoding the whole time window in one shot, which loses the temporal resolution, and time-resolved decoding at each time point, which ignores temporal patterns of activity (<xref ref-type="bibr" rid="B39">Karimi-Rouzbahani et al., 2021b</xref>).</p>
<p>While this study provided insights about how neural codes might be detected from EEG activations, there remain two main limitations in understanding the nature of neural codes in EEG. First, physiological evidence is limited about how neurons produce, often such complicated codes, even in studies where the mathematical features of this study were first introduced. There are theories and mathematical justifications to explain why these complicated codes are helpful (<xref ref-type="bibr" rid="B67">Schaefer et al., 2006</xref>; <xref ref-type="bibr" rid="B42">Kayser et al., 2009</xref>; <xref ref-type="bibr" rid="B68">Schroeder and Lakatos, 2009</xref>, etc.) but not on how neurons produce them. Second, it seems unlikely that the distinctly-defined mathematical features necessarily extract distinct attributes/neural codes. In fact, many of the extracted features overlap: some of them are slightly different ways of quantifying similar characteristics of the neural activity (e.g., variance vs. power, which both quantify the strength of variability of the signal). Therefore, there are not necessarily distinct neural underpinnings for each feature.</p>
<p>There are several future directions for this research. First, as the encoding protocols for different cognitive processes might be different from object category processing (<xref ref-type="bibr" rid="B56">Panzeri et al., 2010</xref>), the generalization of our results to other domains of cognitive neuroscience needs to be evaluated. Second, previous results (<xref ref-type="bibr" rid="B56">Panzeri et al., 2010</xref>) suggest that different aspects of information (e.g., category processing, decision making and motor response) may be encoded using different encoding protocols. Our data did not allow us to tease those aspects apart, which is interesting area for future investigation. Third, following previous suggestions that even different aspects of <italic>visual</italic> information (e.g., color, variations, and task) might also be encoded using different encoding protocols (<xref ref-type="bibr" rid="B18">Gawne et al., 1996</xref>), the number of selected features might need to be varied from one dataset to another. Ideally, we would only keep the informative features above a certain threshold. Here, we chose an arbitrary threshold of 5 included, but it would be interesting to explore the impact of this parameter in the future.</p>
<p>The large-scale EEG analysis of this study aligns with the recent shift to cross-dataset meta-analyses for different human cognitive abilities such as working memory (<xref ref-type="bibr" rid="B1">Adam et al., 2020</xref>) and sustained attention (<xref ref-type="bibr" rid="B43">Langner and Eickhoff, 2013</xref>). Such studies lead to more generalizable conclusions and provide deeper insights into the human cognition. Here, across three very different datasets we showed that, the brain seems to implement a temporally and spatially flexible and multiscale encoding strategy rather than a combinatorial or switching encoding strategy, at least in object category processing.</p>
</sec>
<sec id="S5" sec-type="data-availability">
<title>Data Availability Statement</title>
<p>Datasets 1 and 2 of this study are available online at <ext-link ext-link-type="uri" xlink:href="https://osf.io/wbvpn/">https://osf.io/wbvpn/</ext-link> and dataset 3 at <ext-link ext-link-type="uri" xlink:href="https://exhibits.stanford.edu/data/catalog/tc919dd5388">https://exhibits.stanford.edu/data/catalog/tc919dd5388</ext-link>.</p>
</sec>
<sec id="S6">
<title>Ethics Statement</title>
<p>The datasets used in this study were obtained from experiments that were approved by Shahid Rajaee University Ethics Committee, Iran, and Institutional Review Board of Stanford University, United States. The patients/participants provided their written informed consent to participate in this study.</p>
</sec>
<sec id="S7">
<title>Author Contributions</title>
<p>HK-R: conceptualization, methodology, formal analysis, writing &#x2013; original draft, visualization, data curation, and funding acquisition. AW: writing &#x2013; review and editing and funding acquisition. Both authors contributed to the article and approved the submitted version.</p>
</sec>
<sec id="conf1" sec-type="COI-statement">
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec id="pudiscl1" sec-type="disclaimer">
<title>Publisher&#x2019;s Note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
</body>
<back>
<sec id="S8" sec-type="funding-information">
<title>Funding</title>
<p>This research was funded by the UK Royal Society&#x2019;s Newton International Fellowship NIF\R\192608 to HK-R and MRC intramural funding SUAG/052/G101400 to AW.</p>
</sec>
<sec id="S9" sec-type="supplementary-material">
<title>Supplementary Material</title>
<p>The Supplementary Material for this article can be found online at: <ext-link ext-link-type="uri" xlink:href="https://www.frontiersin.org/articles/10.3389/fnins.2022.825746/full#supplementary-material">https://www.frontiersin.org/articles/10.3389/fnins.2022.825746/full#supplementary-material</ext-link></p>
<supplementary-material xlink:href="Presentation_1.pdf" id="PS1" mimetype="application/pdf" xmlns:xlink="http://www.w3.org/1999/xlink"/>
<supplementary-material xlink:href="Presentation_2.pdf" id="PS2" mimetype="application/pdf" xmlns:xlink="http://www.w3.org/1999/xlink"/>
</sec>
<ref-list>
<title>References</title>
<ref id="B1"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Adam</surname> <given-names>K. C.</given-names></name> <name><surname>Vogel</surname> <given-names>E. K.</given-names></name> <name><surname>Awh</surname> <given-names>E.</given-names></name></person-group> (<year>2020</year>). <article-title>Multivariate analysis reveals a generalizable human electrophysiological signature of working memory load.</article-title> <source><italic>bioRxiv</italic></source> [<comment>Preprint</comment>]. <pub-id pub-id-type="doi">10.1111/psyp.13691</pub-id> <pub-id pub-id-type="pmid">33040349</pub-id></citation></ref>
<ref id="B2"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ahmadi-Pajouh</surname> <given-names>M. A.</given-names></name> <name><surname>Ala</surname> <given-names>T. S.</given-names></name> <name><surname>Zamanian</surname> <given-names>F.</given-names></name> <name><surname>Namazi</surname> <given-names>H.</given-names></name> <name><surname>Jafari</surname> <given-names>S.</given-names></name></person-group> (<year>2018</year>). <article-title>Fractal-based classification of human brain response to living and non-living visual stimuli.</article-title> <source><italic>Fractals</italic></source> <volume>26</volume>:<fpage>1850069</fpage>.</citation></ref>
<ref id="B3"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Alimardani</surname> <given-names>F.</given-names></name> <name><surname>Cho</surname> <given-names>J. H.</given-names></name> <name><surname>Boostani</surname> <given-names>R.</given-names></name> <name><surname>Hwang</surname> <given-names>H. J.</given-names></name></person-group> (<year>2018</year>). <article-title>Classification of bipolar disorder and schizophrenia using steady-state visual evoked potential based features.</article-title> <source><italic>IEEE Access</italic></source> <volume>6</volume> <fpage>40379</fpage>&#x2013;<lpage>40388</lpage>.</citation></ref>
<ref id="B4"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bastos</surname> <given-names>A. M.</given-names></name> <name><surname>Vezoli</surname> <given-names>J.</given-names></name> <name><surname>Bosman</surname> <given-names>C. A.</given-names></name> <name><surname>Schoffelen</surname> <given-names>J. M.</given-names></name> <name><surname>Oostenveld</surname> <given-names>R.</given-names></name> <name><surname>Dowdall</surname> <given-names>J. R.</given-names></name><etal/></person-group> (<year>2015</year>). <article-title>Visual areas exert feedforward and feedback influences through distinct frequency channels.</article-title> <source><italic>Neuron</italic></source> <volume>85</volume> <fpage>390</fpage>&#x2013;<lpage>401</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuron.2014.12.018</pub-id> <pub-id pub-id-type="pmid">25556836</pub-id></citation></ref>
<ref id="B5"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Belitski</surname> <given-names>A.</given-names></name> <name><surname>Gretton</surname> <given-names>A.</given-names></name> <name><surname>Magri</surname> <given-names>C.</given-names></name> <name><surname>Murayama</surname> <given-names>Y.</given-names></name> <name><surname>Montemurro</surname> <given-names>M. A.</given-names></name> <name><surname>Logothetis</surname> <given-names>N. K.</given-names></name><etal/></person-group> (<year>2008</year>). <article-title>Low-frequency local field potentials and spikes in primary visual cortex convey independent visual information.</article-title> <source><italic>J. Neurosci.</italic></source> <volume>28</volume> <fpage>5696</fpage>&#x2013;<lpage>5709</lpage>. <pub-id pub-id-type="doi">10.1523/JNEUROSCI.0009-08.2008</pub-id> <pub-id pub-id-type="pmid">18509031</pub-id></citation></ref>
<ref id="B6"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Benucci</surname> <given-names>A.</given-names></name> <name><surname>Ringach</surname> <given-names>D. L.</given-names></name> <name><surname>Carandini</surname> <given-names>M.</given-names></name></person-group> (<year>2009</year>). <article-title>Coding of stimulus sequences by population responses in visual cortex.</article-title> <source><italic>Nat. Neurosci.</italic></source> <volume>12</volume> <fpage>1317</fpage>&#x2013;<lpage>1324</lpage>. <pub-id pub-id-type="doi">10.1038/nn.2398</pub-id> <pub-id pub-id-type="pmid">19749748</pub-id></citation></ref>
<ref id="B7"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Carlson</surname> <given-names>T.</given-names></name> <name><surname>Tovar</surname> <given-names>D. A.</given-names></name> <name><surname>Alink</surname> <given-names>A.</given-names></name> <name><surname>Kriegeskorte</surname> <given-names>N.</given-names></name></person-group> (<year>2013</year>). <article-title>Representational dynamics of object vision: the first 1000 ms.</article-title> <source><italic>J. Vis.</italic></source> <volume>13</volume>:<fpage>1</fpage>. <pub-id pub-id-type="doi">10.1167/13.10.1</pub-id></citation></ref>
<ref id="B8"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Celebrini</surname> <given-names>S.</given-names></name> <name><surname>Thorpe</surname> <given-names>S.</given-names></name> <name><surname>Trotter</surname> <given-names>Y.</given-names></name> <name><surname>Imbert</surname> <given-names>M.</given-names></name></person-group> (<year>1993</year>). <article-title>Dynamics of orientation coding in area V1 of the awake primate.</article-title> <source><italic>Vis. Neurosci.</italic></source> <volume>10</volume> <fpage>811</fpage>&#x2013;<lpage>825</lpage>. <pub-id pub-id-type="doi">10.1017/s0952523800006052</pub-id> <pub-id pub-id-type="pmid">8217934</pub-id></citation></ref>
<ref id="B9"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chan</surname> <given-names>A. M.</given-names></name> <name><surname>Halgren</surname> <given-names>E.</given-names></name> <name><surname>Marinkovic</surname> <given-names>K.</given-names></name> <name><surname>Cash</surname> <given-names>S. S.</given-names></name></person-group> (<year>2011</year>). <article-title>Decoding word and category-specific spatiotemporal representations from MEG and EEG.</article-title> <source><italic>Neuroimage</italic></source> <volume>54</volume> <fpage>3028</fpage>&#x2013;<lpage>3039</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuroimage.2010.10.073</pub-id> <pub-id pub-id-type="pmid">21040796</pub-id></citation></ref>
<ref id="B10"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Cichy</surname> <given-names>R. M.</given-names></name> <name><surname>Pantazis</surname> <given-names>D.</given-names></name> <name><surname>Oliva</surname> <given-names>A.</given-names></name></person-group> (<year>2014</year>). <article-title>Resolving human object recognition in space and time.</article-title> <source><italic>Nat. Neurosci.</italic></source> <volume>17</volume>:<fpage>455</fpage>. <pub-id pub-id-type="doi">10.1038/nn.3635</pub-id> <pub-id pub-id-type="pmid">24464044</pub-id></citation></ref>
<ref id="B11"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Contini</surname> <given-names>E. W.</given-names></name> <name><surname>Wardle</surname> <given-names>S. G.</given-names></name> <name><surname>Carlson</surname> <given-names>T. A.</given-names></name></person-group> (<year>2017</year>). <article-title>Decoding the time-course of object recognition in the human brain: from visual features to categorical decisions.</article-title> <source><italic>Neuropsychologia</italic></source> <volume>105</volume> <fpage>165</fpage>&#x2013;<lpage>176</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuropsychologia.2017.02.013</pub-id> <pub-id pub-id-type="pmid">28215698</pub-id></citation></ref>
<ref id="B12"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>da Silva</surname> <given-names>F. L.</given-names></name></person-group> (<year>2013</year>). <article-title>EEG and MEG: relevance to neuroscience.</article-title> <source><italic>Neuron</italic></source> <volume>80</volume> <fpage>1112</fpage>&#x2013;<lpage>1128</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuron.2013.10.017</pub-id> <pub-id pub-id-type="pmid">24314724</pub-id></citation></ref>
<ref id="B13"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Dash</surname> <given-names>M.</given-names></name> <name><surname>Liu</surname> <given-names>H.</given-names></name></person-group> (<year>1997</year>). <article-title>Feature selection for classification.</article-title> <source><italic>Intell. Data Anal.</italic></source> <volume>1</volume> <fpage>131</fpage>&#x2013;<lpage>156</lpage>.</citation></ref>
<ref id="B14"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>DiCarlo</surname> <given-names>J. J.</given-names></name> <name><surname>Zoccolan</surname> <given-names>D.</given-names></name> <name><surname>Rust</surname> <given-names>N. C.</given-names></name></person-group> (<year>2012</year>). <article-title>How does the brain solve visual object recognition?</article-title> <source><italic>Neuron</italic></source> <volume>73</volume> <fpage>415</fpage>&#x2013;<lpage>434</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuron.2012.01.010</pub-id> <pub-id pub-id-type="pmid">22325196</pub-id></citation></ref>
<ref id="B15"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Dienes</surname> <given-names>Z.</given-names></name></person-group> (<year>2014</year>). <article-title>Using Bayes to get the most out of non-significant results.</article-title> <source><italic>Front. Psychol.</italic></source> <volume>5</volume>:<fpage>781</fpage>. <pub-id pub-id-type="doi">10.3389/fpsyg.2014.00781</pub-id> <pub-id pub-id-type="pmid">25120503</pub-id></citation></ref>
<ref id="B16"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Eckhorn</surname> <given-names>R.</given-names></name> <name><surname>Bauer</surname> <given-names>R.</given-names></name> <name><surname>Jordan</surname> <given-names>W.</given-names></name> <name><surname>Brosch</surname> <given-names>M.</given-names></name> <name><surname>Kruse</surname> <given-names>W.</given-names></name> <name><surname>Munk</surname> <given-names>M.</given-names></name><etal/></person-group> (<year>1988</year>). <article-title>Coherent oscillations: a mechanism of feature linking in the visual cortex?</article-title> <source><italic>Biol. Cybernet.</italic></source> <volume>60</volume> <fpage>121</fpage>&#x2013;<lpage>130</lpage>. <pub-id pub-id-type="doi">10.1007/BF00202899</pub-id> <pub-id pub-id-type="pmid">3228555</pub-id></citation></ref>
<ref id="B17"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Garrett</surname> <given-names>D. D.</given-names></name> <name><surname>Epp</surname> <given-names>S. M.</given-names></name> <name><surname>Kleemeyer</surname> <given-names>M.</given-names></name> <name><surname>Lindenberger</surname> <given-names>U.</given-names></name> <name><surname>Polk</surname> <given-names>T. A.</given-names></name></person-group> (<year>2020</year>). <article-title>Higher performers upregulate brain signal variability in response to more feature-rich visual input.</article-title> <source><italic>Neuroimage</italic></source> <volume>217</volume>:<fpage>116836</fpage>. <pub-id pub-id-type="doi">10.1016/j.neuroimage.2020.116836</pub-id> <pub-id pub-id-type="pmid">32283277</pub-id></citation></ref>
<ref id="B18"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gawne</surname> <given-names>T. J.</given-names></name> <name><surname>Kjaer</surname> <given-names>T. W.</given-names></name> <name><surname>Richmond</surname> <given-names>B. J.</given-names></name></person-group> (<year>1996</year>). <article-title>Latency: another potential code for feature binding in striate cortex.</article-title> <source><italic>J. Neurophysiol.</italic></source> <volume>76</volume> <fpage>1356</fpage>&#x2013;<lpage>1360</lpage>. <pub-id pub-id-type="doi">10.1152/jn.1996.76.2.1356</pub-id> <pub-id pub-id-type="pmid">8871243</pub-id></citation></ref>
<ref id="B19"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gelman</surname> <given-names>A.</given-names></name> <name><surname>Tuerlinckx</surname> <given-names>F.</given-names></name></person-group> (<year>2000</year>). <article-title>Type S error rates for classical and Bayesian single and multiple comparison procedures.</article-title> <source><italic>Comput. Stat.</italic></source> <volume>15</volume> <fpage>373</fpage>&#x2013;<lpage>390</lpage>.</citation></ref>
<ref id="B20"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gelman</surname> <given-names>A.</given-names></name> <name><surname>Hill</surname> <given-names>J.</given-names></name> <name><surname>Yajima</surname> <given-names>M.</given-names></name></person-group> (<year>2012</year>). <article-title>Why we (usually) don&#x2019;t have to worry about multiple comparisons.</article-title> <source><italic>J. Res. Educ. Effect.</italic></source> <volume>5</volume> <fpage>189</fpage>&#x2013;<lpage>211</lpage>.</citation></ref>
<ref id="B21"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gollisch</surname> <given-names>T.</given-names></name> <name><surname>Meister</surname> <given-names>M.</given-names></name></person-group> (<year>2008</year>). <article-title>Rapid neural coding in the retina with relative spike latencies.</article-title> <source><italic>Science</italic></source> <volume>319</volume> <fpage>1108</fpage>&#x2013;<lpage>1111</lpage>. <pub-id pub-id-type="doi">10.1126/science.1149639</pub-id> <pub-id pub-id-type="pmid">18292344</pub-id></citation></ref>
<ref id="B22"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Grootswagers</surname> <given-names>T.</given-names></name> <name><surname>Wardle</surname> <given-names>S. G.</given-names></name> <name><surname>Carlson</surname> <given-names>T. A.</given-names></name></person-group> (<year>2017</year>). <article-title>Decoding dynamic brain patterns from evoked responses: a tutorial on multivariate pattern analysis applied to time series neuroimaging data.</article-title> <source><italic>J. Cogn. Neurosci.</italic></source> <volume>29</volume> <fpage>677</fpage>&#x2013;<lpage>697</lpage>. <pub-id pub-id-type="doi">10.1162/jocn_a_01068</pub-id></citation></ref>
<ref id="B23"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Guo</surname> <given-names>L.</given-names></name> <name><surname>Rivero</surname> <given-names>D.</given-names></name> <name><surname>Seoane</surname> <given-names>J. A.</given-names></name> <name><surname>Pazos</surname> <given-names>A.</given-names></name></person-group> (<year>2009</year>). &#x201C;<article-title>Classification of EEG signals using relative wavelet energy and artificial neural networks</article-title>,&#x201D; in <source><italic>Proceedings of the first ACM/SIGEVO Summit on Genetic and Evolutionary Computation</italic></source>, <publisher-loc>Shanghai China</publisher-loc>, <fpage>177</fpage>&#x2013;<lpage>184</lpage>. <pub-id pub-id-type="doi">10.3233/SHTI210538</pub-id> <pub-id pub-id-type="pmid">34545817</pub-id></citation></ref>
<ref id="B24"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hebart</surname> <given-names>M. N.</given-names></name> <name><surname>Baker</surname> <given-names>C. I.</given-names></name></person-group> (<year>2018</year>). <article-title>Deconstructing multivariate decoding for the study of brain function.</article-title> <source><italic>Neuroimage</italic></source> <volume>180</volume> <fpage>4</fpage>&#x2013;<lpage>18</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuroimage.2017.08.005</pub-id> <pub-id pub-id-type="pmid">28782682</pub-id></citation></ref>
<ref id="B25"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hermundstad</surname> <given-names>A. M.</given-names></name> <name><surname>Briguglio</surname> <given-names>J. J.</given-names></name> <name><surname>Conte</surname> <given-names>M. M.</given-names></name> <name><surname>Victor</surname> <given-names>J. D.</given-names></name> <name><surname>Balasubramanian</surname> <given-names>V.</given-names></name> <name><surname>Tka&#x010D;ik</surname> <given-names>G.</given-names></name></person-group> (<year>2014</year>). <article-title>Variance predicts salience in central sensory processing.</article-title> <source><italic>Elife</italic></source> <volume>3</volume>:<fpage>e03722</fpage>. <pub-id pub-id-type="doi">10.7554/eLife.03722</pub-id> <pub-id pub-id-type="pmid">25396297</pub-id></citation></ref>
<ref id="B26"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Houweling</surname> <given-names>A. R.</given-names></name> <name><surname>Brecht</surname> <given-names>M.</given-names></name></person-group> (<year>2008</year>). <article-title>Behavioural report of single neuron stimulation in somatosensory cortex.</article-title> <source><italic>Nature</italic></source> <volume>451</volume> <fpage>65</fpage>&#x2013;<lpage>68</lpage>. <pub-id pub-id-type="doi">10.1038/nature06447</pub-id> <pub-id pub-id-type="pmid">18094684</pub-id></citation></ref>
<ref id="B27"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hung</surname> <given-names>C. P.</given-names></name> <name><surname>Kreiman</surname> <given-names>G.</given-names></name> <name><surname>Poggio</surname> <given-names>T.</given-names></name> <name><surname>DiCarlo</surname> <given-names>J. J.</given-names></name></person-group> (<year>2005</year>). <article-title>Fast readout of object identity from macaque inferior temporal cortex.</article-title> <source><italic>Science</italic></source> <volume>310</volume> <fpage>863</fpage>&#x2013;<lpage>866</lpage>. <pub-id pub-id-type="doi">10.1126/science.1117593</pub-id> <pub-id pub-id-type="pmid">16272124</pub-id></citation></ref>
<ref id="B28"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Iranmanesh</surname> <given-names>S.</given-names></name> <name><surname>Rodriguez-Villegas</surname> <given-names>E.</given-names></name></person-group> (<year>2017</year>). <article-title>An ultralow-power sleep spindle detection system on chip.</article-title> <source><italic>IEEE Trans. Biomed. Circuits Syst.</italic></source> <volume>11</volume> <fpage>858</fpage>&#x2013;<lpage>866</lpage>. <pub-id pub-id-type="doi">10.1109/TBCAS.2017.2690908</pub-id> <pub-id pub-id-type="pmid">28541914</pub-id></citation></ref>
<ref id="B29"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Isik</surname> <given-names>L.</given-names></name> <name><surname>Meyers</surname> <given-names>E. M.</given-names></name> <name><surname>Leibo</surname> <given-names>J. Z.</given-names></name> <name><surname>Poggio</surname> <given-names>T.</given-names></name></person-group> (<year>2014</year>). <article-title>The dynamics of invariant object recognition in the human visual system.</article-title> <source><italic>J. Neurophysiol.</italic></source> <volume>111</volume> <fpage>91</fpage>&#x2013;<lpage>102</lpage>. <pub-id pub-id-type="doi">10.1152/jn.00394.2013</pub-id> <pub-id pub-id-type="pmid">24089402</pub-id></citation></ref>
<ref id="B30"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jacobs</surname> <given-names>A. L.</given-names></name> <name><surname>Fridman</surname> <given-names>G.</given-names></name> <name><surname>Douglas</surname> <given-names>R. M.</given-names></name> <name><surname>Alam</surname> <given-names>N. M.</given-names></name> <name><surname>Latham</surname> <given-names>P. E.</given-names></name> <name><surname>Prusky</surname> <given-names>G. T.</given-names></name><etal/></person-group> (<year>2009</year>). <article-title>Ruling out and ruling in neural codes.</article-title> <source><italic>Proc. Natl. Acad. Sci.</italic></source> <volume>106</volume> <fpage>5936</fpage>&#x2013;<lpage>5941</lpage>. <pub-id pub-id-type="doi">10.1073/pnas.0900573106</pub-id> <pub-id pub-id-type="pmid">19297621</pub-id></citation></ref>
<ref id="B31"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jadidi</surname> <given-names>A. F.</given-names></name> <name><surname>Zargar</surname> <given-names>B. S.</given-names></name> <name><surname>Moradi</surname> <given-names>M. H.</given-names></name></person-group> (<year>2016</year>). &#x201C;<article-title>Categorizing visual objects; using ERP components</article-title>,&#x201D; in <source><italic>2016 23rd Iranian Conference on Biomedical Engineering and 2016 1st International Iranian Conference on Biomedical Engineering ICBME</italic></source> (<publisher-loc>Piscataway, NJ</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>159</fpage>&#x2013;<lpage>164</lpage>.</citation></ref>
<ref id="B32"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jeffreys</surname> <given-names>H.</given-names></name></person-group> (<year>1998</year>). <source><italic>The Theory Of Probability.</italic></source> <publisher-loc>Oxford</publisher-loc>: <publisher-name>Oxford University Press</publisher-name>.</citation></ref>
<ref id="B33"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Joshi</surname> <given-names>D.</given-names></name> <name><surname>Panigrahi</surname> <given-names>B. K.</given-names></name> <name><surname>Anand</surname> <given-names>S.</given-names></name> <name><surname>Santhosh</surname> <given-names>J.</given-names></name></person-group> (<year>2018</year>). <article-title>Classification of targets and distractors present in visual hemifields using time-frequency domain EEG features.</article-title> <source><italic>J. Healthc. Eng.</italic></source> <volume>2018</volume> <fpage>1</fpage>&#x2013;<lpage>10</lpage>. <pub-id pub-id-type="doi">10.1155/2018/9213707</pub-id> <pub-id pub-id-type="pmid">29808111</pub-id></citation></ref>
<ref id="B34"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kaneshiro</surname> <given-names>B.</given-names></name> <name><surname>Guimaraes</surname> <given-names>M. P.</given-names></name> <name><surname>Kim</surname> <given-names>H. S.</given-names></name> <name><surname>Norcia</surname> <given-names>A. M.</given-names></name> <name><surname>Suppes</surname> <given-names>P.</given-names></name></person-group> (<year>2015</year>). <article-title>A representational similarity analysis of the dynamics of object processing using single-trial EEG classification.</article-title> <source><italic>PLoS One</italic></source> <volume>10</volume>:<fpage>e0135697</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pone.0135697</pub-id> <pub-id pub-id-type="pmid">26295970</pub-id></citation></ref>
<ref id="B35"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Karimi-Rouzbahani</surname> <given-names>H.</given-names></name></person-group> (<year>2018</year>). <article-title>Three-stage processing of category and variation information by entangled interactive mechanisms of peri-occipital and peri-frontal cortices.</article-title> <source><italic>Sci. Rep.</italic></source> <volume>8</volume> <fpage>1</fpage>&#x2013;<lpage>22</lpage>. <pub-id pub-id-type="doi">10.1038/s41598-018-30601-8</pub-id> <pub-id pub-id-type="pmid">30111859</pub-id></citation></ref>
<ref id="B36"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Karimi-Rouzbahani</surname> <given-names>H.</given-names></name> <name><surname>Bagheri</surname> <given-names>N.</given-names></name> <name><surname>Ebrahimpour</surname> <given-names>R.</given-names></name></person-group> (<year>2017b</year>). <article-title>Hard-wired feed-forward visual mechanisms of the brain compensate for affine variations in object recognition.</article-title> <source><italic>Neuroscience</italic></source> <volume>349</volume> <fpage>48</fpage>&#x2013;<lpage>63</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuroscience.2017.02.050</pub-id> <pub-id pub-id-type="pmid">28245990</pub-id></citation></ref>
<ref id="B37"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Karimi-Rouzbahani</surname> <given-names>H.</given-names></name> <name><surname>Bagheri</surname> <given-names>N.</given-names></name> <name><surname>Ebrahimpour</surname> <given-names>R.</given-names></name></person-group> (<year>2017a</year>). <article-title>Average activity, but not variability, is the dominant factor in the representation of object categories in the brain.</article-title> <source><italic>Neuroscience</italic></source> <volume>346</volume> <fpage>14</fpage>&#x2013;<lpage>28</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuroscience.2017.01.002</pub-id> <pub-id pub-id-type="pmid">28088488</pub-id></citation></ref>
<ref id="B38"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Karimi-Rouzbahani</surname> <given-names>H.</given-names></name> <name><surname>Ramezani</surname> <given-names>F.</given-names></name> <name><surname>Woolgar</surname> <given-names>A.</given-names></name> <name><surname>Rich</surname> <given-names>A.</given-names></name> <name><surname>Ghodrati</surname> <given-names>M.</given-names></name></person-group> (<year>2021a</year>). <article-title>Perceptual difficulty modulates the direction of information flow in familiar face recognition.</article-title> <source><italic>Neuroimage</italic></source> <volume>233</volume>:<fpage>117896</fpage>. <pub-id pub-id-type="doi">10.1016/j.neuroimage.2021.117896</pub-id> <pub-id pub-id-type="pmid">33667671</pub-id></citation></ref>
<ref id="B39"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Karimi-Rouzbahani</surname> <given-names>H.</given-names></name> <name><surname>Shahmohammadi</surname> <given-names>M.</given-names></name> <name><surname>Vahab</surname> <given-names>E.</given-names></name> <name><surname>Setayeshi</surname> <given-names>S.</given-names></name> <name><surname>Carlson</surname> <given-names>T.</given-names></name></person-group> (<year>2021b</year>). <article-title>Temporal variabilities provide additional category-related information in object category decoding: a systematic comparison of informative EEG features.</article-title> <source><italic>Neural Comput.</italic></source> <volume>33</volume> <fpage>3027</fpage>&#x2013;<lpage>3072</lpage>. <pub-id pub-id-type="doi">10.1162/neco_a_01436</pub-id></citation></ref>
<ref id="B40"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Karimi-Rouzbahani</surname> <given-names>H.</given-names></name> <name><surname>Woolgar</surname> <given-names>A.</given-names></name> <name><surname>Rich</surname> <given-names>A. N.</given-names></name></person-group> (<year>2021c</year>). <article-title>Neural signatures of vigilance decrements predict behavioural errors before they occur.</article-title> <source><italic>ELife</italic></source> <volume>10</volume>:<fpage>e60563</fpage>. <pub-id pub-id-type="doi">10.7554/eLife.60563</pub-id> <pub-id pub-id-type="pmid">33830017</pub-id></citation></ref>
<ref id="B41"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Karimi-Rouzbahani</surname> <given-names>H.</given-names></name> <name><surname>Vahab</surname> <given-names>E.</given-names></name> <name><surname>Ebrahimpour</surname> <given-names>R.</given-names></name> <name><surname>Menhaj</surname> <given-names>M. B.</given-names></name></person-group> (<year>2019</year>). <article-title>Spatiotemporal analysis of category and target-related information processing in the brain during object detection.</article-title> <source><italic>Behav. Brain Res.</italic></source> <volume>362</volume> <fpage>224</fpage>&#x2013;<lpage>239</lpage>. <pub-id pub-id-type="doi">10.1016/j.bbr.2019.01.025</pub-id> <pub-id pub-id-type="pmid">30654124</pub-id></citation></ref>
<ref id="B42"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kayser</surname> <given-names>C.</given-names></name> <name><surname>Montemurro</surname> <given-names>M. A.</given-names></name> <name><surname>Logothetis</surname> <given-names>N. K.</given-names></name> <name><surname>Panzeri</surname> <given-names>S.</given-names></name></person-group> (<year>2009</year>). <article-title>Spike-phase coding boosts and stabilizes information carried by spatial and temporal spike patterns.</article-title> <source><italic>Neuron</italic></source> <volume>61</volume> <fpage>597</fpage>&#x2013;<lpage>608</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuron.2009.01.008</pub-id> <pub-id pub-id-type="pmid">19249279</pub-id></citation></ref>
<ref id="B43"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Langner</surname> <given-names>R.</given-names></name> <name><surname>Eickhoff</surname> <given-names>S. B.</given-names></name></person-group> (<year>2013</year>). <article-title>Sustaining attention to simple tasks: a meta-analytic review of the neural mechanisms of vigilant attention.</article-title> <source><italic>Psychol. Bull.</italic></source> <volume>139</volume>:<fpage>870</fpage>. <pub-id pub-id-type="doi">10.1037/a0030694</pub-id> <pub-id pub-id-type="pmid">23163491</pub-id></citation></ref>
<ref id="B44"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lee</surname> <given-names>M. D.</given-names></name> <name><surname>Wagenmakers</surname> <given-names>E. J.</given-names></name></person-group> (<year>2005</year>). <article-title>Bayesian statistical inference in psychology: comment on trafimow (2003).</article-title> <source><italic>Psychol. Rev.</italic></source> <volume>112</volume> <fpage>662</fpage>&#x2013;<lpage>668</lpage>. <pub-id pub-id-type="doi">10.1037/0033-295X.112.3.662</pub-id> <pub-id pub-id-type="pmid">16060758</pub-id></citation></ref>
<ref id="B45"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lempel</surname> <given-names>A.</given-names></name> <name><surname>Ziv</surname> <given-names>J.</given-names></name></person-group> (<year>1976</year>). <article-title>On the complexity of finite sequences.</article-title> <source><italic>IEEE Trans. Inform. Theor.</italic></source> <volume>22</volume> <fpage>75</fpage>&#x2013;<lpage>81</lpage>.</citation></ref>
<ref id="B46"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname> <given-names>H.</given-names></name> <name><surname>Agam</surname> <given-names>Y.</given-names></name> <name><surname>Madsen</surname> <given-names>J. R.</given-names></name> <name><surname>Kreiman</surname> <given-names>G.</given-names></name></person-group> (<year>2009</year>). <article-title>Timing, timing, timing: fast decoding of object information from intracranial field potentials in human visual cortex.</article-title> <source><italic>Neuron</italic></source> <volume>62</volume> <fpage>281</fpage>&#x2013;<lpage>290</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuron.2009.02.025</pub-id> <pub-id pub-id-type="pmid">19409272</pub-id></citation></ref>
<ref id="B47"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Majima</surname> <given-names>K.</given-names></name> <name><surname>Matsuo</surname> <given-names>T.</given-names></name> <name><surname>Kawasaki</surname> <given-names>K.</given-names></name> <name><surname>Kawai</surname> <given-names>K.</given-names></name> <name><surname>Saito</surname> <given-names>N.</given-names></name> <name><surname>Hasegawa</surname> <given-names>I.</given-names></name><etal/></person-group> (<year>2014</year>). <article-title>Decoding visual object categories from temporal correlations of ECoG signals.</article-title> <source><italic>Neuroimage</italic></source> <volume>90</volume> <fpage>74</fpage>&#x2013;<lpage>83</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuroimage.2013.12.020</pub-id> <pub-id pub-id-type="pmid">24361734</pub-id></citation></ref>
<ref id="B48"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mazaheri</surname> <given-names>A.</given-names></name> <name><surname>Jensen</surname> <given-names>O.</given-names></name></person-group> (<year>2008</year>). <article-title>Asymmetric amplitude modulations of brain oscillations generate slow evoked responses.</article-title> <source><italic>J. Neurosci.</italic></source> <volume>28</volume> <fpage>7781</fpage>&#x2013;<lpage>7787</lpage>. <pub-id pub-id-type="doi">10.1523/JNEUROSCI.1631-08.2008</pub-id> <pub-id pub-id-type="pmid">18667610</pub-id></citation></ref>
<ref id="B49"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Miyakawa</surname> <given-names>N.</given-names></name> <name><surname>Majima</surname> <given-names>K.</given-names></name> <name><surname>Sawahata</surname> <given-names>H.</given-names></name> <name><surname>Kawasaki</surname> <given-names>K.</given-names></name> <name><surname>Matsuo</surname> <given-names>T.</given-names></name> <name><surname>Kotake</surname> <given-names>N.</given-names></name><etal/></person-group> (<year>2018</year>). <article-title>Heterogeneous redistribution of facial subcategory information within and outside the face-selective domain in primate inferior temporal cortex.</article-title> <source><italic>Cereb. Cortex</italic></source> <volume>28</volume> <fpage>1416</fpage>&#x2013;<lpage>1431</lpage>. <pub-id pub-id-type="doi">10.1093/cercor/bhx342</pub-id> <pub-id pub-id-type="pmid">29329375</pub-id></citation></ref>
<ref id="B50"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Montemurro</surname> <given-names>M. A.</given-names></name> <name><surname>Rasch</surname> <given-names>M. J.</given-names></name> <name><surname>Murayama</surname> <given-names>Y.</given-names></name> <name><surname>Logothetis</surname> <given-names>N. K.</given-names></name> <name><surname>Panzeri</surname> <given-names>S.</given-names></name></person-group> (<year>2008</year>). <article-title>Phase-of-firing coding of natural visual stimuli in primary visual cortex.</article-title> <source><italic>Curr. Biol.</italic></source> <volume>18</volume> <fpage>375</fpage>&#x2013;<lpage>380</lpage>. <pub-id pub-id-type="doi">10.1016/j.cub.2008.02.023</pub-id> <pub-id pub-id-type="pmid">18328702</pub-id></citation></ref>
<ref id="B51"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Murphy</surname> <given-names>B.</given-names></name> <name><surname>Poesio</surname> <given-names>M.</given-names></name> <name><surname>Bovolo</surname> <given-names>F.</given-names></name> <name><surname>Bruzzone</surname> <given-names>L.</given-names></name> <name><surname>Dalponte</surname> <given-names>M.</given-names></name> <name><surname>Lakany</surname> <given-names>H.</given-names></name></person-group> (<year>2011</year>). <article-title>EEG decoding of semantic category reveals distributed representations for single concepts.</article-title> <source><italic>Brain Lang.</italic></source> <volume>117</volume> <fpage>12</fpage>&#x2013;<lpage>22</lpage>. <pub-id pub-id-type="doi">10.1016/j.bandl.2010.09.013</pub-id> <pub-id pub-id-type="pmid">21300399</pub-id></citation></ref>
<ref id="B52"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Muthukumaraswamy</surname> <given-names>S.</given-names></name></person-group> (<year>2013</year>). <article-title>High-frequency brain activity and muscle artifacts in MEG/EEG: a review and recommendations.</article-title> <source><italic>Front. Hum. Neurosci.</italic></source> <volume>7</volume>:<fpage>138</fpage>. <pub-id pub-id-type="doi">10.3389/fnhum.2013.00138</pub-id> <pub-id pub-id-type="pmid">23596409</pub-id></citation></ref>
<ref id="B53"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Namazi</surname> <given-names>H.</given-names></name> <name><surname>Ala</surname> <given-names>T. S.</given-names></name> <name><surname>Bakardjian</surname> <given-names>H.</given-names></name></person-group> (<year>2018</year>). <article-title>Decoding of steady-state visual evoked potentials by fractal analysis of the electroencephalographic (EEG) signal.</article-title> <source><italic>Fractals</italic></source> <volume>26</volume>:<fpage>1850092</fpage>.</citation></ref>
<ref id="B54"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ng</surname> <given-names>B. S. W.</given-names></name> <name><surname>Logothetis</surname> <given-names>N. K.</given-names></name> <name><surname>Kayser</surname> <given-names>C.</given-names></name></person-group> (<year>2013</year>). <article-title>EEG phase patterns reflect the selectivity of neural firing.</article-title> <source><italic>Cereb. Cortex</italic></source> <volume>23</volume> <fpage>389</fpage>&#x2013;<lpage>398</lpage>. <pub-id pub-id-type="doi">10.1093/cercor/bhs031</pub-id> <pub-id pub-id-type="pmid">22345353</pub-id></citation></ref>
<ref id="B55"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Orb&#x00E1;n</surname> <given-names>G.</given-names></name> <name><surname>Berkes</surname> <given-names>P.</given-names></name> <name><surname>Fiser</surname> <given-names>J.</given-names></name> <name><surname>Lengyel</surname> <given-names>M.</given-names></name></person-group> (<year>2016</year>). <article-title>Neural variability and sampling-based probabilistic representations in the visual cortex.</article-title> <source><italic>Neuron</italic></source> <volume>92</volume> <fpage>530</fpage>&#x2013;<lpage>543</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuron.2016.09.038</pub-id> <pub-id pub-id-type="pmid">27764674</pub-id></citation></ref>
<ref id="B56"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Panzeri</surname> <given-names>S.</given-names></name> <name><surname>Brunel</surname> <given-names>N.</given-names></name> <name><surname>Logothetis</surname> <given-names>N. K.</given-names></name> <name><surname>Kayser</surname> <given-names>C.</given-names></name></person-group> (<year>2010</year>). <article-title>Sensory neural codes using multiplexed temporal scales.</article-title> <source><italic>Trends Neurosci.</italic></source> <volume>33</volume> <fpage>111</fpage>&#x2013;<lpage>120</lpage>. <pub-id pub-id-type="doi">10.1016/j.tins.2009.12.001</pub-id> <pub-id pub-id-type="pmid">20045201</pub-id></citation></ref>
<ref id="B57"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Pouryazdian</surname> <given-names>S.</given-names></name> <name><surname>Erfanian</surname> <given-names>A.</given-names></name></person-group> (<year>2009</year>). &#x201C;<article-title>Detection of steady-state visual evoked potentials for brain-computer interfaces using PCA and high-order statistics</article-title>,&#x201D; in <source><italic>Proceddings of the World Congress on Medical Physics and Biomedical Engineering, September 7-12, 2009</italic></source> (<publisher-loc>Munich</publisher-loc>: <publisher-name>Springer</publisher-name>), <fpage>480</fpage>&#x2013;<lpage>483</lpage>.</citation></ref>
<ref id="B58"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Pulini</surname> <given-names>A. A.</given-names></name> <name><surname>Kerr</surname> <given-names>W. T.</given-names></name> <name><surname>Loo</surname> <given-names>S. K.</given-names></name> <name><surname>Lenartowicz</surname> <given-names>A.</given-names></name></person-group> (<year>2019</year>). <article-title>Classification accuracy of neuroimaging biomarkers in attention-deficit/hyperactivity disorder: effects of sample size and circular analysis.</article-title> <source><italic>Biol. Psychiatr. Cogn. Neurosci. Neuroimag.</italic></source> <volume>4</volume> <fpage>108</fpage>&#x2013;<lpage>120</lpage>. <pub-id pub-id-type="doi">10.1016/j.bpsc.2018.06.003</pub-id> <pub-id pub-id-type="pmid">30064848</pub-id></citation></ref>
<ref id="B59"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Qin</surname> <given-names>Y.</given-names></name> <name><surname>Zhan</surname> <given-names>Y.</given-names></name> <name><surname>Wang</surname> <given-names>C.</given-names></name> <name><surname>Zhang</surname> <given-names>J.</given-names></name> <name><surname>Yao</surname> <given-names>L.</given-names></name> <name><surname>Guo</surname> <given-names>X.</given-names></name><etal/></person-group> (<year>2016</year>). <article-title>Classifying four-category visual objects using multiple ERP components in single-trial ERP.</article-title> <source><italic>Cogn. Neurodynam.</italic></source> <volume>10</volume> <fpage>275</fpage>&#x2013;<lpage>285</lpage>. <pub-id pub-id-type="doi">10.1007/s11571-016-9378-0</pub-id> <pub-id pub-id-type="pmid">27468316</pub-id></citation></ref>
<ref id="B60"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Racine</surname> <given-names>R.</given-names></name></person-group> (<year>2011</year>). <source><italic>Estimating the Hurst Exponent.</italic></source> <publisher-loc>Zurich</publisher-loc>: <publisher-name>Mosaic Group</publisher-name>.</citation></ref>
<ref id="B61"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Richman</surname> <given-names>J. S.</given-names></name> <name><surname>Moorman</surname> <given-names>J. R.</given-names></name></person-group> (<year>2000</year>). <article-title>Physiological time-series analysis using approximate entropy and sample entropy.</article-title> <source><italic>Am. J. Physiol. Heart Circ. Physiol.</italic></source> <volume>278</volume> <fpage>H2039</fpage>&#x2013;<lpage>H2049</lpage>. <pub-id pub-id-type="doi">10.1152/ajpheart.2000.278.6.H2039</pub-id> <pub-id pub-id-type="pmid">10843903</pub-id></citation></ref>
<ref id="B62"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ritchie</surname> <given-names>J. B.</given-names></name> <name><surname>Tovar</surname> <given-names>D. A.</given-names></name> <name><surname>Carlson</surname> <given-names>T. A.</given-names></name></person-group> (<year>2015</year>). <article-title>Emerging object representations in the visual system predict reaction times for categorization.</article-title> <source><italic>PLoS Comput. Biol.</italic></source> <volume>11</volume>:<fpage>e1004316</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pcbi.1004316</pub-id> <pub-id pub-id-type="pmid">26107634</pub-id></citation></ref>
<ref id="B63"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Roffo</surname> <given-names>G.</given-names></name></person-group> (<year>2016</year>). <article-title>Feature selection library (MATLAB toolbox).</article-title> <source><italic>arXiv</italic></source> [<comment>Preprint</comment>]. <fpage>1607.01327</fpage></citation></ref>
<ref id="B64"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Roth</surname> <given-names>Z. N.</given-names></name> <name><surname>Ryoo</surname> <given-names>M.</given-names></name> <name><surname>Merriam</surname> <given-names>E. P.</given-names></name></person-group> (<year>2020</year>). <article-title>Task-related activity in human visual cortex.</article-title> <source><italic>PLoS Biol.</italic></source> <volume>18</volume>:<fpage>e3000921</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pbio.3000921</pub-id> <pub-id pub-id-type="pmid">33156829</pub-id></citation></ref>
<ref id="B65"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Rouder</surname> <given-names>J. N.</given-names></name> <name><surname>Morey</surname> <given-names>R. D.</given-names></name> <name><surname>Speckman</surname> <given-names>P. L.</given-names></name> <name><surname>Province</surname> <given-names>J. M.</given-names></name></person-group> (<year>2012</year>). <article-title>Default Bayes factors for ANOVA designs.</article-title> <source><italic>J. Math. Psychol.</italic></source> <volume>56</volume> <fpage>356</fpage>&#x2013;<lpage>374</lpage>.</citation></ref>
<ref id="B66"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Rupp</surname> <given-names>K.</given-names></name> <name><surname>Roos</surname> <given-names>M.</given-names></name> <name><surname>Milsap</surname> <given-names>G.</given-names></name> <name><surname>Caceres</surname> <given-names>C.</given-names></name> <name><surname>Ratto</surname> <given-names>C.</given-names></name> <name><surname>Chevillet</surname> <given-names>M.</given-names></name><etal/></person-group> (<year>2017</year>). <article-title>Semantic attributes are encoded in human electrocorticographic signals during visual object recognition.</article-title> <source><italic>Neuroimage</italic></source> <volume>148</volume> <fpage>318</fpage>&#x2013;<lpage>329</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuroimage.2016.12.074</pub-id> <pub-id pub-id-type="pmid">28088485</pub-id></citation></ref>
<ref id="B67"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Schaefer</surname> <given-names>A. T.</given-names></name> <name><surname>Angelo</surname> <given-names>K.</given-names></name> <name><surname>Spors</surname> <given-names>H.</given-names></name> <name><surname>Margrie</surname> <given-names>T. W.</given-names></name></person-group> (<year>2006</year>). <article-title>Neuronal oscillations enhance stimulus discrimination by ensuring action potential precision.</article-title> <source><italic>PLoS Biol.</italic></source> <volume>4</volume>:<fpage>e163</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pbio.0040163</pub-id> <pub-id pub-id-type="pmid">16689623</pub-id></citation></ref>
<ref id="B68"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Schroeder</surname> <given-names>C. E.</given-names></name> <name><surname>Lakatos</surname> <given-names>P.</given-names></name></person-group> (<year>2009</year>). <article-title>Low-frequency neuronal oscillations as instruments of sensory selection.</article-title> <source><italic>Trends Neurosci.</italic></source> <volume>32</volume> <fpage>9</fpage>&#x2013;<lpage>18</lpage>. <pub-id pub-id-type="doi">10.1016/j.tins.2008.09.012</pub-id> <pub-id pub-id-type="pmid">19012975</pub-id></citation></ref>
<ref id="B69"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Seeliger</surname> <given-names>K.</given-names></name> <name><surname>Fritsche</surname> <given-names>M.</given-names></name> <name><surname>G&#x00FC;&#x00E7;l&#x00FC;</surname> <given-names>U.</given-names></name> <name><surname>Schoenmakers</surname> <given-names>S.</given-names></name> <name><surname>Schoffelen</surname> <given-names>J. M.</given-names></name> <name><surname>Bosch</surname> <given-names>S. E.</given-names></name><etal/></person-group> (<year>2018</year>). <article-title>Convolutional neural network-based encoding and decoding of visual object recognition in space and time.</article-title> <source><italic>Neuroimage</italic></source> <volume>180</volume> <fpage>253</fpage>&#x2013;<lpage>266</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuroimage.2017.07.018</pub-id> <pub-id pub-id-type="pmid">28723578</pub-id></citation></ref>
<ref id="B70"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Shourie</surname> <given-names>N.</given-names></name> <name><surname>Firoozabadi</surname> <given-names>M.</given-names></name> <name><surname>Badie</surname> <given-names>K.</given-names></name></person-group> (<year>2014</year>). <article-title>Analysis of EEG signals related to artists and nonartists during visual perception, mental imagery, and rest using approximate entropy.</article-title> <source><italic>BioMed Res. Int.</italic></source> <volume>2014</volume>:<fpage>764382</fpage>. <pub-id pub-id-type="doi">10.1155/2014/764382</pub-id> <pub-id pub-id-type="pmid">25133180</pub-id></citation></ref>
<ref id="B71"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Simanova</surname> <given-names>I.</given-names></name> <name><surname>Van Gerven</surname> <given-names>M.</given-names></name> <name><surname>Oostenveld</surname> <given-names>R.</given-names></name> <name><surname>Hagoort</surname> <given-names>P.</given-names></name></person-group> (<year>2010</year>). <article-title>Identifying object categories from event-related EEG: toward decoding of conceptual representations.</article-title> <source><italic>PLoS One</italic></source> <volume>5</volume>:<fpage>e14465</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pone.0014465</pub-id> <pub-id pub-id-type="pmid">21209937</pub-id></citation></ref>
<ref id="B72"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Stewart</surname> <given-names>A. X.</given-names></name> <name><surname>Nuthmann</surname> <given-names>A.</given-names></name> <name><surname>Sanguinetti</surname> <given-names>G.</given-names></name></person-group> (<year>2014</year>). <article-title>Single-trial classification of EEG in a visual object task using ICA and machine learning.</article-title> <source><italic>J. Neurosci. methods</italic></source> <volume>228</volume> <fpage>1</fpage>&#x2013;<lpage>14</lpage>. <pub-id pub-id-type="doi">10.1016/j.jneumeth.2014.02.014</pub-id> <pub-id pub-id-type="pmid">24613798</pub-id></citation></ref>
<ref id="B73"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Storey</surname> <given-names>J. D.</given-names></name></person-group> (<year>2002</year>). <article-title>A direct approach to false discovery rates.</article-title> <source><italic>J. R. Stat. Soc.</italic></source> <volume>64</volume> <fpage>479</fpage>&#x2013;<lpage>498</lpage>.</citation></ref>
<ref id="B74"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Szczepa&#x0144;ski</surname> <given-names>J.</given-names></name> <name><surname>Amig&#x00F3;</surname> <given-names>J. M.</given-names></name> <name><surname>Wajnryb</surname> <given-names>E.</given-names></name> <name><surname>Sanchez-Vives</surname> <given-names>M. V.</given-names></name></person-group> (<year>2003</year>). <article-title>Application of Lempel&#x2013;Ziv complexity to the analysis of neural discharges.</article-title> <source><italic>Network</italic></source> <volume>14</volume>:<fpage>335</fpage>.</citation></ref>
<ref id="B75"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Taghizadeh-Sarabi</surname> <given-names>M.</given-names></name> <name><surname>Daliri</surname> <given-names>M. R.</given-names></name> <name><surname>Niksirat</surname> <given-names>K. S.</given-names></name></person-group> (<year>2015</year>). <article-title>Decoding objects of basic categories from electroencephalographic signals using wavelet transform and support vector machines.</article-title> <source><italic>Brain Topogr.</italic></source> <volume>28</volume> <fpage>33</fpage>&#x2013;<lpage>46</lpage>. <pub-id pub-id-type="doi">10.1007/s10548-014-0371-9</pub-id> <pub-id pub-id-type="pmid">24838816</pub-id></citation></ref>
<ref id="B76"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Torabi</surname> <given-names>A.</given-names></name> <name><surname>Jahromy</surname> <given-names>F. Z.</given-names></name> <name><surname>Daliri</surname> <given-names>M. R.</given-names></name></person-group> (<year>2017</year>). <article-title>Semantic category-based classification using nonlinear features and wavelet coefficients of brain signals.</article-title> <source><italic>Cogn. Comput.</italic></source> <volume>9</volume> <fpage>702</fpage>&#x2013;<lpage>711</lpage>.</citation></ref>
<ref id="B77"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tzovara</surname> <given-names>A.</given-names></name> <name><surname>Murray</surname> <given-names>M. M.</given-names></name> <name><surname>Plomp</surname> <given-names>G.</given-names></name> <name><surname>Herzog</surname> <given-names>M. H.</given-names></name> <name><surname>Michel</surname> <given-names>C. M.</given-names></name> <name><surname>De Lucia</surname> <given-names>M.</given-names></name></person-group> (<year>2012</year>). <article-title>Decoding stimulus-related information from single-trial EEG responses based on voltage topographies.</article-title> <source><italic>Pattern Recognit.</italic></source> <volume>45</volume> <fpage>2109</fpage>&#x2013;<lpage>2122</lpage>.</citation></ref>
<ref id="B78"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Victor</surname> <given-names>J. D.</given-names></name></person-group> (<year>2000</year>). <article-title>How the brain uses time to represent and process visual information.</article-title> <source><italic>Brain Res.</italic></source> <volume>886</volume> <fpage>33</fpage>&#x2013;<lpage>46</lpage>. <pub-id pub-id-type="doi">10.1016/s0006-8993(00)02751-7</pub-id></citation></ref>
<ref id="B79"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Vidaurre</surname> <given-names>D.</given-names></name> <name><surname>Myers</surname> <given-names>N. E.</given-names></name> <name><surname>Stokes</surname> <given-names>M.</given-names></name> <name><surname>Nobre</surname> <given-names>A. C.</given-names></name> <name><surname>Woolrich</surname> <given-names>M. W.</given-names></name></person-group> (<year>2019</year>). <article-title>Temporally unconstrained decoding reveals consistent but time-varying stages of stimulus processing.</article-title> <source><italic>Cereb. Cortex</italic></source> <volume>29</volume> <fpage>863</fpage>&#x2013;<lpage>874</lpage>. <pub-id pub-id-type="doi">10.1093/cercor/bhy290</pub-id> <pub-id pub-id-type="pmid">30535141</pub-id></citation></ref>
<ref id="B80"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Voloh</surname> <given-names>B.</given-names></name> <name><surname>Oemisch</surname> <given-names>M.</given-names></name> <name><surname>Womelsdorf</surname> <given-names>T.</given-names></name></person-group> (<year>2020</year>). <article-title>Phase of firing coding of learning variables across the fronto-striatal network during feature-based learning.</article-title> <source><italic>Nat. Commun.</italic></source> <volume>11</volume> <fpage>1</fpage>&#x2013;<lpage>16</lpage>. <pub-id pub-id-type="doi">10.1038/s41467-020-18435-3</pub-id> <pub-id pub-id-type="pmid">32938940</pub-id></citation></ref>
<ref id="B81"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wairagkar</surname> <given-names>M.</given-names></name> <name><surname>Zoulias</surname> <given-names>I.</given-names></name> <name><surname>Oguntosin</surname> <given-names>V.</given-names></name> <name><surname>Hayashi</surname> <given-names>Y.</given-names></name> <name><surname>Nasuto</surname> <given-names>S.</given-names></name></person-group> (<year>2016</year>). &#x201C;<article-title>Movement intention based Brain Computer Interface for Virtual Reality and Soft Robotics rehabilitation using novel autocorrelation analysis of EEG</article-title>,&#x201D; in <source><italic>Proceedings of the 2016 6th IEEE International Conference on Biomedical Robotics and Biomechatronics (BioRob)</italic></source> (<publisher-loc>Piscataway, NJ</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>685</fpage>&#x2013;<lpage>685</lpage>.</citation></ref>
<ref id="B82"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>C.</given-names></name> <name><surname>Xiong</surname> <given-names>S.</given-names></name> <name><surname>Hu</surname> <given-names>X.</given-names></name> <name><surname>Yao</surname> <given-names>L.</given-names></name> <name><surname>Zhang</surname> <given-names>J.</given-names></name></person-group> (<year>2012</year>). <article-title>Combining features from ERP components in single-trial EEG for discriminating four-category visual objects.</article-title> <source><italic>J. Neural Eng.</italic></source> <volume>9</volume>:<fpage>056013</fpage>. <pub-id pub-id-type="doi">10.1088/1741-2560/9/5/056013</pub-id></citation></ref>
<ref id="B83"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>Y.</given-names></name> <name><surname>Wang</surname> <given-names>P.</given-names></name> <name><surname>Yu</surname> <given-names>Y.</given-names></name></person-group> (<year>2018</year>). <article-title>Decoding english alphabet letters using EEG phase information.</article-title> <source><italic>Front. Neurosci.</italic></source> <volume>12</volume>:<fpage>62</fpage>. <pub-id pub-id-type="doi">10.3389/fnins.2018.00062</pub-id> <pub-id pub-id-type="pmid">29467615</pub-id></citation></ref>
<ref id="B84"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wark</surname> <given-names>B.</given-names></name> <name><surname>Fairhall</surname> <given-names>A.</given-names></name> <name><surname>Rieke</surname> <given-names>F.</given-names></name></person-group> (<year>2009</year>). <article-title>Timescales of inference in visual adaptation.</article-title> <source><italic>Neuron</italic></source> <volume>61</volume> <fpage>750</fpage>&#x2013;<lpage>761</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuron.2009.01.019</pub-id> <pub-id pub-id-type="pmid">19285471</pub-id></citation></ref>
<ref id="B85"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Waschke</surname> <given-names>L.</given-names></name> <name><surname>Kloosterman</surname> <given-names>N. A.</given-names></name> <name><surname>Obleser</surname> <given-names>J.</given-names></name> <name><surname>Garrett</surname> <given-names>D. D.</given-names></name></person-group> (<year>2021</year>). <article-title>Behavior needs neural variability.</article-title> <source><italic>Neuron</italic></source> <volume>109</volume> <fpage>751</fpage>&#x2013;<lpage>766</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuron.2021.01.023</pub-id> <pub-id pub-id-type="pmid">33596406</pub-id></citation></ref>
<ref id="B86"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Watrous</surname> <given-names>A. J.</given-names></name> <name><surname>Deuker</surname> <given-names>L.</given-names></name> <name><surname>Fell</surname> <given-names>J.</given-names></name> <name><surname>Axmacher</surname> <given-names>N.</given-names></name></person-group> (<year>2015</year>). <article-title>Phase-amplitude coupling supports phase coding in human ECoG.</article-title> <source><italic>Elife</italic></source> <volume>4</volume>:<fpage>e07886</fpage>.</citation></ref>
<ref id="B87"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Williams</surname> <given-names>M. A.</given-names></name> <name><surname>Dang</surname> <given-names>S.</given-names></name> <name><surname>Kanwisher</surname> <given-names>N. G.</given-names></name></person-group> (<year>2007</year>). <article-title>Only some spatial patterns of fMRI response are read out in task performance.</article-title> <source><italic>Nat. Neurosci.</italic></source> <volume>10</volume> <fpage>685</fpage>&#x2013;<lpage>686</lpage>. <pub-id pub-id-type="doi">10.1038/nn1900</pub-id> <pub-id pub-id-type="pmid">17486103</pub-id></citation></ref>
<ref id="B88"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wong</surname> <given-names>K. F. K.</given-names></name> <name><surname>Galka</surname> <given-names>A.</given-names></name> <name><surname>Yamashita</surname> <given-names>O.</given-names></name> <name><surname>Ozaki</surname> <given-names>T.</given-names></name></person-group> (<year>2006</year>). <article-title>Modelling non-stationary variance in EEG time series by state space GARCH model.</article-title> <source><italic>Comput. Biol. Med.</italic></source> <volume>36</volume> <fpage>1327</fpage>&#x2013;<lpage>1335</lpage>. <pub-id pub-id-type="doi">10.1016/j.compbiomed.2005.10.001</pub-id> <pub-id pub-id-type="pmid">16293239</pub-id></citation></ref>
<ref id="B89"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Woolgar</surname> <given-names>A.</given-names></name> <name><surname>Dermody</surname> <given-names>N.</given-names></name> <name><surname>Afshar</surname> <given-names>S.</given-names></name> <name><surname>Williams</surname> <given-names>M. A.</given-names></name> <name><surname>Rich</surname> <given-names>A. N.</given-names></name></person-group> (<year>2019</year>). <article-title>Meaningful patterns of information in the brain revealed through analysis of errors.</article-title> <source><italic>bioRxiv</italic></source> [<comment>Preprint</comment>].</citation></ref>
<ref id="B90"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zellner</surname> <given-names>A.</given-names></name> <name><surname>Siow</surname> <given-names>A.</given-names></name></person-group> (<year>1980</year>). <article-title>Posterior odds ratios for selected regression hypotheses.</article-title> <source><italic>Trabajos De Estad&#x00ED;stica y De Investigaci&#x00F3;n Operativa</italic></source> <volume>31</volume> <fpage>585</fpage>&#x2013;<lpage>603</lpage>.</citation></ref>
</ref-list>
<fn-group>
<fn id="footnote1">
<label>1</label>
<p><ext-link ext-link-type="uri" xlink:href="https://www.mathworks.com/matlabcentral/fileexchange/38211-calc_lz_complexity">https://www.mathworks.com/matlabcentral/fileexchange/38211-calc_lz_complexity</ext-link></p></fn>
<fn id="footnote2">
<label>2</label>
<p><ext-link ext-link-type="uri" xlink:href="https://ww2.mathworks.cn/matlabcentral/fileexchange/50290-higuchi-and-katz-fractal-dimension-measures">https://ww2.mathworks.cn/matlabcentral/fileexchange/50290-higuchi-and-katz-fractal-dimension-measures</ext-link></p></fn>
<fn id="footnote3">
<label>3</label>
<p><ext-link ext-link-type="uri" xlink:href="https://www.mathworks.com/matlabcentral/fileexchange/32427-fast-approximate-entropy">https://www.mathworks.com/matlabcentral/fileexchange/32427-fast-approximate-entropy</ext-link></p></fn>
<fn id="footnote4">
<label>4</label>
<p><ext-link ext-link-type="uri" xlink:href="https://klabhub.github.io/bayesFactor/">https://klabhub.github.io/bayesFactor/</ext-link></p></fn>
</fn-group>
</back>
</article>
