ORIGINAL RESEARCH article

Front. Hum. Neurosci., 01 May 2025

Sec. Sensory Neuroscience

Volume 19 - 2025 | https://doi.org/10.3389/fnhum.2025.1569282

This article is part of the Research TopicNeuro-Behavioral Insights on Low Vision and BeyondView all 5 articles

Video-evoked neuromarkers of visual function in age-related macular degeneration


Angela I. Renton
Angela I. Renton1*David J. KleinDavid J. Klein2Jesse A. LivezeyJesse A. Livezey2Dan NemrodovDan Nemrodov2Stephanie WolferStephanie Wolfer2Adam HaninaAdam Hanina2Dimitri Van De Ville,Dimitri Van De Ville1,3
  • 1Neuro-X Institute, School of Engineering, École polytechnique fédérale de Lausanne, Geneva, Switzerland
  • 2Dandelion Science Corp, Hoboken, NJ, United States
  • 3Department of Radiology and Medical Informatics, Faculty of Medicine, University of Geneva, Geneva, Switzerland

Neural markers of visual function in age-related macular degeneration (AMD) allow clinicians and researchers to directly evaluate the functional changes in visual processing which occur as a result of the progressive loss of afferent input from the macula. Unfortunately, few protocols exist that elicit such neural markers, and most of these are poorly adapted to AMD. Here, we propose a novel method of embedding frequency tags into full color and motion videos by periodically manipulating the contrast of visual information of different spatial frequencies at different temporal frequencies. These videos elicit steady-state visual evoked potentials (SSVEPS) in viewers which, when measured using electrophysiological neuroimaging methods, independently represent the responses of populations of neurons tuned to the tagged spatial frequencies. We used electroencephalography (EEG) to record the SSVEPs of 15 AMD patients and 16 age-matched healthy controls watching a 6-min series of natural scene videos filtered with this spatial frequency tagging method. Compared with healthy controls, AMD patients showed a lower SSVEP to high spatial frequency information, and a stronger response to the low spatial frequency information in the video set. The ratio of the SSVEP to lower relative to higher spatial frequency information was strongly predictive of both visual acuity and contrast sensitivity, and the topographic distributions of these responses suggested retinotopic reorganization of the neural response to spatial frequency information.

Introduction

Age related macular degeneration (AMD) is the most common cause of vision loss in older adults, affecting 8.7% of people aged 45–85 years globally (Stahl, 2020). This amounted to an estimated 196 million people suffering from AMD in 2020, with a projected increase to 288 million people by 2040 (Wong et al., 2014). This chronic disease affects the macular region of the retina and is characterized by progressive loss of central vision (Lim et al., 2012; Rattner and Nathans, 2006). AMD can severely impinge on patient's quality of life and independence by interfering with visual discrimination tasks such as reading (Varadaraj et al., 2018), face and emotion recognition (Boucart et al., 2008), object and scene recognition (Boucart et al., 2013; Thibaut et al., 2018), interaction with electronic devices (Taylor et al., 2016), and driving (Rovner and Casten, 2002; Wood et al., 2018). Clinicians and researchers who work to understand and treat AMD require accurate and comprehensive measures of AMD patients' changing visual function. Central vision loss is known to result in structural and functional changes in the brain's visual system, yet standard visual function tests rarely directly measure these changes (Baker et al., 2005; Broadhead et al., 2020; Chandramohan et al., 2016; Cheung and Legge, 2005; Ramanoël et al., 2018). In turn, researchers still have a limited understanding of what adaptive vs. maladaptive changes in neural activity look like in the face of central vision loss (Rosa et al., 2013). To address this gap, we developed and tested a novel neuroimaging-based visual function test for AMD which harnesses visual evoked neural responses to full-color natural scene videos.

The macular region of the retina is characterized by a high density of both light sensitive photoreceptor cells and the retinal ganglion cells they innervate, with the density of these photosensitive cells decreasing systematically toward the peripheral visual field. The macula thus senses changes in light and color on a fine scale, allowing the brain to resolve visual details such as the shapes of letters on a page (Ambati and Fowler, 2012). Spatial frequency, the rate at which information changes over space, is therefore differentially detectable across the visual field, with the highest spatial frequencies detectable in central vision (Drasdo et al., 2007; Hadjikhani and Tootell, 2000; Ramanoël et al., 2018). In turn, spatial frequency sensitivity in the brain's visual system is retinotopically mapped, such that neurons sensitive to higher spatial frequencies are more populous in regions of the cortex responsive to the central visual field, and neurons sensitive to lower spatial frequencies are more populous in regions of the cortex mapped to the periphery of the visual field (Kauffmann et al., 2014). The detection of contrasted edges at distinct spatial frequencies is one of the first steps in hierarchical visual processing; with this information feeding forward through the visual system to allow for more complex perceptual processing of visual information (Tovée, 1996). As such, the deafferentation of visual cortical tissue mapped to the macular region of the visual field should be expected to have a profound effect on neural activity in the brain's visual cortex.

There is substantial evidence suggesting that AMD patients undergo both structural and functional changes in vision. Several studies have shown that long-term AMD patients present with atrophy within cortical areas retinotopically mapped to central vision (Boucard et al., 2009; Prins et al., 2016) and cortical thickening in regions mapped to more peripheral regions of the visual field (Burge et al., 2016). The cortical thickening is suggestive of compensatory changes as visual perceptual processing adjusts to the loss of central visual input. Indeed, many functional studies in both animals and humans support this notion; finding evidence of retinotopic reorganization following central vision loss (see Cheung and Legge, 2005 for a review). These changes can be seen even in short term adaptation to scotopic vision, such that neurons whose receptive fields overlap with a central scotoma shift their receptive fields to encompass a larger area centered more peripherally in the visual field (Barton and Brewer, 2015). This adaptive reorganization has been shown to be more dramatic in long-term AMD patients, with visual cortical areas which map to central vision in healthy adults found to respond to peripheral visual stimulation in AMD patients (Baker et al., 2008, 2005; Dilks et al., 2014; Liu et al., 2010; Plank et al., 2021; Schumacher et al., 2008). Further evidence for functional reorganization comes from the tendency of AMD patients to develop a preferred retinal locus (PRL); i.e., a location in the peripheral retina adopted as pseudo-fovea by the oculomotor system, setting a reliable and automatic new location for fixation (Crossland et al., 2011; Maniglia et al., 2023, 2020; Rees et al., 2005). Psychophysical testing has revealed that visual perceptual function in the visual field around the PRL mimics that of the visual field around the fovea in healthy vision (Chen et al., 2019; Chung, 2014). Together, these results strongly suggest that the visual system undergoes functional reorganization to accommodate the loss of central visual information from the retina.

The functional changes in neural activity found for AMD patients can be either adaptive or maladaptive, either allowing patients to optimize their remaining visual function or engendering ancillary visual perceptual pathologies. As an example of an adaptive change, Shanidze and Verghese (2019) found that motion discrimination is well preserved in AMD patients. This is likely because motion sensitive areas of the visual cortex are typically innervated by the magnocellular pathway, which largely contains information from peripheral vision (Hadjikhani and Tootell, 2000). Crucially, the authors found a positive correlation between motion discrimination performance and the time since AMD diagnosis, suggesting that patients can adapt to improve their preserved visual function as their ability to discriminate central visual information wanes. By contrast, up to 40% of ocular pathology patients develop a condition called Charles Bonnet syndrome as a result of the loss of part of their visual field (Teunisse et al., 1996). Charles Bonnet patients, who have no comorbid psychiatric conditions, experience vivid long-term hallucinations ranging from simple geometric shapes, patterns, and flashing lights to complex hallucinations of animals, faces and even entire scenes (ffytche, 2009; Santhouse et al., 2000; Christoph et al., 2025). Painter et al., 2018 found compelling evidence to support a long-held hypothesis that Charles Bonnet syndrome in AMD can be attributed to cortical hyperexcitability, as Charles Bonnet AMD patients display strikingly elevated visual cortical responses to peripheral visual field stimulation compared with control AMD patients. These examples highlight the importance of visually evoked neural activity in forming a full picture of any AMD patient's visual function.

To date, several tests of visual function have been developed that directly measure visually evoked neural activity. For example, steady-state visual evoked potential (SSVEP) measures of visual acuity use electroencephalography (EEG) to measure evoked neural responses to a sweep of simple pattern reversing grating stimuli across a range of spatial frequencies. Sweep protocols aim to harness these SSVEPs at the grating flicker-frequency to measure the threshold highest granularity of visual information that a patients' visual cortex can resolve (Hamilton et al., 2021a). If a patient has lost the ability to resolve a grating of a particular high spatial frequency, the pattern-reversal flicker appears as a uniform gray surface and will not evoke an SSVEP. Unfortunately, while visual acuity scores derived with this SSVEP thresholding approach align well with objective behavioral measurements in general, this correlation is less strong for AMD patients (Hamilton et al., 2021b). Neuroimaging-based protocols that specifically aim to measure visual field losses rely on similar principles to sweep SSVEP protocols, typically presenting a dartboard-like checkerboard composed of concentric rings of black and white gratings. These methods measure neural response to pattern reversals in specific regions of the checkerboard stimulus (Bach, 2006; de Santiago et al., 2019; Horn et al., 2016; Liu et al., 2021). These techniques require patients to maintain fixation at the center of the display, to ensures that patients' visual fields line up correctly with the checkerboard stimulus. Unfortunately, this poses a significant barrier for AMD patients who struggle to maintain central fixation due to their central vision loss. Thus, current neuroimaging-based measures of visual function are relatively poorly adapted to AMD.

Currently available neuroimaging measures of visual function share two additional limitations in their applicability to AMD patients. The first is that they rely on detecting a threshold spatial frequency or location at which neural responses are no longer detectable. However, many visual impairments will lead to reduced, heightened or shifted (i.e., in time or location in the brain) neural responses rather than a simple absence of response. These more subtle shifts in neural activity are not measured by the standard application of currently available tests. The second limitation is that these protocols rely on simple monochrome visual stimuli, such as gratings and checkerboards. These high contrast stimuli optimally stimulate early visual cortical activity and maximize the signal-to-noise ratio or the measured neural response. However, these stimuli also have low ecological validity as test stimuli, and have been shown to be less effective than more complex visual stimuli in stimulating neural activity through purely extra-foveal stimulation (Nemrodov et al., 2024). Natural visual scenes contain color and motion information and have a complex hierarchical mathematical structure; i.e., spatial frequencies are grouped across time and space following scale-invariant fractal geometry, and exhibit a 1/f power law (Zetzsche, 2005). The human brain is optimized to this structure. Further, due to both the anatomical structure of the eye and the functional organization of the visual system, there are retinotopic spatial and temporal patterns to color, motion, and spatial frequency sensitivity which cannot be detected using monochrome patterns. For example, lower spatial frequency information, largely sensed in peripheral vision, is conveyed to the visual cortex through a faster cellular pathway than the higher spatial frequency information detected in central vision (Johnson and Johnson, 2014; Neri, 2014; Roberts et al., 2022). Low spatial frequency information, processed earlier, is used to extract scene context which is fed back down the visual hierarchy to allow predictive guidance in the processing of high spatial frequency information (Kauffmann et al., 2014). This fast and automatic feedforward/feedback process strongly impacts neural responses to visual information of distinct spatial frequencies but cannot be measured using simple monochrome pattern stimuli.

Here, we developed and benchmarked a novel neuroimaging-based visual function test which aims to address limitations in past neuroimaging measures of visual function in AMD. We propose a novel method to embed frequency tags within full-color and motion natural scene videos, such that the contrast of information at different spatial frequencies is periodically modulated at different temporal frequencies to elicit SSVEPs. Next, we propose a test of visual function in AMD which measures the changing sensitivity to spatial frequency information in these videos. This test relies on the retinotopic organization of spatial frequency sensitivity and would thus allow for the identification of central visual field deficits while participants freely and naturally shift their gaze across the display. We hypothesize that the novel video spatial frequency tagging method will evoke SSVEPs, allowing for concurrent but independent measurement of the population responses of neurons tuned to each of the tagged spatial frequencies. Further, we hypothesize that compared with healthy, age-matched controls, AMD patients should present a reduced SSVEP to higher spatial frequency information due to the loss visual sensitivity in central vision, but a larger response to lower spatial frequency information due to adaptive changes in visual perceptual processing. SSVEPs are particularly well suited to this method because of their resilience to oculomotor artifacts. Eye-movements and blinks minimally affect SSVEP amplitudes and these minimal effects are distributed across tagged frequencies, adding noise equally to all conditions. Thus, changes in subjects' gaze patterns cannot induce artifacts that would alter the relative power of SSVEPs to higher vs. lower spatial frequencies. To assess the hypotheses, we created a stimulus set in which we tagged two spatial frequency ranges: a higher range which should only be resolvable given the density of retinal ganglion cells within paracentral vision (<8°/visual angle within visual field, >3.2 c/d), and a lower range to which cortical regions retinotopically mapped to peripheral vision should be most sensitive (<3.2 c/d, Metha and Lennie, 2001). In total 15 patients diagnosed with bilateral dry AMD and 16 age-matched healthy controls viewed a 6-min series of these spatial frequency tagged videos while we recorded EEG data to allow for the measurement of SSVEPs. In analyzing these data, we found that the novel video tagging method elicited SSVEPs at the tagged frequencies. In turn, we showed that these SSVEPs were sensitive to the changes in visual processing associated with AMD. We therefore propose that video-evoked SSVEPs to spatial frequency information in natural scenes can be used as a neural marker of visual function in AMD.

Methods

Participants

N = 15 patients diagnosed with bilateral dry AMD (13 females, age M = 74.75 years, SD = 6.81) and 16 healthy older adults (14 females, age M = 61.3 years, SD = 3.3) volunteered to participate in the experiment after providing informed consent. Healthy controls were recruited through Institutional Review Board (IRB)-approved recruitment materials including flyers, online advertising on social media platforms, and business cards. AMD patients were referred and pre-screened by an ophthalmologist to the research team who completed the screening and enrolment process. AMD patients were included based on their age (at least 50 years old), a confirmed diagnosis of bilateral Dry AMD by their ophthalmologist, corrected visual acuity in both eyes between 20/30 and 20/100, Mini-Mental State Exam (MMSE) score of at least 25, and ability to provide written informed consent. Geographic atrophy (GA) was present in many, but not all of the AMD patients (see Figure 1). No data were collected on whether any of the AMD patients who participated in this study experienced Charles Bonnet Syndrome. Healthy controls were included based on age (at least 50 years old), best corrected visual acuity better than 20/25, and no history of any eye or optic nerve conditions. Participants were screened for psychiatric neurological conditions that could affect their vision or cognitive abilities, photosensitivity to flickering images and lights and history of epilepsy or seizures, no recent eye exam within the last 2 years, evidence of wet AMD or active neovascular leakage, any implanted electronic devices, history of cardiac problems, and any active infection or inflammation in the eyes. AMD participants were paid $88/h and healthy controls were paid $38/h for their participation. AMD patients performed this experiment as part of a larger battery of tests not presented here. The study was approved by the Biomedical Research Alliance of New York (BRANY) Ethics Committee, and the experiment was performed according to the relevant guidelines and regulations.

Figure 1
www.frontiersin.org

Figure 1. Optical coherence tomography (OCT) images of left and right eyes in example patients. Visual function scores for these patients can be found in Table 2.

Experiment design

All participants were instructed to freely view a 6-min series of frequency-tagged natural-scene videos while we recorded EEG data. Participants were not instructed to fixate at any position on the display. The video set consisted of 6 unique, 7.5 s videos (see Figure 2a for a sample frame from each video or visit https://osf.io/rp4q5/to view videos). For each of these videos, we frequency-tagged relatively higher and relatively lower spatial frequency information (>3.2 cyc/deg, <3.2 cyc/deg), manipulating the contrast of information within these spatial frequency ranges at different rates (7 Hz, 9 Hz) to induce SSVEPs. We choose 3.2 cyc/deg as the threshold spatial frequency to which reginal ganglion cells retinotopically mapped to regions of the visual field outside of paracentral vision (>8°/visual angle) should no longer be sensitive. Temporal frequency and spatial frequency (SF) were counterbalanced, such that each unique video was presented under two conditions (Condition 1: high SF−7 Hz, low SF−9 Hz | Cond 2: high SF: 9 Hz, low SF−7 Hz). The 6-min series was therefore composed of each of the 6 unique videos, presented under each of the 2 tagging conditions, repeated 4 times per condition for a total 48 video presentations. Videos were presented in random order with no breaks between videos.

Figure 2
www.frontiersin.org

Figure 2. Overview of visual stimulus design. (a) Sample frames of each of the 6, 7.5 s natural scene video clips used in the experiment. Participants were free to shift their gaze at will. (b) Illustrative steered 2D wavelet functions used for signal embedding. Each frame was decomposed into 5 spatial frequency bands at 8 different orientations to allow the contrast of these visual features to be independently manipulated. (c) By manipulating the power of different spatial frequencies and orientations in an image's steerable wavelet pyramid, it is possible to reconstruct the image with altered contrast in specific orientation and spatial frequency bands. For the example image shown here, this illustration shows the effect of applying this approach to filter out (d) relatively higher or lower spatial frequencies at all orientations, or (e) each of the two oblique orientations at all spatial frequencies. (f) To filter out orientation information, a filter is applied across orientations, smoothly down-regulating the power of filtered-out orientations. These are the filters used to generate the images shown in (e, g). An example video frame before and after filtering to embed the flicker signal. This example filtered frame highlights low spatial frequency information.

Frequency tag embedding

The novel spatial frequency video tagging method presented here relies on embedding temporal frequency tags into the spatial frequency structure of natural scene videos. To achieve this, we applied a steerable 2D wavelet pyramid method to alter the power with which different spatial frequencies and orientations were represented in each frame (Karasaridis and Simoncelli, 1996; Simoncelli et al., 1992; Simoncelli and Freeman, 1995). This method was used instead of a conventional 2D Fast-Fourier Transform (FFT), because it allows for segmentation across both orientation and spatial frequency; thus, creating the possibility to flexibly choose how different orientations are tagged. Indeed, this method could also be applied purely to tag oriented information, regardless of spatial frequency. Steerable wavelet transforms involve convolving an input signal with a “mother wavelet” function, which is scaled (stretched wider and narrower) to capture the input signal at different spatial frequencies, and rotated to respond to different orientations at a given spatial frequency (Van De Ville and Unser, 2008). Here, we used a Simoncelli isotropic wavelet function as the mother wavelet (Figure 2b) to build the steerable wavelet pyramid, which provides good estimates for early visual processing (Portilla and Simoncelli, 2000). One of the computational benefits of this method is that the orientation-spatial frequency space representation is perfectly invertible, and thus it is possible to adjust the power with which different spatial frequencies and orientations are represented across the image and then reconstruct meaningful images incorporating these changes. For example, consider the image in Figure 2c, which varies in both spatial frequency and orientation over space, for illustrative purposes. Using the 2-D steerable wavelet pyramid, it is possible to reconstruct the image after suppressing the power within specific spatial frequency (Figure 2d) or orientation (Figure 2e) channels. When the image is filtered for high spatial frequencies (Figure 2d, left), high spatial frequency regions and edges throughout the image are emphasized and smooth, low frequency areas are filtered (averaged into gray). In contrast, when the image is filtered for low spatial frequencies (Figure 2d, right), high spatial frequency regions turn into gray and only the larger spatial structures are preserved. Similarly, areas with oriented edges are differentially preserved or filtered depending on the orientation of the filter (Figure 2e). The steerable wavelet decomposition for this project was applied using the Steerable Wavelet Transform Toolbox in MATLAB (Püspöki and Unser, 2015; Unser and Chenouard, 2013). We recommend the Plenoptic toolbox for implementation in Python, which we have used successfully for the same purpose in other projects (Duong et al., 2024).

To apply this method, we converted each full-color video frame to hue, saturation, value (HSV) color space and extracted the value (luminance) channel (see example single frame in Figure 2f). We parsed this luminance information to a 2D steerable wavelet pyramid to extract the power within 5 spatial frequency bands for 8 orientations (0π-7π/8, Figure 2b). This allowed us to separate out spatial frequency information larger and smaller than the threshold (3.2 cyc/degree). For each of these spatial frequency ranges, we sought to specifically tag information at oblique (rather than cardinal orientations). In general, cardinally oriented information is more prevalent in natural scenes, and in turn the visual system is more sensitive to visual information at cardinal orientations (Tovée, 1996). Thus, applying the embedded tag to obliquely orientated information allowed us to probe neural responses to distinct spatial frequencies while minimizing the semantic distortion to the natural scene videos. This method could also be applied to tag cardinal orientations to elicit a stronger response. To achieve the tagging of oblique information, we applied a filter to scale the power across the 8 recovered orientations (for the relevant spatial frequency band) according to a phase of the embedded frequency tag (7 or 9 Hz sin wave). The power filter F was computed as shown in Equation 1:

F=(2A-1)×|sin2θ|    (1)

where A is the amplitude of embedded sinusoidal flicker signal, which oscillates between 0 and 1, and theta is the orientation of the tagged information. See Figure 2f for an illustration of this type of power filter. Effectively, this filter up and down regulated information at oblique orientations according to the phase of the flicker signal, while leaving information cardinal angles unmodulated over time. Note that the filter was applied to the full frame, and any natural scene video will, by definition, also vary in orientation and spatial frequency power over time and space. However, as these natural variations are stochastic in nature, averaging the neural response across successive video presentations allows for the exclusive measurement of signals time-locked to the embedded flicker signal (e.g., those evoked by the embedded changes in contrast of specific orientations and spatial frequencies). After applying the filters, we reconstructed the luminance information using the modulated spatial frequency pyramid. In turn, this new “Value” channel (HSV) was recombined with its original hue and saturation channels and converted back to RGB color space for presentation during the experiment. See Figure 2f for an example of one such modulated frame, noting that different orientations and spatial frequencies will be more relevant in different frames depending on the interactions of the two embedded frequency tags. Each raw video clip was subjected to this procedure twice, once to tag relatively higher spatial frequencies at 7 Hz and lower spatial frequencies at 9 Hz, and a second time to tag relatively higher spatial frequencies at 9 Hz and lower spatial frequencies at 7 Hz. See Supplementary material to view the spatial frequency tagged videos.

Behavioral visual function testing

All AMD patients and healthy controls underwent behavioral visual function testing to assess visual acuity (logMAR) and contrast sensitivity (logCS) using the Freiburg vision test (FrACT, Bach, 2024, 1996). Testing was performed from a viewing distance of 93 cm.

Other cognitive tasks

In addition to this experimental task, AMD patients in this study also performed a larger battery of cognitive tasks which were out of the scope of this research question. The study was performed over 3 consecutive days, with sessions lasting 1–2 h per day. Behavioral visual function tests were performed at the beginning of the first session and at the end of the last session. These included FrACT and MNRead (Calabrèse et al., 2016). The visual function scores reported for this experiment were those gathered using FrACT at the beginning of the first session. On each of the three days participants performed a lateral masking task requiring participants to identify low-contrast Gabor patches in crowded visual displays and, a letter identification task requiring participants to report the identity of letters which appeared for only 100 ms. These tasks are described in more detail by (Maniglia et al., 2020). Participants also performed a novel reading-speed task, in which they were asked to make speeded responses identifying the semantic grouping of single printed words. These words could be either “natural” (e.g., leaf, mouse, lake) or “man-made” (e.g., hammer, car, computer). The video-watching task which comprised this experiment was presented as the first task on the second day of the study. It should be noted that the lateral masking task which was presented on each day is a visual perceptual learning task with the aim to improve visual discrimination performance. However, the video-watching task was presented after only one experimental session, whereas training effects for this task have previously been reported after 24 sessions (Maniglia et al., 2020). Statistical analysis found no significant difference in visual function scores from the first to the third day of this study, thus any results should not be unduly biased by the inclusion of the lateral masking task in the experimental protocol.

Display computer specifications

All displays were presented at a viewing distance of 64 cm on Alienware 27 inch AW2721D monitor with a refresh rate of 120 Hz and resolution of 2,560 x 1,440. Stimuli were presented using custom software incorporating PsychoPy video presentation software (Peirce et al., 2019). The experiment was run on a Lenovo Thinkpad laptop and Lambda workstation.

EEG recording

EEG data were sampled at 1,000 Hz using a BioSemi Active Two amplifier (BioSemi, Amsterdam, Netherlands) from 68 active Ag/AgCl scalp electrodes arranged according to the international standard 10–20 system for electrode placement in a nylon head cap (Oostenveld and Praamstra, 2001). Four electrooculography (EOG) electrodes were used to record eye-movement muscle artifacts. The Common Mode Sense (CMS) and Driven Right Leg (DRL) electrodes were placed to the left and to the right of POz, respectively. Eye movements were recorded using the Tobii Pro Fusion 250 Hz eye-tracking system and synchronized with the EEG data using Lab Streaming Layer (LSL, Wang et al., 2023).

EEG analysis

Pre-processing

EEG data were pre-processed for offline analysis using the MNE-Python package (Gramfort et al., 2013). Noisy electrodes, identified via visual inspection by an expert, were replaced with cubic-spline interpolation based on the nearest channels. A tutorial outlining this process can be found within the documentation for the MNE MEG and EEG Analysis and Visualization toolbox. EEG data were average-referenced and bandpass filtered using a zero-phase bandpass filter from 1–100 Hz. Additionally, a notch filter at 60 Hz was applied to eliminate line noise. An independent component analysis (ICA) with 20 components was used to eliminate oculomotor artifacts, using the MNE-Python automated algorithms to identify and exclude components likely to be blink artifacts, eye-movement artifacts, or muscle noise.

SSVEP analysis

For each 7.5 s video, we computed a 7 s epoch beginning 0.5 s after video onset. For each of these epochs, EEG data across all channels were linearly detrended and baseline corrected. Videos on which the absolute amplitude at any occipitoparietal electrode (Iz, I1, I2, Oz, O1, O2, POz, PO3, PO4, PO7, PO8, Pz, P1, P2, P3, P4) exceeded 150 μV were excluded from further analysis. Using these epochs, we averaged across all videos for each of the two spatial frequency-flicker frequency conditions (lower SF 7 Hz, higher SF 9 Hz | lower SF 9 Hz, higher SF 7 Hz). Thus, we computed ERPs representing the neural activity synchronized to the embedded flickering signals, and averaged out spontaneous endogenous neural activity which changed in phase across successive video presentations. These ERPs were submitted to Fast Fourier Transforms (FFTs), and SSVEPs were taken as the average power across the two occipitoparietal electrodes that showed the strongest SSVEP for each flicker frequency (7 Hz, 9 Hz). As flicker frequency and spatial frequency were fully counterbalanced, this approach optimized SSVEP amplitudes equally for the relatively lower and higher spatial frequencies. Signal to noise ratio (SNR) was calculated by dividing the power at each frequency by the average of the four neighboring frequencies on either side (Meigen and Bach, 1999). Given the signal length of 7 s, this corresponded to 0.57 Hz. This was then converted to decibels as shown in Equation 2:

SNR (dB)=10log10(SNR)    (2)

Machine learning

Electrode sites used in machine learning analyses were: Iz, I1, I2, Oz, O1, O2, POz, PO3, PO4, PO7, PO8, Pz, P1, P2, P3, P4. Regression was implemented using Lasso (L1) regularized regression, with parameters estimated using the union of intersects (UoI) method (Sachdeva et al., 2021). UoI Lasso regression was implemented through the PyUoI package, using the default parameters with the exception of stability_selection (Bouchard et al., 2017; Sachdeva et al., 2021). The stability_selection parameter was set to 0.5 to reduce overfitting by encouraging the selection of more electrodes as features. We used 5-fold cross validation to estimate goodness of fit on unseen data. Splitting the data into 5 groups, we predicted visual acuity for the members of each group using a model trained on data from the 4 remaining groups. Goodness of fit was evaluated on these predictions using the coefficient of determination (R2), confidence intervals on R2 were calculated as per Cohen et al. (2013). 5-fold cross validation was fit using scikit-learn. The K-nearest neighbors (KNN), logistic regression (LR) and multi-layer perceptron classifiers were all fit using scikit-learn. KNN was fit using 5 nearest neighbors. The MLP was trained with two hidden layers (sizes: 10, 2) using Adam optimisation. Logistic regression was fit using L1 regularization and the liblinear solver.

Statistical tests

Statistical tests were conducted using the BayesFactor package in R (Package version 0.9.2+, R version 3.6.1, Rouder et al., 2012). Pairwise differences and differences between groups were assessed using the JZS t-test (Rouder et al., 2009). Bayes factors for main effects and interactions in Bayesian ANOVA models were assessed by comparing the full model (main effects + interaction + random effects) with the model containing all effects bar the effect of interest (Rouder et al., 2012). Bayes factors are reported with proportional error estimates unless the estimate of proportional error was <0.01%. Bayes factors are interpreted according to Jeffrey's criteria (Jarosz and Wiley, 2014), as follows:

Data and code availability

Python code used to implement data analysis and R code used to implement statistical analysis can be found at https://github.com/MIPLabCH/VENM-AMD. The raw data supporting these results are available upon request to the authors.

Results

Groupwise frequency tagging results

Here, we proposed a novel method of eliciting SSVEPs by periodically altering the contrast of distinct spatial frequencies in natural scene videos. Thus, we first sought to confirm that the videos tagged using this method elicited SSVEPs. As an initial step we investigated whether the video processing method had embedded the frequency tags as intended. To this end we submitted the processed videos to the 2D steerable wavelet decomposition, extracting the power with which different orientations and spatial frequencies were represented in each frame (Supplementary Figure 1a). As expected for natural scenes, we found that power peaked at the lowest spatial frequencies, with progressively higher spatial frequencies less strongly represented (1/f power distribution). Further, cardinal angles were more strongly represented than obliques (Supplementary Figure 1b). To interrogate the fidelity of the frequency tagging procedure we computed the average power for oblique angles at each spatial frequency in each video-frame and subjected this power to an FFT. This procedure showed that the tags were embedded as intended, with higher and lower spatial frequencies displaying peaks in the power spectra for their tagged temporal frequencies (Supplementary Figure 1c). As a point of interest, we performed the same procedure for cardinal angles and found that the tags were less strong, though still present (Supplementary Figure 1d). Thus, for future studies intending to use this procedure to investigate differences in orientation (rather than spatial frequency) representation, we recommend a sharper cutoff in the filter used to apply frequency tags across orientation than the sinusoidal taper applied here.

Once we had satisfied that spatial frequency information in the video set had been tagged as intended, we turned our attention to the EEG data. Using these data, we computed an ERP for each frequency-tagging condition (low SF 7 Hz, high SF 9 Hz | low SF 9 Hz, high SF 7 Hz), and subjected these ERPs to FFTs. Finally, we computed the SNR (dB) across the frequency spectrums output by the FFT. The grand-average frequency spectrums computed using this method revealed SSVEPs at the tagged frequencies (7 Hz, 9 Hz), for both AMD patients and controls (Figure 3a). We next visualized the topographic distribution of the grand-mean SSVEP (Db) across tagged frequencies (7 Hz, 9 Hz). In line with previous frequency tagging protocols, SSVEPs were found to peak at occipitoparietal electrode sites, indicating a visual cortical response (Norcia et al., 2015; Renton et al., 2022, Figure 3b). Thus, we concluded that the novel method of frequency tagging was effective in eliciting SSVEPs.

Figure 3
www.frontiersin.org

Figure 3. Overview of frequency tagging results. (a) Grand average frequency spectrum across all participants. To generate these spectrums, ERPs to video presentations for each of the two spatial frequency/flicker frequency conditions were submitted to FFTs. For each participant, spectrums were taken as the average across the occipitoparietal electrode sites where SSVEPs peaked for each frequency. Spectrums were then averaged across all participants tagging conditions for each group (AMD, controls). Flicker frequencies (7 Hz, 9 Hz) and their harmonics (14 Hz, 18 Hz) are marked by grey stars. The shaded areas mark 95% confidence intervals. (b) Topographical distribution of grand-mean SSVEPs (dB) across all tagged spatial frequencies. Topographies are shown for AMD patients, healthy controls, and the difference of mean SSVEPs for AMD patients—controls. All scalp topographies were visualized using the MNE-Python toolbox. (c) Violin plot showing SSVEPs (dB) for the relatively higher and lower spatial frequencies, for AMD patients and controls. SSVEPs were taken at two occipitoparietal electrode sites for which SSVEPs peaked for each participant. (d) Violin plot showing the mean SSVEPs across spatial frequencies, for AMD patients and controls. Statistical significance is illustrated by both the Bayes Factor, and a star system representing the p value computed using frequentist statistics: n.s. p > 0.05, *p < 0.05, **p < 0.01, ***p < 0.001.

Interestingly, we noted a topographical difference in the peak of the SSVEP between AMD patients and healthy controls; on average, the SSVEP for healthy controls peaked at the electrode sites Oz, O1, and O2. By contrast, the SSVEP for AMD patients was shifted anteriorly, toward the electrode sites POz, PO3, and PO4 (Figure 3b). To investigate this further, we interrogated the topographic distribution of the SSVEP separately for each spatial frequency (high SF, low SF). This revealed that the topographic difference was driven by the lower spatial frequency response; while both AMD patients and healthy controls SSVEPs r to high spatial frequencies peaked at the electrode site Oz, the lower spatial frequency response was shifted anteriorly to POz in AMD patients. This difference in response likely reflects the retinotopic organization of the visual system; AMD patients progressively lose the ability to resolve all information in central vision, and thus this low spatial frequency information must be detected more peripherally. Indeed, such anteriorly shifted topographies have been linked to peripheral visual stimulation in a previous study aiming to retinotopically map visually evoked responses in EEG (Capilla et al., 2016). By contrast, the higher spatial frequencies tagged in these video stimuli should only be possible to resolve within the macular region of the eye and not in the periphery. Thus, in line with these findings, we would expect the neural response to high spatial frequencies to be reduced at all electrode sites in AMD patients. These topographic differences represent compelling first evidence of the efficacy of video spatial frequency tagging as a neural marker for visual field loss in AMD.

We next sought to understand how, on average, AMD patients differed from healthy controls in their SSVEPs to the spatial-frequency tagged videos. For each participant, we computed an SSVEP amplitude for each spatial frequency by taking the average response across the two electrodes where the SSVEP peaked for each tagged spatial frequency and flicker frequency condition (low SF 7 Hz, high SF 9 Hz | low SF 9 Hz, high SF 7 Hz). We submitted these SSVEPs (dB) to a Bayesian ANOVA with spatial frequency (low SF, high SF) and group (AMD, control) as factors, and found decisive evidence for an interaction between these two factors (BF10 = 3005.81 ± 3.64%, Figure 3c). Note that the terminology used to describe the strength of evidence (e.g., decisive evidence) is derived Jeffrey's criteria (Jarosz and Wiley, 2014). These criteria are outlined in Table 1. Interestingly, there was only anecdotal evidence for an effect of group on SSVEP (dB), such that AMD patients (M = 9.61, SD = 4.29) did not differ significantly from controls (M = 11.40, SD = 3.99) in the overall magnitude of their neural response to the tagged videos (BF10 = 1.51 ± 23.87%, Figure 3d). This may be surprising, as AMD patients progressively lose the ability to resolve their central visual fields. In turn, less visual information innervates the early visual cortex, and one might expect a significantly lower amplitude neural response to the same visual information. Indeed, we found results in line with this supposition for SSVEPs to the higher spatial frequency visual information, which could only be resolved given the density of retinal ganglion cells typically found in paracentral vision. There was decisive evidence that SSVEPs (dB) to higher spatial frequency information in the tagged videos were weaker for AMD patients (M = 5.99, SD = 2.21) than for controls (M = 11.70, SD = 4.76, BF10 = 106.61, Figure 3c). By contrast, however, we found anecdotal evidence for the opposite effect for lower spatial frequency information. SSVEPs to lower spatial frequencies were stronger for AMD patients (M = 13.20, SD = 2.26) than for controls (M = 11.00, SD = 3.17, BF10 = 1.98 ± 0.01%, Figure 3c). This interaction between group (AMD, control) and spatial frequency response (low SF, high SF) may reflect a compensatory mechanism to the loss of central vision, such that AMD patients become more sensitive to lower spatial frequency information as they lose the ability to resolve higher spatial frequency information within the macular region.

Table 1
www.frontiersin.org

Table 1. Jeffrey's criteria for Bayes Factor inference.

Given that AMD patients differed from healthy controls in their responses to both higher and lower spatial frequencies, we sought to combine these effects into a single neural marker by computing the ratio of SSVEPs to lower relative to higher spatial frequencies. Initial visual inspection, confirmed by a Shapiro-Wilk's test for normality (W = 0.78, p < 0.001), suggested that these ratios were heavily right skewed. We therefore computed the log of the SSVEP ratios, resulting in normally distributed data as confirmed by a non-significant Shapiro-Wilk's test (W = 0.94, p = 0.073). To visualize the topographical distribution of the SSVEP ratios, we computed the log(low SF/high SF) SSVEP ratio for each electrode. Visual inspection suggested that this neural marker differentiated well between AMD patients and controls (Figure 4a): healthy controls showed a weak negative log(ratio) centered posteriorly on the scalp and peaking at electrode O2. By contrast, AMD patients showed a strong positive log(ratio) centered more anteriorly on the scalp and peaking around electrode PO4. The slight right-shifted asymmetry of these scalp topographies is a common finding in visual perceptual studies and likely reflects the underlying asymmetries in visual attention networks (Corbetta and Shulman, 2002). To confirm these observed differences, we computed the log(ratio) of the peak neural response to lower vs. higher spatial frequencies (Figure 4b). SSVEPs for each spatial frequency were computed at the electrode sites where they peaked. Note that this approach means that the high vs. low spatial frequency SSVEPs used to compute the log(Ratio) were derived from different electrode sites. The SSVEP values used to compute the log(ratio) were the SNR amplitudes before conversion to decibels. Using these ratios of lower to higher spatial frequency responses, we found decisive evidence that the log(ratio) was higher for AMD patients (M = 1.67, SD = 0.61) than controls (M = −0.14, SD = 1.23, BF10 = 940.47. Thus, the novel spatial frequency video tagging method allowed for the derivation of a neural marker which was sensitive to differences in visual processing between AMD patients and healthy controls.

Figure 4
www.frontiersin.org

Figure 4. Measures of visual function. (a) Topographical distribution and (b) violin plot of log(Ratio) of SSVEPs to lower relative to higher spatial frequencies for AMD patients vs. controls. (c) Violin plot showing visual acuity (logMAR) for AMD patients vs. controls. (d) Violin plot showing contrast sensitivity (logCS) for AMD patients vs. controls. ***p < 0.001.

Individual differences: video-evoked SSVEPs vs. visual function

To contextualize the SSVEP results, we measured visual function in both AMD patients and controls using standardized behavioral measures of visual acuity and contrast sensitivity. The visual acuity (logMAR), contrast sensitivity (logCS), and SSVEP metrics for each participant in the study are shown in Table 2. As expected, we found decisive evidence that visual acuity for AMD patients (logMAR, range: 0.07– 0.81, M = 0.31, SD = 0.20) was significantly poorer than for healthy controls (range:−0.22–0.23, M = 0.03, SD = 0.10, | BF10 = 450.28, Figure 4c). We also found very strong evidence that AMD patients had lower contrast sensitivity (logCS, range: 0.20 −1.94, M = 1.22, SD = 0.52) than healthy controls (range: 1.42–2.11, M = 1.74, SD = 2.11, | BF10 = 48.84, Figure 4d).

Table 2
www.frontiersin.org

Table 2. SSVEP and behavioral measures of visual function for each participant, sorted by visual acuity (logMAR).

Having established that the log(SSVEP Ratio) of lower relative to higher spatial frequencies could be used as a neural marker of AMD, we next sought to understand how well log(SSVEP Ratio) aligned with traditional behavioral measures of visual function. To this end, we fit a Bayesian general linear model to each of the behavioral visual function measures (logMAR, logCS) with group membership (AMD, control), and log(SSVEP Ratio) as predictors. For visual acuity, we found that data were best explained by both a linear effect of log(SSVEP Ratio) on logMAR (β = 0.54) and a main effect of group membership on logMAR (BF10 = 609.29 ± 0.60% vs. intersect only model, Figure 5a). There was anecdotal evidence for a null interaction between these terms, suggesting that the linear relationship between logMAR and log(SSVEP Ratio) did not change across groups (BF10 = 2.26 ± 1.07%). For contrast sensitivity, data were best explained only by an effect of group membership on logCS (BF10 = 48.84 vs. intersect only model, Figure 5b). We found anecdotal evidence that there was no additional linear relationship between log(SSVEP Ratio) and logCS (β = −0.41) beyond that explained by group (BF10 = 1.52 ± 0.64%).

Figure 5
www.frontiersin.org

Figure 5. Individual differences in log(SSVEP ratio) and relationship to behavioral visual function metrics. (a) log(SSVEP ratio) vs. visual acuity (logMAR) for AMD patients and healthy controls. (b) log(SSVEP ratio) vs. contrast sensitivity (logCS) for AMD patients and healthy controls. (c) Results of an UoI Lasso –regularized regression using group identity (AMD, Control) and the log(SSVEP Ratio) of low to high spatial frequencies at occipitoparietal electrode sites to predict visual acuity. (d) Results of an UoI Lasso –regularized regression using group identity (AMD, Control) and SSVEPs to high and low spatial frequency information at occipitoparietal electrode sites to predict visual acuity (logMAR). Predicted data in (c, d) were acquired through 5–fold cross validation. (e) EEG channel coefficients for the UoI Lasso regression shown in (c, d). Topoplots were generated using “nearest” interpolation. (f) Confusion matrices for the prediction of participant group (AMD, control) using behavioral visual function measures (left), SSVEPs to high and low spatial frequency information (center), and all of the above (right) using the best performing classification model (MLP).

The log(SSVEP Ratio) metric was calculated using the SSVEPs to lower and higher spatial frequency information in the video set at the electrode sites where these SSVEPs peaked for each person. It is therefore possible that there is an underlying topographic pattern in the SSVEPs which is more strongly predictive of behavioral visual function metrics. To investigate this possibility, we used Lasso (L1) regularized regression using the union of intersections (UoI) method for parameter inference (Sachdeva et al., 2021). Lasso regularization is known to be useful in feature selection, pushing the coefficients of less predictive features to zero. Here, we can apply this method to determine which electrode-sites were most predictive of visual acuity for higher and lower spatial frequency SSVEPs. However, application of the L1 regularization penalty is also known to bias non-zero weights to smaller values (“shrinking”) and often incorrectly identifies non-zero weights, reducing predictive accuracy (Sachdeva et al., 2021; Tibshirani, 1996). The UoI method aims to achieve both stable feature-selection and high predictive accuracy by utilizing ensemble model training and separating feature selection and estimation across two steps (Sachdeva et al., 2021). We therefore used the UoI method to fit a Lasso regression to predict visual acuity (logMAR) using group membership (AMD, Control) and SSVEP (dB) to lower and higher spatial frequencies across occipitoparietal electrode sites as predictors. Visual acuity was used as it was more strongly linearly predicted by log(SSVEP Ratio) than contrast sensitivity.

To establish baseline performance, we first fit a standard linear regression with group membership (AMD, Control) as the only predictor of visual acuity. Cross validated goodness of fit for this model was R2 = 0.33, 95% CI [0.08, 0.57]. By contrast, a UoI-Lasso model which included the log(SSVEP Ratio) at each occipitoparietal electrode-site as predictors provided a better and more reliable fit, R2 = 0.38, 95% CI [0.27, 0.49] (Figure 5c). However, the best fitting model was the UoI-Lasso model which included individual SSVEPs to low and high spatial frequencies as predictors, providing the best and most reliable prediction of visual acuity, R2 = 0.61, 95% CI [0.59, 0.63] (Figure 5d). These results show that not just the ratio, but also the topographic distribution of SSVEPs to high and low spatial frequencies was predictive of visual function. To estimate how these topographic patterns contributed to the prediction of visual acuity scores, we computed the mean of the coefficients across the 5 cross-validation folds. These average coefficients are plotted topographically in Figure 5e. The topographic plots of the coefficients show that in general, weaker responses to higher spatial frequencies and stronger responses to lower spatial frequencies across occipitoparietal electrode sites were predictive of worse visual acuity (Figure 5e). This is evident in the coefficients for the log(SSVEP Ratio) regression with the presence of a strong positive coefficient at electrode site POz. Interestingly, there is also a weak negative coefficient at Iz, showing that at this posterior electrode site, a stronger log(SSVEP Ratio) was actually predictive of better visual acuity. Coefficients for low-spatial frequency weights were largely positive and peaked at the electrode PO2. Thus, stronger SSVEPs to low spatial frequencies at this site were predictive of worse visual acuity. This large positive peak was flanked by electrodes with weaker negative coefficients (Iz, PO3, P2), suggesting that stronger low spatial frequency responses at these flanking locations were associated with better visual acuity. By contrast, high spatial frequency SSVEPs were associated with a strong negative coefficient at the electrode site POZ, and weaker positive coefficients at PO2, I1, and I2. This is a near, though not exact, inverse of the pattern found for low-spatial frequency responses. Together, these results are indicative of the retinotopic remapping previously reported for AMD patients; suggesting AMD patients experience a shift toward prioritized processing of lower spatial-frequency information and a topographic change in sensitivity to high and low spatial frequency information (Baker et al., 2008, 2005; Dilks et al., 2014).

These findings suggest that SSVEPs elicited by the spatial frequency tagged video set are predictive of visual function in AMD. However, while the neural and behavioral measures share variance, they are not completely aligned; i.e., SSVEP measures do not capture all the variance in behavioral measures, and vice versa. This misalignment suggests that SSVEP measures will likely contain additional information which may be relevant to the classification of AMD. To asses this possibility, we trained several machine learning models to discriminate between AMD patients and controls based on (1) their visual function scores (logMAR, logCS), (2) their spatial-frequency evoked SSVEPs (low SF, high SF) at central occipitoparietal electrode sites (Oz, O1, O2, POz, PO3, PO4), and (3) all of the above. We selected three machine learning models for this task: a logistic regression with L1 regularization as a linear model and K-nearest neighbors (KNN) and multi-layer perceptron (MLP) classifiers as non-linear models. The sensitivity, specificity, and accuracy of each of these models is shown in Table 3. Confusion plots for the MLP classifier, which performed best overall, are shown in Figure 5f. Overall, classification accuracy was better using the behavioral function scores (Mean performance: sensitivity: 83.00%, specificity = 72.33%, Accuracy = 76.33%) compared with SSVEPs (sensitivity: 71.67%, specificity = 70.67%, Accuracy = 71.00%) as predictors. However, the best result was for the combination of SSVEPs and visual acuity (sensitivity: 85.00%, specificity = 87.67%, Accuracy = 86.00%). Critically, we do not suggest that video-evoked SSVEPS should be used to diagnose AMD. Rather, these results highlight that the SSVEPs elicited by the spatial frequency tagging method provide additional information related to AMD, unique from that measured by behavioral visual function tests. This measure could therefore be gathered as part of forming a full picture of an individual patient's visual function.

Table 3
www.frontiersin.org

Table 3. Classification performance for the three classifiers, Logistic Regression (LR), K–nearest neighbors (KNN) and multi–layer perceptron (MLP).

Test optimisation

In this first evaluation of the novel video spatial frequency tagging method for probing spatial frequency sensitivity in AMD patients, the test parameters were determined based on first principles and early pilot results. We therefore sought to investigate the influence of video content and experiment duration on the results, with the aim of aiding decision-making for future research using this method.

The 6 videos used in this test were selected with the aim of spanning a diverse range of color, motion, perspective and subject matter. However, there likely exists a subset of video features that optimally elicit the SSVEP-ratio marker of visual function. While it is impossible to map this full feature-space with only 6 videos, we still sought to quantify how SSVEP metrics varied across this stimulus set. To this end, we computed the log(SSVEP Ratio) metric independently for each video in the set. It should be noted that this meaningfully increases the noise in the SSVEP metric, as per-video SSVEPs could only be calculated with a cell-size of 8 video repeats compared with the full 48 video repeats available when calculating SSVEPs for the full experiment. Interestingly, visual inspection suggested that the differences in log(SSVEP Ratio) between AMD patients and controls did vary by video ID (Figure 6a). To investigate this effect, log(SSVEP Ratio) data for each video were subjected to JZS t-tests to compare AMD patients vs. controls. The results, in order of significance were as follows: V3: BF10 = 4.84, V2: BF10 = 1.42 ± 0.01%, V6: BF10 = 1.19, V1: BF10 = 0.43, V5: BF10 = 0.37, V4: BF10 = 0.36. Thus, only V3 elicited a log(SSVEP Ratio) with substantial evidence for a difference between groups (threshold: BF10 > 3). Notably though, there was also no video which, when analyzed alone, elicited a log(SSVEP Ratio) with substantial evidence for a null difference between groups (threshold: BF10 < 0.33). A second variable of interest in assessing the efficacy of each video is the alignment with behavioral measures of visual function. As such, we computed the Pearson correlation between log(SSVEP Ratio) and each behavioral visual function metric (logMAR, logCS) for each video ID (Figure 6b). Again, we found variability in the strength of these relationships across videos. Similarly to the group difference results (AMD patients vs. controls), the correlation analysis showed that the strongest alignment with behavioral measures of visual function was elicited by videos V3, V2, and V6, while the weakest alignment was for videos V4, V5, and V1. Thus, the semantic content and visual feature set of videos used in eliciting SSVEP metrics is likely to be an important variable in test performance.

Figure 6
www.frontiersin.org

Figure 6. SSVEP metrics by video id and experiment duration. (a) Violin plot showing log(SSVEP ratio) between low and high spatial frequencies for AMD patients and controls, independently calculated for each video in the tagged video set. The videos are ordered on the axis from the smallest to largest mean difference between AMD patients and controls. (b) Pearson correlation between log(SSVEP ratio) and behavioral measures of visual function (logMAR, logCS). The video–order is the same as in (a). (c) log(SSVEP ratio) between low and high spatial frequencies by simulated experiment duration. Results are shown for AMD patients, healthy controls, and the mean difference between AMD patients and controls. (d) Pearson correlation between log(SSVEP ratio) and behavioral measures of visual function (logMAR, logCS) by simulated experiment duration. Shaded areas on (c, d) indicate 95% CI, computed by Monte–Carlo permutation.

Another important factor in future experiment design is the duration of the test. Neural responses to each probed spatial frequency range can be estimated more precisely with additional video presentation. Conflictingly, shorter diagnostic tests are valuable for both clinicians and researchers to maximize the number of distinct tests which can be performed and minimize demands on patients' time. As such, we performed a permutation test to estimate the variance on the mean of each of the test metrics for increasing experiment durations (15 s, 1 min, 2 min, 3 min, 4 min, 5 min, 6 min). For each experiment duration, over 30 permutations, we randomly sampled (with replacement) N/2 trials from each of the two frequency tagging conditions (low SF 7 Hz, high SF 9 Hz | low SF 9 Hz, high SF 7 Hz), where N is the number of 7.5 s trials needed to reach the targeted experiment duration. Thus, for example, in the 15 s condition, we sampled 1, 7.5 s trial from each frequency tagging condition, 30 times for each subject, to create 30 permuted samples in which each participants' SSVEPs where computed using only 2 trials.

Using these permuted data, we calculated the mean log(SSVEP Ratio) for each group (AMD, control, Figure 6c). Visual inspection suggested that the log(SSVEP Ratio) for the control group converged to a stable value within 1 min of video presentation. By contrast, the log(SSVEP Ratio) for the AMD group increased with each successive minute of video presentation. To estimate when the difference in log(SSVEP Ratio) between AMD patients and controls might converge to a stable value, we fit an inverse exponential function to this curve, as shown in Equation 3:

Δlog(Ratio)=a(1-e-s(T-i))    (3)

Where T is the number of training trials, a is the asymptote, s is the scaling factor and i is the x-axis intercept (Scolari et al., 2007). This model was found to provide a very good fit for the data (R2 = 0.99, Figure 6c). The inverse exponential model suggested that the mean log(SSVEP Ratio) difference between groups would converge to an asymptote of 0.99 after 10.5 min of presentation (experiment duration required to achieve 99% of asymptote value). This result suggests that a slightly longer presentation time than used here may have been optimal and should be used in future results.

Another relevant metric is that of alignment with behavioral measures of visual function. We therefore also correlated the permuted log(SSVEP Ratio) data with visual acuity (logMAR) and contrast sensitivity (logCS) data for each experiment duration. The inverse exponential model used for the log(SSVEP Ratio) difference between groups also provided a good fit for the changing r-values in this correlation as experiment duration increased (logMAR: R2 = 0.97, logCS: R2 = 0.94, Figure 6d). Thus, using the same model, we determined that the correlation between log(SSVEP Ratio) and visual acuity (logMAR) could be expected to converge to r = 0.46, within a 7-min experiment duration. For the correlation between log(SSVEP Ratio) and contrast sensitivity (logCS), the model suggested that an asymptote value of r = −0.37 could be reached within 5 min of experiment duration. As such, the relationship with contrast sensitivity is likely already optimally represented within the experiment duration presented here. However, effects for group differences and the relationship with visual acuity may be more strongly evident with the additional power of a longer experiment duration.

Discussion

Here, we developed and benchmarked a neuroimaging-based visual function test which harnesses full-color and motion videos of natural scenes. To facilitate this test, we proposed a novel technique of embedding frequency tags within video stimuli through periodic modulation of the contrast of higher and lower spatial frequencies. Neurons tuned to the tagged spatial frequencies should respond at the tagged flicker frequencies of 7 and 9 Hz, eliciting SSVEPs detectable at the scalp through EEG. Here, we demonstrated the efficacy of the novel video tagging method in eliciting SSVEPs. In turn, we showed that these SSVEPs were sensitive to the changes in visual processing associated with AMD and could be used as a neural marker of visual function in this patient group.

As hypothesized, we found that AMD patients differed from controls in their SSVEPs to both higher and lower spatial frequencies. The macular region of the retina is characterized by a high density of both photosensitive cone cells and the retinal ganglion cells they innervate. Retinal ganglion cells at the fovea may respond to only a single photoreceptor, as compared to thousands photoreceptors per retinal ganglion cell in peripheral vision (Watson, 2014). Thus, increasing eccentricity in the visual field corresponds with lower spatial frequency sensitivity in retina, a pattern reflected by the retinotopic organization of spatial frequency sensitivity in the visual cortices of healthy adults. As the macula region of the retina degenerates, AMD patients progressively lose the ability to resolve higher spatial frequencies, and thus should be expected to display a decreased neural response to this information. In turn, research suggests that the visual system adapts to the loss of central vision through structural and functional reorganization, reallocating processing resources toward lower spatial frequency information in peripheral vision (Cheung and Legge, 2005; Ramanoël et al., 2018). SSVEPs to the tagged video stimuli were in line with these predictions. Compared with healthy controls, AMD patients displayed a lower neural response to higher spatial frequencies across occipitoparietal electrode sites. In turn, they displayed a larger neural response to lower spatial frequencies. This lower spatial frequency response was topographically shifted toward the expected response pattern from cortical regions retinotopically mapped to peripheral regions of the visual field (Capilla et al., 2016). Thus, the SSVEPs elicited by this video set were sensitive to the functional changes in visual processing which occur in AMD.

We established that the ratio of SSVEPs to lower relative to higher spatial frequency information in the tagged video set acts as a neural marker of AMD. This marker differed significantly between AMD patients and age-matched healthy controls and could be used to predict visual acuity and contrast sensitivity in individual participants. This neural marker is likely attributable to two distinct effects in the visual cortex; the loss of afferent information from the central retina, and the neuroplastic changes in visual perceptual processing which occur in response to these lost inputs. The most profound of these adaptations is the development of a preferred retinal location (PRL), a “pseudo-fovea” in peripheral vision which is adopted when scotoma prevents central fixation. Researchers have found a large body of evidence to suggest that retinotopic remapping occurs toward the PRL. Functional magnetic resonance imaging (fMRI) data has shown that cortical regions which, in healthy adults, typically respond to visual stimulation at the fovea become responsive to visual stimulation at the PRL in long-term AMD patients (Baker et al., 2008, 2005; Dilks et al., 2014; Schumacher et al., 2008). Further, psychophysical testing has revealed changes in how visual stimuli interfere with the processing of other nearby stimuli in AMD patients (an effect termed “crowding”). Indeed, behavioral responses around the PRL begin to resemble those typically seen around the fovea (Chung, 2014, 2011). Further research has shown that visual cortical neurons whose receptive fields overlap with a central visual scotoma exhibit a shift in receptive field over time (Barton and Brewer, 2015). These receptive fields typically grow larger and shift outward from the scotoma, resulting in greater cortical sensitivity to lower spatial frequency visual information from peripheral vision as AMD progresses. Given the likely contribution these functional changes in visual processing to the log(SSVEP Ratio), it would be useful for future research to concretely map the relationship between SSVEPs, visual field loss, and time since AMD diagnosis.

The results presented here highlight the importance of neural markers of visual function. While behavioral visual function tests capture the outcome of the full process of visual perceptual processing and perceptual decision making, neural markers allow for exploration of subtle changes within specific stages of visual processing. Prior research suggests that the results of the behavioral visual function tests commonly administered for AMD patients often correspond poorly with AMD patients' reported everyday visual quality of life (Broadhead et al., 2020; Taylor et al., 2016). Here, we found that AMD could be best classified using both SSVEPs and behavioral visual function scores, suggesting that these two measures explore different sources of variance in visual function. It would be interesting to know whether this neural marker might also account for some of the unexplained variance in visual quality of life scores. Behavioral measures of visual acuity and contrast sensitivity represent the end result of a chain of neural events; visual information is sensed at the eye, undergoes hierarchical visual processing, is subject to perceptual decision-making processes which, finally, lead to a behavioral action (e.g., reporting a letter's identity). In later stages of processing, the brain is able to lean on cognitive skills such as selective attention, visual statistical learning, and predictive coding to enhance and pattern complete missing visual information. While these visual processing steps improve perceptual performance, they are also effortful and cannot be continuously applied to full capacity. As such, behavioral metrics can be considered to test patients' best possible visual function. By contrast, the SSVEP measure proposed here indexes neural responses at the first stages of visual processing; spatial frequency sensitive neurons are typically found in visual areas V1 and V2 (Foster et al., 1985). This measure can therefore be considered a baseline of which visual information is reaching the visual system before processing. Notably, while these initial results are promising, the proposed method will need to be standardized and validated with a larger cohort before it can be deployed as a formal diagnostic test.

Several open questions should be answered in the development of such a standardized protocol. Primarily, the feature set and semantic content of the test video set should be optimally selected. In this first study, we purposefully selected six natural scene video clips, aiming to span a diverse range of color, motion, perspective and subject matter. However, there likely exists a subset of video features that optimally elicit the SSVEP-ratio marker of visual function. To address these questions, we propose a neuroadaptive Bayesian optimization approach; by simulating scotoma in healthy participants, one could explore the feature space of videos which optimally elicit differences in log(SSVEP Ratio) between the scotoma vs. no-scotoma condition (Lorenz et al., 2017). An important point to note in any such experiment would be that while high spatial frequency responses should be immediately dampened by artificial scotoma, changes in the low-spatial frequency response and topographical distribution of responses are likely to largely reflect longer term neuroplastic changes. As such, results should always be validated with a patient population. The optimal test duration should also be established with this population using an optimized video test set; A model trained on the current dataset suggests that 10.5 min should be a sufficient duration to converge on the maximal power to differentiate between AMD patients and controls using the log(SSVEP Ratio). However, this should be confirmed using a longer test duration. Finally, the current results were obtained using a full 68-channel EEG cap. This setup is expensive and time-consuming to set up, making it inappropriate for clinical settings. However, examination of the coefficients in the machine learning regression analyses suggested that the electrode sites POz, PO2, PO3, Iz, I1, and I2 were most important in predicting behavioral measures of visual function and classifying AMD patients. A lower density EEG setup with this small subset of electrode sites would be more suitable for a clinical setting and should be sufficient for deriving the log(SSVEP Ratio) neural marker.

Here, we presented the novel video spatial frequency tagging method in the context of diagnostic applications in AMD. However, while neural markers of visual function are useful for diagnosis and patient monitoring, they also may be useful in therapeutic applications. For example, closed-loop neurofeedback protocols aim to shift neural activity toward a pattern of activity associated with a targeted cognitive state. Advances in compute power, experimental protocols, and analytical methods over the past 15 years have meaningfully shifted the horizon of viable cognitive objectives for neurofeedback interventions. Neurofeedback protocols are rapidly developing to allow more effective and specific cognitive enhancement, with compelling evidence of training effects in visual perceptual processing; including to improve visual selective attention (Bagherzadeh et al., 2020), sustained attention (deBettencourt et al., 2015), and perceptual confidence (Cortese et al., 2020) and visual perceptual learning (Amano et al., 2016; Shibata et al., 2011). The first step in developing any such neurofeedback protocol is to identify a neural marker of the targeted cognitive objective. Neurofeedback therapy represents an especially promising treatment avenue for patients suffering from Charles Bonnet Syndrome, a common condition in AMD patients leading to vivid visual hallucinations (Christoph et al., 2025; Teunisse et al., 1996). A leading hypothesis for the cause of these hallucinations is cortical hyperexcitability in the visual cortex (Bridge et al., 2024). Notably, chronic tinnitus has been attributed to the same cause in the auditory cortex (Chai et al., 2019; Langguth et al., 2007), and has been demonstrated to be treatable through neurofeedback (Gninenko et al., 2024). Previous research has shown that, due to this cortical hyperexcitability, Charles Bonnet AMD patients display profound differences in the strength of SSVEPs as compared to control AMD patients (Painter et al., 2018). Thus, SSVEPs elicited by the video spatial frequency tagging method will likely allow for a more nuanced exploration of this hyperexcitability and would provide a strong candidate for neurofeedback training aiming to regulate this maladaptive response.

Here, we presented a novel spatial frequency video tagging method to elicit neural markers of AMD, with applications in the diagnosis, monitoring, and therapeutic treatment of this patient group. This method is quick to administer, requires few electrodes, employs an ecologically valid test set of full color and motion stimuli with natural scene statistics, does not require patients to hold fixation, and allows for exploration of subtle differences in the amplitude and retinotopy of neural responses to distinct spatial frequencies. A limitation of this study is that even though all patients included experienced bilateral AMD, and thus presented with central scotoma, they still all had relatively good visual acuity. Indeed, by the World Health Organisation's International Classification of Diseases (ICD-11), corrected visual acuity scores for the patients included in this study could be categorized in a range from no vision impairment to moderate vision impairment. In future studies, it will be important to explore whether the pattern of results found here can be extrapolated as AMD progresses toward more severe vision impairment. For the patients included here, we found that evoked SSVEPS were sensitive to differences between AMD patients and age-matched controls with normal vision, aligned well with existing behavioral measures of visual function, and represented a source of variance in visual function not computed by these behavioral tests. Finally, we explored how test parameters might be altered to optimize the power of this test in scoring visual function and recommended several avenues for future research using this method.

Data availability statement

The example video files can be found on OSF: https://osf.io/rp4q5/.

Ethics statement

The studies involving humans were approved by Biomedical Research Alliance of New York (BRANY) Institutional Review Board. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.

Author contributions

DK: Funding acquisition, Investigation, Project administration, Writing – review & editing. JL: Investigation, Software, Writing – review & editing. DN: Investigation, Writing – review & editing. SW: Investigation, Writing – review & editing. AH: Funding acquisition, Project administration, Resources, Writing – review & editing. DV: Conceptualization, Funding acquisition, Methodology, Project administration, Supervision, Writing – review & editing. AR: Conceptualization, Data curation, Formal analysis, Funding acquisition, Methodology, Software, Validation, Visualization, Writing – original draft, Writing – review & editing.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. This work was funded by the American Macular Degeneration Foundation (AMDF) Breakthrough award and the Swiss Innovation Agency (Innosuisse, Project No. 116.587 IP-LS).

Conflict of interest

DK, JL, DN, AH, and SW are employees of Dandelion Science Corp. AR and DV have received financial support from Dandelion Science Corp in the form of research funding. All authors are listed on a pending patent describing this work.

Generative AI statement

The author(s) declare that no Gen AI was used in the creation of this manuscript.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fnhum.2025.1569282/full#supplementary-material

References

Amano, K., Shibata, K., Kawato, M., Sasaki, Y., and Watanabe, T. (2016). Learning to associate orientation with color in early visual areas by associative decoded fMRI neurofeedback. Curr. Biol. 26, 1861–1866. doi: 10.1016/j.cub.2016.05.014

PubMed Abstract | Crossref Full Text | Google Scholar

Ambati, J., and Fowler, B. J. (2012). Mechanisms of age-related macular degeneration. Neuron 75, 26–39. doi: 10.1016/j.neuron.2012.06.018

PubMed Abstract | Crossref Full Text | Google Scholar

Bach, M. (1996). The Freiburg visual acuity test—automatic measurement of visual acuity. Optom. Vis. Sci. Off. Publ. Am. Acad. Optom. 73, 49–53. doi: 10.1097/00006324-199601000-00008

PubMed Abstract | Crossref Full Text | Google Scholar

Bach, M. (2006). Latency of the mfVEP to diagnose glaucoma? Br. J. Ophthalmol. 90, 1076–1077. doi: 10.1136/bjo.2006.097626

PubMed Abstract | Crossref Full Text | Google Scholar

Bach, M. (2024). Freiburg vision test (FrACT): optimal number of trials? Graefes Arch. Clin. Exp. Ophthalmol. doi: 10.1007/s00417-024-06638-z

PubMed Abstract | Crossref Full Text | Google Scholar

Bagherzadeh, Y., Baldauf, D., Pantazis, D., and Desimone, R. (2020). Alpha synchrony and the neurofeedback control of spatial attention. Neuron 105, 577–587.e5. doi: 10.1016/j.neuron.2019.11.001

PubMed Abstract | Crossref Full Text | Google Scholar

Baker, C. I., Dilks, D. D., Peli, E., and Kanwisher, N. (2008). Reorganization of visual processing in macular degeneration: replication and clues about the role of foveal loss. Vision Res. 48, 1910–1919. doi: 10.1016/j.visres.2008.05.020

PubMed Abstract | Crossref Full Text | Google Scholar

Baker, C. I., Peli, E., Knouf, N., and Kanwisher, N. G. (2005). Reorganization of visual processing in macular degeneration. J. Neurosci. Off. J. Soc. Neurosci. 25, 614–618. doi: 10.1523/JNEUROSCI.3476-04.2005

PubMed Abstract | Crossref Full Text | Google Scholar

Barton, B., and Brewer, A. A. (2015). fMRI of the rod scotoma elucidates cortical rod pathways and implications for lesion measurements. Proc. Natl. Acad. Sci. 112, 5201–5206. doi: 10.1073/pnas.1423673112

PubMed Abstract | Crossref Full Text | Google Scholar

Boucard, C. C., Hernowo, A. T., Maguire, R. P., Jansonius, N. M., Roerdink, J. B. T. M., Hooymans, J. M. M., et al. (2009). Changes in cortical grey matter density associated with long-standing retinal visual field defects. Brain J. Neurol. 132, 1898–1906. doi: 10.1093/brain/awp119

PubMed Abstract | Crossref Full Text | Google Scholar

Boucart, M., Dinon, J.-F., Despretz, P., Desmettre, T., Hladiuk, K., and Oliva, A. (2008). Recognition of facial emotion in low vision: a flexible usage of facial features. Vis. Neurosci. 25, 603–609. doi: 10.1017/S0952523808080656

PubMed Abstract | Crossref Full Text | Google Scholar

Boucart, M., Moroni, C., Szaffarczyk, S., and Tran, T. H. C. (2013). Implicit processing of scene context in macular degeneration. invest. ophthalmol. Vis. Sci. 54, 1950–1957. doi: 10.1167/iovs.12-9680

PubMed Abstract | Crossref Full Text | Google Scholar

Bouchard, K., Bujan, A., Roosta, F., Ubaru, S., Prabhat, M., Snijders, A., et al. (2017). “Union of Intersections (UoI) for Interpretable Data Driven Discovery and Prediction,” in: Advances in Neural Information Processing Systems. (Newry, Curran Associates, Inc.).

Google Scholar

Bridge, H., Wyllie, A., Kay, A., Rand, B., Starling, L., Millington-Truby, R. S., et al. (2024). Neurochemistry and functional connectivity in the brain of people with Charles Bonnet syndrome. Ther. Adv. Ophthalmol. 16:25158414241280201. doi: 10.1177/25158414241280201

PubMed Abstract | Crossref Full Text | Google Scholar

Broadhead, G. K., Hong, T., Grigg, J. R., McCluskey, P., Schlub, T. E., Spooner, K., et al. (2020). Does functional assessment predict everyday visual functioning? Visual function testing and quality of life in mild/moderate age-related macular degeneration. Int. Ophthalmol. 40, 3241–3249. doi: 10.1007/s10792-020-01508-z

PubMed Abstract | Crossref Full Text | Google Scholar

Burge, W. K., Griffis, J. C., Nenert, R., Elkhetali, A., DeCarlo, D. K., ver Hoef, L. W., et al. (2016). Cortical thickness in human V1 associated with central vision loss. Sci. Rep. 6:23268. doi: 10.1038/srep23268

PubMed Abstract | Crossref Full Text | Google Scholar

Calabrèse, A., Owsley, C., McGwin, G., and Legge, G. E. (2016). Development of a reading accessibility index using the MNREAD acuity chart. JAMA Ophthalmol. 134, 398–405. doi: 10.1001/jamaophthalmol.2015.6097

PubMed Abstract | Crossref Full Text | Google Scholar

Capilla, A., Melcón, M., Kessel, D., Calderón, R., Pazo-Álvarez, P., and Carreti,é, L. (2016). Retinotopic mapping of visual event-related potentials. Biol. Psychol. 118, 114–125. doi: 10.1016/j.biopsycho.2016.05.009

PubMed Abstract | Crossref Full Text | Google Scholar

Chai, Z., Ma, C., and Jin, X. (2019). Cortical stimulation for treatment of neurological disorders of hyperexcitability: a role of homeostatic plasticity. Neural Regen. Res. 14:34. doi: 10.4103/1673-5374.243696

PubMed Abstract | Crossref Full Text | Google Scholar

Chandramohan, A., Stinnett, S. S., Petrowski, J. T., Schuman, S. G., Toth, C. A., Cousins, S. W., et al. (2016). Visual function measures in early and intermediate age-related macular degeneration. Retina 36:1021. doi: 10.1097/IAE.0000000000001002

PubMed Abstract | Crossref Full Text | Google Scholar

Chen, N., Shin, K., Millin, R., Song, Y., Kwon, M., and Tjan, B. S. (2019). Cortical reorganization of peripheral vision induced by simulated central vision loss. J. Neurosci. 39, 3529–3536. doi: 10.1523/JNEUROSCI.2126-18.2019

PubMed Abstract | Crossref Full Text | Google Scholar

Cheung, S.-H., and Legge, G. E. (2005). Functional and cortical adaptations to central vision loss. Vis. Neurosci. 22, 187–201. doi: 10.1017/S0952523805222071

PubMed Abstract | Crossref Full Text | Google Scholar

Christoph, S. E., Boden, K. T., Siegel, R., Seitz, B., Szurman, P., and Schulz, A. (2025). The prevalence of Charles-Bonnet syndrome in ophthalmic patients: A systematic review and meta-analysis. Brain Res. Bull. 111282. doi: 10.1016/j.brainresbull.2025.111282

PubMed Abstract | Crossref Full Text | Google Scholar

Chung, S. T. L. (2011). Improving reading speed for people with central vision loss through perceptual learning. invest. Ophthalmol. Vis. Sci. 52, 1164–1170. doi: 10.1167/iovs.10-6034

PubMed Abstract | Crossref Full Text | Google Scholar

Chung, S. T. L. (2014). Size or spacing: which limits letter recognition in people with age-related macular degeneration? Vision Res. 101, 167–176. doi: 10.1016/j.visres.2014.06.015

PubMed Abstract | Crossref Full Text | Google Scholar

Cohen, J., Cohen, P., West, S. G., and Aiken, L. S. (2013). Applied Multiple Regression/Correlation Analysis for the Behavioral Sciences. Abingdon: Routledge.

Google Scholar

Corbetta, M., and Shulman, G. L. (2002). Control of goal-directed and stimulus-driven attention in the brain. Nat. Rev. Neurosci. 3, 201–215. doi: 10.1038/nrn755

PubMed Abstract | Crossref Full Text | Google Scholar

Cortese, A., Lau, H., and Kawato, M. (2020). Unconscious reinforcement learning of hidden brain states supported by confidence. Nat. Commun. 11:4429. doi: 10.1038/s41467-020-17828-8

PubMed Abstract | Crossref Full Text | Google Scholar

Crossland, M. D., Engel, S. A., and Legge, G. E. (2011). The preferred retinal locus in macular disease: toward a consensus definition. Retina 31, 2109–2114. doi: 10.1097/IAE.0b013e31820d3fba

PubMed Abstract | Crossref Full Text | Google Scholar

de Santiago, L., Sánchez Morla, E. M., Ortiz, M., López, E., Amo Usanos, C., Alonso-Rodríguez, M. C., et al. (2019). A computer-aided diagnosis of multiple sclerosis based on mfVEP recordings. PLoS ONE 14:e0214662. doi: 10.1371/journal.pone.0214662

PubMed Abstract | Crossref Full Text | Google Scholar

deBettencourt, M. T., Cohen, J. D., Lee, R. F., Norman, K. A., and Turk-Browne, N. B. (2015). Closed-loop training of attention with real-time brain imaging. Nat. Neurosci. 18, 470–475. doi: 10.1038/nn.3940

PubMed Abstract | Crossref Full Text | Google Scholar

Dilks, D. D., Julian, J. B., Peli, E., and Kanwisher, N. (2014). Reorganization of visual processing in age-related macular degeneration depends on foveal loss. Optom. Vis. Sci. 91, e199–e206. doi: 10.1097/OPX.0000000000000325

PubMed Abstract | Crossref Full Text | Google Scholar

Drasdo, N., Millican, C. L., Katholi, C. R., and Curcio, C. A. (2007). The length of Henle fibers in the human retina and a model of ganglion receptive field density in the visual field. Vision Res. 47, 2901–2911. doi: 10.1016/j.visres.2007.01.007

PubMed Abstract | Crossref Full Text | Google Scholar

Duong, L., Bonnen, K., Broderick, W., Fiquet, P. É., Parthasarathy, N., Yerxa, T., et al. (2023). Plenoptic: a platform for synthesizing model-optimized visual stimuli. J. Vis. 23, 5822–5822. doi: 10.1167/jov.23.9.5822

Crossref Full Text | Google Scholar

ffytche, D.H. (2009). Visual hallucinations in eye disease. Curr. Opin. Neurol. 22, 28–35. doi: 10.1097/WCO.0b013e32831f1b3f

PubMed Abstract | Crossref Full Text | Google Scholar

Foster, K. H., Gaska, J. P., Nagler, M., and Pollen, D. A. (1985). Spatial and temporal frequency selectivity of neurones in visual cortical areas V1 and V2 of the macaque monkey. J. Physiol. 365, 331–363. doi: 10.1113/jphysiol.1985.sp015776

PubMed Abstract | Crossref Full Text | Google Scholar

Gninenko, N., Trznadel, S., Daskalou, D., Gramatica, L., Vanoy, J., Voruz, F., et al. (2024). Functional MRI neurofeedback outperforms cognitive behavioral therapy for reducing tinnitus distress: a prospective randomized clinical trial. Radiology 310:e231143. doi: 10.1148/radiol.231143

PubMed Abstract | Crossref Full Text | Google Scholar

Gramfort, A., Luessi, M., Larson, E., Engemann, D. A., Strohmeier, D., Brodbeck, C., et al. (2013). MEG and EEG data analysis with MNE-Python. Front. Neurosci. 7:267. doi: 10.3389/fnins.2013.00267

PubMed Abstract | Crossref Full Text | Google Scholar

Hadjikhani, N., and Tootell, R. B. H. (2000). Projection of rods and cones within human visual cortex. Hum. Brain Mapp. 9, 55–63. doi: 10.1002/(sici)1097-0193(2000)9:1<55::aid-hbm6>3.0.co;2-u

Crossref Full Text | Google Scholar

Hamilton, R., Bach, M., Heinrich, S. P., Hoffmann, M. B., Odom, J. V., McCulloch, D. L., et al. (2021a). ISCEV extended protocol for VEP methods of estimation of visual acuity. Doc. Ophthalmol. 142, 17–24. doi: 10.1007/s10633-020-09780-1

PubMed Abstract | Crossref Full Text | Google Scholar

Hamilton, R., Bach, M., Heinrich, S. P., Hoffmann, M. B., Odom, J. V., McCulloch, D. L., et al. (2021b). VEP estimation of visual acuity: a systematic review. Doc. Ophthalmol. 142, 25–74. doi: 10.1007/s10633-020-09770-3

PubMed Abstract | Crossref Full Text | Google Scholar

Horn, F. K., Selle, F., Hohberger, B., and Kremers, J. (2016). Steady-state multifocal visual evoked potential (ssmfVEP) using dartboard stimulation as a possible tool for objective visual field assessment. Graefes Arch. Clin. Exp. Ophthalmol. 254, 259–268. doi: 10.1007/s00417-015-3214-9

PubMed Abstract | Crossref Full Text | Google Scholar

Jarosz, A. F., and Wiley, J. (2014). What are the odds? A practical guide to computing and reporting Bayes factors. J. Probl. Solving 7:2. doi: 10.7771/1932-6246.1167

PubMed Abstract | Crossref Full Text | Google Scholar

Johnson, M. R., and Johnson, M. K. (2014). Decoding individual natural scene representations during perception and imagery. Front. Hum. Neurosci. 8:59. doi: 10.3389/fnhum.2014.00059

PubMed Abstract | Crossref Full Text | Google Scholar

Karasaridis, A., and Simoncelli, E. (1996). “A filter design technique for steerable pyramid image transforms,” in 1996 IEEE International Conference on Acoustics, Speech, and Signal Processing Conference Proceedings. Presented at the 1996 IEEE International Conference on Acoustics, Speech, and Signal Processing Conference Proceedings, pp. 2387–2390 vol. 4.

Google Scholar

Kauffmann, L., Ramanoël, S., and Peyrin, C. (2014). The neural bases of spatial frequency processing during scene perception. Front. Integr. Neurosci. 8:37. doi: 10.3389/fnint.2014.00037

PubMed Abstract | Crossref Full Text | Google Scholar

Langguth, B., Kleinjung, T., Marienhagen, J., Binder, H., Sand, P. G., Hajak, G., et al. (2007). Transcranial Magnetic Stimulation for the treatment of tinnitus: effects on cortical excitability. BMC Neurosci. 8:45. doi: 10.1186/1471-2202-8-45

PubMed Abstract | Crossref Full Text | Google Scholar

Lim, L. S., Mitchell, P., Seddon, J. M., Holz, F. G., and Wong, T. Y. (2012). Age-related macular degeneration. Lancet 379, 1728–1738. doi: 10.1016/S0140-6736(12)60282-7

PubMed Abstract | Crossref Full Text | Google Scholar

Liu, H., Liao, F., Blanco, R., and de la Villa, P. (2021). Multifocal visual evoked potentials (mfvep) for the detection of visual field defects in glaucoma: systematic review and meta-analysis. J. Clin. Med. 10:4165. doi: 10.3390/jcm10184165

PubMed Abstract | Crossref Full Text | Google Scholar

Liu, T., Cheung, S.-H., Schuchard, R. A., Glielmi, C. B., Hu, X., He, S., et al. (2010). Incomplete cortical reorganization in macular degeneration. invest. Ophthalmol. Vis. Sci. 51, 6826–6834. doi: 10.1167/iovs.09-4926

PubMed Abstract | Crossref Full Text | Google Scholar

Lorenz, R., Hampshire, A., and Leech, R. (2017). Neuroadaptive bayesian optimization and hypothesis testing. Trends Cogn. Sci. 21, 155–167. doi: 10.1016/j.tics.2017.01.006

PubMed Abstract | Crossref Full Text | Google Scholar

Maniglia, M., Soler, V., and Trotter, Y. (2020). Combining fixation and lateral masking training enhances perceptual learning effects in patients with macular degeneration. J. Vis. 20:19. doi: 10.1167/jov.20.10.19

PubMed Abstract | Crossref Full Text | Google Scholar

Maniglia, M., Visscher, K. M., and Seitz, A. R. (2023). Consistency of preferred retinal locus across tasks and participants trained with a simulated scotoma. Vision Res. 203:108158. doi: 10.1016/j.visres.2022.108158

PubMed Abstract | Crossref Full Text | Google Scholar

Meigen, T., and Bach, M. (1999). On the statistical significance of electrophysiological steady-state responses. Doc. Ophthalmol. 98, 207–232. doi: 10.1023/A:1002097208337

PubMed Abstract | Crossref Full Text | Google Scholar

Metha, A., and Lennie, P. (2001). Transmission of spatial information in S-cone pathways. Vis. Neurosci. 18, 961–72. doi: 10.1017/S095252380118613X

PubMed Abstract | Crossref Full Text | Google Scholar

Nemrodov, D., Klein, D., Livezey, J., Wolfer, S., Kupersmith, M. J., Kondapalli, S., et al. (2024). Generative video stimulation enhances neural-based visual performance biomarkers in patients with dry age-related macular degeneration. Invest. Ophthalmol. Vis. Sci. 65:5434.

Google Scholar

Neri, P. (2014). Semantic control of feature extraction from natural scenes. J. Neurosci. 34, 2374–2388. doi: 10.1523/JNEUROSCI.1755-13.2014

PubMed Abstract | Crossref Full Text | Google Scholar

Norcia, A. M., Appelbaum, L. G., Ales, J. M., Cottereau, B. R., and Rossion, B. (2015). The steady-state visual evoked potential in vision research: a review. J. Vis. 15, 4–4. doi: 10.1167/15.6.4

PubMed Abstract | Crossref Full Text | Google Scholar

Oostenveld, R., and Praamstra, P. (2001). The five percent electrode system for high-resolution EEG and ERP measurements. Clin. Neurophysiol. 112, 713–719. doi: 10.1016/S1388-2457(00)00527-7

PubMed Abstract | Crossref Full Text | Google Scholar

Painter, D. R., Dwyer, M. F., Kamke, M. R., and Mattingley, J. B. (2018). Stimulus-driven cortical hyperexcitability in individuals with charles bonnet hallucinations. Curr. Biol. 28, 3475–3480.e3. doi: 10.1016/j.cub.2018.08.058

PubMed Abstract | Crossref Full Text | Google Scholar

Peirce, J., Gray, J. R., Simpson, S., MacAskill, M., Höchenberger, R., Sogo, H., et al. (2019). PsychoPy2: experiments in behavior made easy. Behav. Res. Methods 51, 195–203. doi: 10.3758/s13428-018-01193-y

PubMed Abstract | Crossref Full Text | Google Scholar

Plank, T., Benkowitsch, E. M. A., Beer, A. L., Brandl, S., Malania, M., Frank, S. M., et al. (2021). Cortical thickness related to compensatory viewing strategies in patients with macular degeneration. Front. Neurosci. 15:718737. doi: 10.3389/fnins.2021.718737

PubMed Abstract | Crossref Full Text | Google Scholar

Portilla, J., and Simoncelli, E. P. (2000). A parametric texture model based on joint statistics of complex wavelet coefficients. Int. J. Comput. Vis. 40, 49–70. doi: 10.1023/A:1026553619983

Crossref Full Text | Google Scholar

Prins, D., Plank, T., Baseler, H. A., Gouws, A. D., Beer, A., Morland, A. B., et al. (2016). Surface-based analyses of anatomical properties of the visual cortex in macular degeneration. PLoS ONE 11:e0146684. doi: 10.1371/journal.pone.0146684

PubMed Abstract | Crossref Full Text | Google Scholar

Püspöki, Z., and Unser, M. (2015). Template-free wavelet-based detection of local symmetries. IEEE Trans. Image Process. 24, 3009–3018. doi: 10.1109/TIP.2015.2436343

PubMed Abstract | Crossref Full Text | Google Scholar

Ramanoël, S., Chokron, S., Hera, R., Kauffmann, L., Chiquet, C., Krainik, A., et al. (2018). Age-related macular degeneration changes the processing of visual scenes in the brain. Vis. Neurosci. 35:E006. doi: 10.1017/S0952523817000372

PubMed Abstract | Crossref Full Text | Google Scholar

Rattner, A., and Nathans, J. (2006). Macular degeneration: recent advances and therapeutic opportunities. Nat. Rev. Neurosci. 7, 860–872. doi: 10.1038/nrn2007

PubMed Abstract | Crossref Full Text | Google Scholar

Rees, A. L., Kabanarou, S. A., Culham, L. E., and Rubin, G. S. (2005). Can retinal eccentricity predict visual acuity and contrast sensitivity at the PRL in AMD patients? Int. Congr. Ser. Vision 1282, 694–698. doi: 10.1016/j.ics.2005.05.172

Crossref Full Text | Google Scholar

Renton, A. I., Painter, D. R., and Mattingley, J. B. (2022). Optimising the classification of feature-based attention in frequency-tagged electroencephalography data. Sci. Data. 9:296. doi: 10.1038/s41597-022-01398-z

PubMed Abstract | Crossref Full Text | Google Scholar

Roberts, M. M., Schira, M. M., Spehar, B., and Isherwood, Z. J. (2022). Nature in motion: the tuning of the visual system to the spatiotemporal properties of natural scenes. J. Vis. 22:7. doi: 10.1167/jov.22.6.7

PubMed Abstract | Crossref Full Text | Google Scholar

Rosa, A. M., Silva, M. F., Ferreira, S., Murta, J., and Castelo-Branco, M. (2013). Plasticity in the human visual cortex: an ophthalmology-based perspective. BioMed Res. Int. 2013:568354. doi: 10.1155/2013/568354

PubMed Abstract | Crossref Full Text | Google Scholar

Rouder, J. N., Morey, R. D., Speckman, P. L., and Province, J. M. (2012). Default bayes factors for ANOVA designs. J. Math. Psychol. 56, 356–374. doi: 10.1016/j.jmp.2012.08.001

Crossref Full Text | Google Scholar

Rouder, J. N., Speckman, P. L., Sun, D., Morey, R. D., and Iverson, G. (2009). Bayesian t tests for accepting and rejecting the null hypothesis. Psychon. Bull. Rev. 16, 225–237. doi: 10.3758/PBR.16.2.225

PubMed Abstract | Crossref Full Text | Google Scholar

Rovner, B. W., and Casten, R. J. (2002). Activity loss and depression in age-related macular degeneration. Am. J. Geriatr. Psychiatry 10, 305–310. doi: 10.1097/00019442-200205000-00010

Crossref Full Text | Google Scholar

Sachdeva, P. S., Livezey, J. A., Dougherty, M. E., Gu, B.-M., Berke, J. D., and Bouchard, K. E. (2021). Improved inference in coupling, encoding, and decoding models and its consequence for neuroscientific interpretation. J. Neurosci. Methods 358:109195. doi: 10.1016/j.jneumeth.2021.109195

PubMed Abstract | Crossref Full Text | Google Scholar

Santhouse, A. M., Howard, R. J., and ffytche, D.H. (2000). Visual hallucinatory syndromes and the anatomy of the visual brain. Brain 123, 2055–2064. doi: 10.1093/brain/123.10.2055

PubMed Abstract | Crossref Full Text | Google Scholar

Schumacher, E. H., Jacko, J. A., Primo, S. A., Main, K. L., Moloney, K. P., Kinzel, E. N., et al. (2008). Reorganization of visual processing is related to eccentric viewing in patients with macular degeneration. Restor. Neurol. Neurosci. 26, 391–402. doi: 10.3233/RNN-2008-00423

PubMed Abstract | Crossref Full Text | Google Scholar

Scolari, M., Kohnen, A., Barton, B., and Awh, E. (2007). Spatial attention, preview, and popout: Which factors influence critical spacing in crowded displays? J. Vis. 7, 1–23. doi: 10.1167/7.2.7

PubMed Abstract | Crossref Full Text | Google Scholar

Shanidze, N., and Verghese, P. (2019). Motion perception in central field loss. J. Vis. 19:20. doi: 10.1167/19.14.20

PubMed Abstract | Crossref Full Text | Google Scholar

Shibata, K., Watanabe, T., Sasaki, Y., and Kawato, M. (2011). Perceptual learning incepted by decoded fmri neurofeedback without stimulus presentation. Science 334, 1413–1415. doi: 10.1126/science.1212003

PubMed Abstract | Crossref Full Text | Google Scholar

Simoncelli, E. P., and Freeman, W. T. (1995). The steerable pyramid: a flexible architecture for multi-scale derivative computation, in Proceedings., International Conference on Image Processing. Presented at the Proceedings. International Conference on Image Processing, pp. 444–447 vol.3.

Google Scholar

Simoncelli, E. P., Freeman, W. T., Adelson, E. H., and Heeger, D. J. (1992). Shiftable multiscale transforms. IEEE Trans. Inf. Theory 38, 587–607. doi: 10.1109/18.119725

Crossref Full Text | Google Scholar

Stahl, A. (2020). The diagnosis and treatment of age-related macular degeneration. Dtsch. Ärztebl. Int. 117, 513–520. doi: 10.3238/arztebl.2020.0513

PubMed Abstract | Crossref Full Text | Google Scholar

Taylor, D. J., Hobby, A. E., Binns, A. M., and Crabb, D. P. (2016). How does age-related macular degeneration affect real-world visual ability and quality of life? A systematic review. BMJ Open 6:e011504. doi: 10.1136/bmjopen-2016-011504

PubMed Abstract | Crossref Full Text | Google Scholar

Teunisse, R. J., Zitman, F. G., Cruysberg, J. R. M., Hoefnagels, W. H. L., and Verbeek, A. L. M. (1996). Visual hallucinations in psychologically normal people: Charles Bonnet's syndrome. Lancet 347, 794–797. doi: 10.1016/S0140-6736(96)90869-7

PubMed Abstract | Crossref Full Text | Google Scholar

Thibaut, M., Tran, T., Szaffarczyk, S., and Boucart, M. (2018). Impact of age-related macular degeneration on object searches in realistic panoramic scenes. Clin. Exp. Optom. 101, 372–379. doi: 10.1111/cxo.12644

PubMed Abstract | Crossref Full Text | Google Scholar

Tibshirani, R. (1996). Regression shrinkage and selection via the lasso. J. R. Stat. Soc. Ser. B Methodol. 58, 267–288. doi: 10.1111/j.2517-6161.1996.tb02080.x

Crossref Full Text | Google Scholar

Tovée, M. J. (1996). An Introduction to the Visual System.. Cambridge: Cambridge University Press.

Google Scholar

Unser, M., and Chenouard, N. (2013). A unifying parametric framework for 2D steerable wavelet transforms. SIAM J. Imaging Sci. 6, 102–135. doi: 10.1137/120866014

Crossref Full Text | Google Scholar

Van De Ville, D., and Unser, M. (2008). Complex wavelet bases, steerability, and the marr-like pyramid. IEEE Trans. Image Process. 17, 2063–2080. doi: 10.1109/TIP.2008.2004797

PubMed Abstract | Crossref Full Text | Google Scholar

Varadaraj, V., Lesche, S., Ramulu, P. Y., and Swenor, B. K. (2018). Reading speed and reading comprehension in age-related macular degeneration. Am. J. Ophthalmol. 186, 138–143. doi: 10.1016/j.ajo.2017.11.026

PubMed Abstract | Crossref Full Text | Google Scholar

Wang, Q., Zhang, Q., Sun, W., Boulay, C., Kim, K., and Barmaki, R. L. (2023). A scoping review of the use of lab streaming layer framework in virtual and augmented reality research. Virtual Real. 27, 2195–2210. doi: 10.1007/s10055-023-00799-8

Crossref Full Text | Google Scholar

Watson, A. B. (2014). A formula for human retinal ganglion cell receptive field density as a function of visual field location. J. Vis. 14:15. doi: 10.1167/14.7.15

PubMed Abstract | Crossref Full Text | Google Scholar

Wong, W. L., Su, X., Li, X., Cheung, C. M. G., Klein, R., Cheng, C.-Y., et al. (2014). Global prevalence of age-related macular degeneration and disease burden projection for 2020 and 2040: a systematic review and meta-analysis. Lancet Glob. Health 2, e106-116. doi: 10.1016/S2214-109X(13)70145-1

PubMed Abstract | Crossref Full Text | Google Scholar

Wood, J. M., Black, A. A., Mallon, K., Kwan, A. S., and Owsley, C. (2018). Effects of age-related macular degeneration on driving performance. Invest. Ophthalmol. Vis. Sci. 59, 273–279. doi: 10.1167/iovs.17-22751

PubMed Abstract | Crossref Full Text | Google Scholar

Zetzsche, C. (2005). CHAPTER 37 – “Natural scene statistics and salient visual features,” in: Neurobiology of Attention, Eds. L. Itti, G. Rees, and J. K. Tsotsos (Burlington: Academic Press), pp. 226–232.

Google Scholar

Keywords: age-related macular degenerateion (AMD), visual function, neural marker, frequency tagging, steady-state visual evoked potentials (SSVEP), steerable wavelet pyramid, spatial frequency

Citation: Renton AI, Klein DJ, Livezey JA, Nemrodov D, Wolfer S, Hanina A and Van De Ville D (2025) Video-evoked neuromarkers of visual function in age-related macular degeneration. Front. Hum. Neurosci. 19:1569282. doi: 10.3389/fnhum.2025.1569282

Received: 31 January 2025; Accepted: 10 April 2025;
Published: 01 May 2025.

Edited by:

Silvio Ionta, Fondation Asile des aveugles, Switzerland

Reviewed by:

Markey Cierra Olson, Barrow Neurological Institute (BNI), United States
Saurabh Gandhi, Indian Institute of Technology Delhi, India
Ivan Markovinović, University of Rijeka, Croatia

Copyright © 2025 Renton, Klein, Livezey, Nemrodov, Wolfer, Hanina and Van De Ville. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Angela I. Renton, YW5naWUucmVudG9uMjNAZ21haWwuY29t

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.