ORIGINAL RESEARCH article

Front. Psychiatry, 07 April 2025

Sec. Digital Mental Health

Volume 16 - 2025 | https://doi.org/10.3389/fpsyt.2025.1559202

This article is part of the Research TopicAutism: The Movement (Sensing) Perspective a Decade LaterView all 28 articles

Hidden social and emotional competencies in autism spectrum disorders captured through the digital lens

  • 1Psychology Department, Sensory Motor Integration Laboratory, Rutgers University, Piscataway, NJ, United States
  • 2Computer Science Department, Center for Biomedicine Imaging and Modeling (CBIM), Rutgers University, Piscataway, NJ, United States
  • 3Rutgers University Center for Cognitive Science (RUCCS), Piscataway, NJ, United States

Background/objectives: The current deficit model of autism leaves us ill-equipped to connect with persons on the spectrum, thus creating disparities and inequalities in all aspects of social exchange in which autistic individuals try to participate. Traditional research models also tend to follow the clinical definition of impairments in social communication and emotions without offering personalized therapeutic help to autistic individuals. There is a critical need to redefine autism with the aim of co-adapting and connecting with this exponentially growing sector of society. Here, we hypothesize that there are social and emotional competencies hidden in the movements’ nuances that escape the naked eye. Further, we posit that we can extract such information using highly scalable means such as videos from smartphones.

Methods: Using a phone/tablet app, we recorded brief face videos from 126 individuals (56 on the spectrum of autism) to assess their facial micro-motions during several emotional probes in relation to their resting state. We extracted the micro-movement spikes (MMSs) from the motion speed along 68 points of the OpenFace grid and empirically determined the continuous family of probability distribution functions best characterizing the MMSs in a maximum likelihood sense. Further, we analyzed the action units across the face to determine their presence and intensity across the cohort.

Results: We find that the continuous Gamma family of probability distribution functions describes best the empirical face speed variability and offers several parameter spaces to automatically classify participants. Unambiguous separation at rest denotes marked differences in stochastic patterns between neurotypicals and autistic individuals amenable to further separate autistic individuals according to the required level of support. Both groups have comparable action units present during emotional probes. They, however, operate within parameter ranges that fall outside our perceptual umwelt and, as such, do not meet our expectations from prior experiences. We cannot detect them.

Conclusions: This work offers new methods to detect hidden facial features and begin the path of augmenting our perception to include those signatures of the autism spectrum that can enhance our capacity for social interactions, communication, and emotional support to meet theirs.

1 Introduction

The current societal perception of individuals with autism spectrum disorders (ASD) inherently depends on the diagnostic criteria and the research narrative that such criteria advance (1). Every paper on autism starts with the description of ASD as deficits in social interactions and communication (2, 3). There are no metrics indicative of readiness potential to socially interact or to communicate, despite the recognized plasticity of the developing nervous system (4).

The screening and diagnostic process of neurodevelopmental disorders relies exclusively on observation, so it remains a challenge to identify any potential competencies for social and emotional exchange. Surely, the system survived a developmental insult and learned to function in the world over the period of 3–4 years that, on average, takes to receive such a diagnosis (5). During this time, it is possible that coping mechanisms developed without the type of support that would be required to gradually learn to bridge the mental intent of a person to the physical execution of a person’s goal-directed thoughts into congruent actions under volitional control (6, 7). The disconnect that autistic individuals report between mental intent and physical action may mislead us, as we have already built-in certain expectations and biases from interactions with others whose actions match their intended consequences.

Because large portions of these aspects of motor control constitute behavioral nuances that transpire largely beneath awareness, we tend to miss them when relying on visual observation alone. Blurred among our expectations and biased subjective experiences may be those competencies of the autistic system longing to build social rapport (8). We may be missing the opportunity to truly help and connect with autistic individuals by relying exclusively on observation.

Observational instruments in the USA, like the Autism Diagnostic Observation Schedule (ADOS; currently in version 2; 2) and the Diagnostic and Statistical Manual of Mental Disorders (DSM; currently in version 5, from the American Psychiatric Association (9)) are built upon subjective opinion. Both instruments are consistently and broadly used in basic research to drive research questions that would eventually be translated into practice. They are also used in clinical practices to recommend services and to drive funding and societal support for this exponentially growing sector of the population.

These tools, which rely exclusively on observation (10), have led to several prominent theories of autism (11, 12), which in turn, have resulted in stigma and even denial of basic education rights to some autistic individuals in the USA. The very notion that autistic individuals lack empathy (13, 14), lack the social desire to communicate (15, 16), are deviant or do not have a theory of mind (12, 17), and, more generally, are deprived of human emotions (1820) has systematically contributed to a narrative that negatively impacts these people’s lives (1, 8). The science behind such opinions is grounded in subjective observational techniques. It merely exposes the tip of the iceberg in a rather misleading way (Figure 1A). The physiological underpinnings of what we observe and interpret are starting to reveal a different picture of autism (8). New computational techniques are emerging to complement the observational instruments. Indeed, they promise to help advance diagnosis and screening tools to a new generation of a more dynamic, objective, quantitative science of autism, a science that would help uncover the readiness potential for social, emotional, and communication exchange (8, 21).

Figure 1
www.frontiersin.org

Figure 1. Current deficit model of autism (A) is the tip of the iceberg. New advances in computer vision amid a digital revolution begin to reveal a new picture of autism (B) whereby the child needs personalized help and objective science-based support to thrive. Here, 5 seconds of video data captured at rest; smiling and surprised show the dysregulated facial micro motions of the child, surely impacting visual perception of others in the social scene and having heightened uncertainty from a brain that is receiving excessive random noise as reafferent input. Under such uncertain world, anxiety and dysregulated states of pain and hypervigilance are not uncommon in ASD. ASD, autism spectrum disorder.

The new digital tools reach beyond the limits of the naked eye and can get to a micro-level of social and emotional nuances critical for conveying gestural communication and building rapport and trust among humans (22, 23). Bringing awareness of such nuances could then make them more accessible to diagnosticians and significantly improve the current clinical methods. Examining people through a new digital lens may shed new light on the social and emotional competencies of autistic individuals. What if we have been wrong all along and, in fact, autistic individuals do have empathy, do have emotions, and do want social interactions, but we cannot see it through the traditional lens that relies exclusively on visual observation?

Quantitative biometrics are beginning to become more ubiquitous in our behavioral sciences. They include methods based on eye tracking (24, 25), facial expressions (26), reaching movements (22, 2729), gait (30), neural correlates from Electroencephalography (EEG) (3133), Magnetoencephalography (MEG) (34), and functional Magnetic Resonance Imaging (fMRI) resting-state motion (3539), among others. Nevertheless, most of this basic research heavily depends on instrumentation that is much too invasive for autistic individuals who suffer from hyper- and hypo-sensitivity to touch (40), movement (4148), temperature-pain (4951), and self-generated involuntary motions (39) that tend to keep the autonomic systems highly dysregulated (52, 53).

New non-intrusive data acquisition techniques amenable to capturing biological motions embedded as nuance in our behaviors are now becoming available thanks to recent advances in computer vision (54, 55). Among them are facial recognition and facial tracking methods that can extract behavioral states from video data (54, 55). They hold promise to identify differences between autistic individuals and controls (24, 26, 56). The few methods that exist are nevertheless still too taxing on autistic individuals. They may require a long time to acquire the data, which requires focus—a challenge for some autistic individuals. They may also heavily rely on heuristics and manual setting of threshold values required to reduce and analyze the large amounts of data that lengthy assays produce.

In this work, we combine a brief assay requiring merely 5 seconds per task to acquire video data in non-intrusive ways using commonly available means such as a webcam, an iPhone, or a tablet. We combine this brief data acquisition assay with new analytics that maximize data extraction and permit a direct personalized assessment of the micro-movements of the face as a person makes facial gestures on command. We seek to identify commonalities across the autistic and neurotypical facial coding ranges at a micro-level to characterize facial nuances beneath awareness that could nevertheless play a key role in building social rapport between autistic individuals and neurotypicals. We posit that by bringing those nuances to awareness, both groups could better co-adapt and coexist while minimizing judgmental opinions and rather presuming that competencies do exist across the human spectrum at a level that we can now make accessible to share.

2 Materials and methods

2.1 Participants

The data set comprised individuals on the autism spectrum and typically developing or typically developed controls (TDs). The broad spectrum included non-speakers with mid to high support needs (ASD-HS) who have a diagnosis of ASD and within this group a subgroup with a diagnosis of apraxia. The ASD group also comprised individuals with low support needs (ASD-LS) who can communicate through spoken language. Among the TDs are also adults who are parents of the non-speaker ASD-HS participants and of those with apraxia. Some of the moms reported diagnoses of acquired autoimmune disorders, depression, bipolar disorder, and attention deficit hyperactivity disorder (ADHD), which we have generally grouped under TD MomM. Other moms reported healthy aging (denoted TD Mom). TD dads did not report any neuropsychiatric disorders. Table 1 reports on the various participating sites. These included a school of children with special needs, two clinics, and one social event.

Table 1
www.frontiersin.org

Table 1. Participants recruitment and demographics.

2.1.1 Strategies for recruiting participants

Participants were recruited through word of mouth, using Institutional Review Board (IRB)-approved flyers distributed across the various Rutgers University campuses. Furthermore, non-profit organizations were contacted, and the IRB-approved flyers were distributed at conferences and various events, schools, and clinics. The Principal Investigator (PI) and students traveled to various events, schools, and clinics to record participants. Clinicians and school principals distributed IRB-approved letters to parents to further assist in student enrollment. Those who expressed interest in participating in the study were recruited.

2.2 Data acquisition and processing

We acquired the data using video cameras embedded in iPhones, tablets, or webcams. We also used a research app that facilitates doing so at home, clinics, or schools without having to come to our lab (see picture in Supplementary Figure 1A). Upon IRB consent, the app guided the person to a session of practice (5 seconds), then resting (5 seconds), and then a series of micro-expressions from instructed emotional gestures, each also lasting 5 seconds. In some participants, we only used smiling (as in happy) and opening of the mouth (as in surprised) (see picture in Supplementary Figure 1B). In other participants, we used those micro-expressions and other emotional gestures such as anger, disgust, contempt, fear, and sadness. The latter combined with the former formed seven facial micro-expressions popularized by Paul Eckman (57) and other researchers. We used the instructions by Vanessa Van Edwards (see Note 1) to provide guidance (to the caregiver) on how to produce motions of a facial expression to instruct the participant. In some cases, we directly instructed the participant.

The purpose of the experiment was not to identify emotions but rather to capture micro-movements of the facial action units (AUs) and facial grid points that we can extract from the videos using the OpenFace software (54). Figure 1B shows examples of the grid that we can extract from brief videos in different populations (neurotypical, ASD-HS, and ASD-HS with the additional diagnosis of apraxia). Importantly, we can see the motion trajectories of the pixels’ positions as they change during resting vs. other states of smiling or surprise. Figure 2 explains the pipeline of data parsing. Upon collecting the videos, we ran OpenFace (see Note 2, https://github.com/TadasBaltrusaitis/OpenFace/wiki) and retained the facial grid and AUs (see Supplementary Figure 2 and Note 2). We then used the trigeminal nerve (cranial nerve V) regions shown in Figure 2C to parse out three subregions of the face (ophthalmic, maxillary, and mandibular), which we named V1, V2, and V3, respectively. These were colored differently to map those grid points of Figure 2B in correspondence with the regions in Figure 2D innervated by the trigeminal nerves. These are important for sensing movement reafference throughout key facial regions that enable social and emotional communication (e.g., the eyes, ears, mouth, lips, tongue, and mandibular regions).

Figure 2
www.frontiersin.org

Figure 2. Pipeline of data acquisition and pre-processing. (A) Sample grid from OpenFace was extracted from 5 seconds of video data captured using a research app. (B) OpenFace grid was used to track the 68 markers numbered from 0 to 67. (C) The trigeminal ganglion-inspired parcellation of the face grid points into three facial subregions of V1, V2, and V3 to analyze the data according to subregions. (D) The 5-second positional pixel trajectories across the subregions. (E) One grid point 5-second positional trajectories of the X and Y coordinates converted from pixels to mm. (F) The smoothed position to enable proper differentiation to obtain the velocity with scalar values (speed) in panel G obtained by gluing all segments of the region (V1 in this case) and (H) obtaining the frequency histogram of the peaks (red dots) in panel (G) Multiple fitting distributions were examined and the continuous Gamma family of probability distributions selected as the best fit in a maximum likelihood sense. Written informed parental consent was obtained from the parents of the individual.

The 5 seconds of the assay during the resting state is shown in Figure 2E for the positional X and Y pixel coordinates (normalized to adjust for discrepancies in distance from the face to the camera). Please see the subsection below. We smoothed out the facial grid points’ trajectories using in-home developed software that employs the spline interpolation toolbox MATLAB version R2023b. We then obtained the speed quantities by proper differentiation and glued all segment trajectories in each point of the grid corresponding to the subregions spanned by 68 points on the grid of Figure 2B. These are 26 points of the grid in Figure 2B, V1 (17:30 33 36-47); 17 points of the grid, V2 (31 32 34 35 48 49:53 54 68:64); and 25 points of the grid, V3 (0:16 55:59 65:67).

Figure 2G shows the speed points corresponding to the speed segments of the 26 grid points of V1. We focused on the peaks (marked as red dots in Figure 2G) to obtain the frequency histogram of their distribution across all these points of the grid during the 5-second task. This is shown in panel Figure 2H. We were not interested in the temporal component at this point, and for that reason, we did not treat this sequence as a time series but rather focused on the set of variations in the amplitude of the speed signal. Shuffling the points in each region to glue them in a different order did not impact the shape of the histogram. The overall result of the work remained, and the analysis produced similar parameter ranges. However, for consistency, to glue the speed segments, we set a fixed order of the points in the grid and used that order for the entire data set. This order is as explained above in V1, V2, and V3 based on the proximity of the grid points because of inherent synergies and co-dependencies of their motions. These co-dependencies are influenced by allometric effects of the distance between points, with lengths that vary across the population due to anatomical and size differences.

A sampling at 30 Hz and 5 seconds worth of data gave us 150 frames per grid point for a total of 3,900, 2,550, and 3,750 frames for V1, V2, and V3, respectively. These in turn provided over 100 peaks of the speed amplitude, such that empirical estimation of the stochastic signatures of speed variability provided enough statistical power and tight 95% confidence intervals. We caution that if the analysis were to focus on the temporal dynamics of the time series data, one would need to double the lower bound of data time to 10 seconds and focus on the individual time series. For other types of temporal analysis, one can use the series of peak width (ms) and analyze the stochastic signatures of the glued data. However, here, we focus on the speed amplitude variations rather than on the temporal information. These are different aspects of the problem, but the methods provided here for analyses of the variations in speed amplitude can also be used for temporal dynamics-related data (see, for example, 30, 58).

2.2.1 Distance normalization

To account for the translation of the subject along the optical axis, we performed the following z-score normalization: for each frame, we took the mean and standard deviation of each positional coordinate component, x and y, and performed a transformation like a z-score calculation:

zx=(xμx)σx,zy=(yμy)σy.

This transformation thereby produces a time series of normalized faces (collections of 68 points), each with a mean of 0 and a standard deviation of 1. While this normalization does not prevent distortions from head rotations or lens compression, it nonetheless is an essential preprocessing step to create invariance to the distance from the camera.

2.2.2 Micro-movement spikes and distribution fitting

The distribution of the peaks’ amplitude is subject to distribution fitting, yielding the continuous Gamma family of probability distributions as a good fit to capture both the autistic and neurotypical stochastic ranges. This is so because the autistic sample also includes the memoryless exponential distribution, which has a shape parameter a=1 in the Gamma family.

We then empirically estimated the Gamma mean in each subregion Γμ=a·b, and for each point on the ordered series of grid points in the subregion, we obtained the absolute deviations from the empirical Γμ. This is shown in Figure 3A with the peaks of the absolute deviations from the Γμ marked red dots. For each peak, we normalized the value using the formula in Figure 3B to scale out neighboring effects related to anatomical differences and disparities in distances between the grid points across the participants (59). Figure 3C then shows the data from the normalized peaks, which we named micro-movement spikes [granted US patents (6063) filed in 2012] and for which we obtained the frequency histogram and Gamma fitting. These are also expressed in Figure 3E as spikes ranging in the real valued interval from 0 to 1 across the original number of points in the series of Figure 3A. Compare these to Figure 3C with fewer points of non-zero absolute deviations from the empirical Gamma mean of the region (V1 in this case). The micro-movement spike (MMS) comprising all frames can be turned into binary spikes and other sets of information theoretical tools employed (as in recent work from our lab; 64) when we used time series data of a longer time range. In the present work, since we deliberately chose the lowest possible time range (5 seconds) that still enables statistical power, we focused on stochastic analyses instead.

Figure 3
www.frontiersin.org

Figure 3. Pipeline of data analysis. (A) Absolute deviations from the empirically estimated Gamma mean ( Γμ) were obtained for each point of the set, and the peaks were extracted for local scaling. (B) Local scaling to normalize the speed amplitude and dampen effects of anatomical differences and lengths across faces. The local peak was scaled by the sum of the peak value and the average value of its neighboring local minima. (C) The micro-movement spikes (MMSs), considering only the frames with the speed deviation peaks above 0 value. (D) The histogram of the MMSs and the fitted Gamma probability distribution function. (E) The full MMSs inclusive of 0-valued entries in (A) retaining all original frames distributed bimodally between small and larger deviations and can be used in other analyses beyond the scope of this paper. (F) The Gamma parameter plane spanned by the shape and the scale values and 2 representative values derived from the empirical speed’s MMSs with 95% confidence intervals. (G) Empirical Gamma PDFs corresponding to the representative values in panels (F, H) Parameter space spanned by the Gamma scale (NSR) obtained from all points in V1, V2, and V3, also for the representative participants of panels (F, I) Parameter plane spanned by the norm of the V1, V2, and V3 Gamma NSR and the norm of the V1, V2, and V3 Gamma Skewness from the example in panel (F).

Upon estimation of the best fitting distribution to the peaks of the non-zero deviations from the empirical Γμ, we found once again that the continuous Gamma family of probability distributions best fits the data (in a maximum likelihood estimation sense). We then plotted the Gamma shape (a) and the Gamma scale (b) parameters of the Gamma family thus obtained on a parameter plane with 95% confidence intervals. In Figure 3F, we show an example of an ASD vs. a TD participant, and in Figure 3G, we show the empirical Gamma Probability Density Function (PDFs) for each of the three facial regions, V1, V2, and V3, that we have defined in this study. Figure 3H shows another parameter space spanned by the Gamma scale (the noise-to-signal ratio ΓNSR=ΓσΓμ=a·b2a·b=b) of each subregion. We then used the scalar value of the vectorial representation of these respective noise levels (V1, V2, and V3), i.e., the Gamma Noise to Signal Ratio (NSR) and the Gamma skewness of each of the empirical distributions, to plot the scalar quantities as points on a parameter plane. This parameter plane spanned by these two dimensions of the data is shown in Figure 3I for the example of ASD and TD representative participants.

Lastly, we measured, pairwise, the distances between the distributions of the MMS peaks for each face region and participant. To that end, we used the Earth Mover’s Distance (EMD) (65, 66) and plotted the color maps for each of the facial gestural assays. We used the resulting matrices as input to a tree classifier, and given the number of subtypes of interest, we then quantified the percentage of ASD vs. TD participants in each cluster that the algorithm found. These were reported on each leaf of the branches denoting emerging subtypes in the tree structure.

2.3 Action unit identification

The OpenFace interface reveals universal AUs present (as a binary matrix of N frames by 18 columns) in each facial expression. This is in the form of an N × 18 matrix, where N is the number of frames collected at 30 Hz and 18 are AUs defined in the OpenFace GitHub site (see Note 2). The intensity of the AUs is also defined in an N × 17 matrix.

Across frames, we used the edit distance (67) of the column-wise binary vector denoting the presence of the AU to measure its frequency across the 5 seconds of motion yielding approximately 150 frames per point in the grid (plus/minus two frames upon OpenFace estimation). Then, we obtained the most common pattern present (the mode) to denote the presence of the AU. This pattern was saved along with its corresponding intensity values. We saved intensities equal to and above the 0.001 level. We did this for the individuals of the ASD and TD control groups. We then pooled across the intensity values and thus identified and reported on the frequency histograms of intensity values for the AUs that are present in each group.

3 Results

3.1 Stochastic differences between ASD and TD are captured at rest

At rest, TD individuals produced separable patterns of facial micro-movements from ASD participants. Each one of the facial regions showed separable patterns, expressing more variability in the ASD participants across all subregions. Figures 4A–C show the patterns of Gamma shape and scale parameters along with the empirical PDFs that shift in ASD relative to controls, with the standardized absolute deviations from the Gamma mean corresponding to the micro-movements’ spikes of the facial subregion.

Figure 4
www.frontiersin.org

Figure 4. Results from the analysis of resting state data, whereby 5 seconds of the face at rest was captured on video, and the pipelines in Figures 2 and 3 were executed. (A–C) Gamma plane results for each of the V1, V2, and V3 facial regions. Inset corresponds to the empirical Gamma PDFs showing shifted density and differences in dispersion and skewness. Notice the span of the ASD scatter with broader variability (along both dimensions) at rest than that of the TD controls. (D) Gamma NSR parameter space also reveals differences in the scatter between ASD and TD groups, with higher noise for the mandibular region V3 in ASD-HS apraxia subgroup. Lower skewness in TD is also evident. Higher skewness in ASD points at heavier tails in the empirical Gamma PDFs with higher speed deviations from the empirical Gamma (MMS speed peaks) mean. These represent rare events relative to controls with lower skewness tendencies. (E) Parameter plane spanned by the norm of the Gamma distributions' NSR (scale parameter) from the (V1, V2, V3) vector and the corresponding norm of the Gamma distributions' skewness of V1, V2, V3). (F) Color map of the EMD taken pairwise for ASD and TD participants across all regions V1, V2, V3. (G) Corresponding tree with cluster composition. (F) Pairwise EMD obtained for both groups, and each of the face subregions separate ASD from TD and provide structure denoting differences in distributions of the normalized speed peaks at rest. (G) Tree clustering of (F) outputs with different compositions in the 6 subtypes (3 face regions × 2 groups). ASD, autism spectrum disorder; TD, typically developed; HS, high support; EMD, Earth Mover’s Distance.

Along the dimensions of the Gamma NSR in Figure 4D, we observed orderly patterns, suggesting that this projection of the high-dimensional data may produce an embedding of contiguous points distinguishable between ASD and TD. These are conserved in Figure 4E along the parameter plane that focuses on the scalar quantities. In Figure 4D, elevated NSR in V3 for a subset of participants on the spectrum revealed correspondence with those with the ASD-HS subset with the diagnosis of apraxia. These patterns expressed statistically significant differences in the Gamma scale (NSR) and higher skewness parameter (indicative of heavy-tailed distributions with more frequent appearance of high-speed motions in the MMSs). Please see Supplementary Figure 3 showing pairwise Wilcoxon rank-sum test statistics (equivalent to the Mann–Whitney U-test) for ASD vs. TD inclusive of micro-expressions during the resting state vs. micro-expressions for anger, happiness, sadness, and surprise. Area V1 was consistently significantly different between TD and ASD participants across the resting state and all other micro-expressions (taken pairwise, p < 0.01). Furthermore, in V1, within each of the cohorts, the resting state micro-expressions were significantly different from those in all other states. In the ASD group, happiness and surprise also differed significantly from all other micro-expressions (p < 0.01). In the TD group, anger vs. surprise was significantly different at p < 0.01, and sadness vs. anger was significantly different at p < 0.05. In areas V2 and V3, statistical differences followed more complex patterns.

In V2, TD at resting state differed from all ASD micro-expressions but only significantly for resting and happiness at p < 0.01. TD anger and surprised differed significantly from ASD happiness at p < 0.01. Within the TD group, resting differed from happiness and sadness, as did happiness and sadness relative to surprise, all at p < 0.05. Within the ASD cohort, anger differed from happiness and happiness from anger at p < 0.01. Surprise in ASD differed from all micro-expressions with significance relative to rest and happiness, p < 0.01.

In V3, the resting micro-expressions in ASD differed from all micro-expressions in TD (p < 0.01). ASD happiness differed significantly from TD anger and TD happiness (p < 0.01). Within the ASD cohort, ASD happiness differed from resting, anger, sadness, and surprise (p < 0.01). ASD surprise differed from all micro-expressions with significance at p < 0.01 for happiness and p < 0.05 for resting, while anger and sadness did not reach significance relative to surprise.

In Figure 4D, we compare the Gamma NSR along each of the facial subregions. The patterns separate participants with ASD-HS apraxia along the V3 (mandibular region) from other ASD participants and TD controls. All ASD distributions from the resting condition showed higher skewness than those from the TD controls and broader ranges of the Gamma NSR, with heavier right-tailed distributions for the ASD-HS participants. This can be appreciated in Figure 4E. Furthermore, there is a visible structure in the EMD matrix of Figure 4F denoting the pairwise differences in distributions according to the similarity metric (EMD) across the participant’s three facial subregions. Within the TD group, we saw similarities in the distributions of standardized speed MMSs in the lower values of the EMD. Furthermore, clear differences between the ASD and TD groups were captured with automatic clustering using six groups (ASD vs. TD and three face regions) that break down the composition of each subtype in Figure 4G.

3.2 Facial micro-movements during expressions of anger, happiness, sadness, and surprise show significant stochastic differences between ASD and TD

In the subset of individuals where, upon resting, we measured anger, happiness, sadness, and surprise micro-expressions as instructed by the person taking the measurements (as in Note 2), profound differences were captured in the Gamma NSR (along all three facial subregions) and Gamma skewness by the empirical distributions’ shifts. Figures 58 show these patterns for each of the parameter spaces of interest. Statistically significant differences were found across all pairwise comparisons using the rank-sum (Wilcoxon) test. For the scalar quantity of the Gamma scale, TD anger vs. ASD anger was significantly different at the alpha 0.05 level, p < 0.02, resting p < 8.4 × 10−8, sadness p < 5.7 × 10−4, surprised p < 7.57 × 10−6, and no significant differences in the ΓNSR micro-expressions for happiness between TD and ASD. The skewness scalar quantity also showed significant differences across all micro-expressions with p << 0.001.

Figure 5
www.frontiersin.org

Figure 5. Results from the analysis of anger micro-expressions, whereby 5 seconds of the face was captured on video upon instruction of how to enact facial micro-expressions of an angry face. The pipelines in Figures 2 and 3 were executed to appreciate the shifts from resting state in Figure 4. (A–C) Scatters on the Gamma plane show broader ranges of motion in TD signaling more variability in the two dimensions across the TD population. (D) The ASD scatter on the Gamma NSR space separates from the TD scatter, which spans broader ranges of noise (variability in speed MMSs) across all facial subregions. (E) Higher skewness scalar quantities in ASD signal heavier-tailed empirical Gamma PDFs during anger expressions, denoting more rare events of faster deviation peaks from the empirical Gamma mean. (F) Pairwise EMD metric shows more similarity among controls (blue tones with low EMD values) than ASD and differences in normalized speed deviations (MMSs) across all regions. ASD group is more similar in V1 (comprising the eyebrow and eye micro-expressions). (G) Different clusters and their composition for the 6 subgroups (2 groups × 3 subregions of the face). TD, typically developed; ASD, autism spectrum disorder; MMSs, micro-movement spikes; EMD, Earth Mover’s Distance.

Figure 6
www.frontiersin.org

Figure 6. Results from the analysis of happiness micro-expressions, whereby 5 seconds of the face was captured on video upon instruction of how to make a happy face by smiling. The pipelines in Figures 2 and 3 were executed to appreciate the shifts from resting state in Figure 4. (A–C) Gamma plane scatters show marked differences in the V3 mandibular area engaged in motions to produce a smile. Subregions V1 and V2 show more overlapping patterns, apart from outliers with higher noise, which correspond to the ASD-HS apraxia subgroup. (D) Separation between the scatters of TD and ASD show higher ranges of speed MMS variability in TD motions and outliers (ASD-HS apraxia subgroup) span broader ranges of noise. (E) The ASD scatter shows higher Gamma skewness ranges indicating heavier right tails of the distributions as with resting and anger, with more rare events (higher deviations from the Gamma speed mean boasting faster changes in position of the points). (E) Parameter plane spanned by the norms of the Gamma probability distributions' NSR and skewness from (V1, V2, V3) showing the higher skewness values for ASD (denoting the accumulation of more rare events on the right tail (higher speed fluctuations away from the empirical Gamma mean) (F) Color map from pairwise EMD across ASD and TD face regions V1, V2, V3. (G) Corresponding tree cluster with compositions from each group. (F) Pairwise EMD shows similarity within V1, most of the groups in V2 and V3 in ASD across the group, and differences in V1–V2 and V2–V3 comparisons with ASD. The TD group boasts more similarity across regions. Pairwise comparisons show V2 as the most similar region between TD and ASD. ASD, autism spectrum disorder; HS, high support; TD, typically developed; MMS, micro-movement spike; EMD, Earth Mover’s Distance.

Figure 7
www.frontiersin.org

Figure 7. Results from the analysis of sad micro-expressions, whereby 5 seconds of the face was captured on video upon instruction of how to make a sad face (as in Note 1). The pipelines in Figures 2 and 3 were executed to appreciate the shifts from resting state in Figure 4. (A–C) Gamma plane scatters and insets showing the empirical Gamma PDFs. Notice the marked shifts in each case with V3 at the highest differentiation. (D) Both scatters on the Gamma NSR parameter space showing the departure of ASD from TD scatter. (E) The Gamma NSR scalar quantity vs. the Gamma skewness scalar quantity also separates the groups. (F) The pairwise EMD matrix showing differentiation within the ASD group and higher similarity within the TD group. Highly structured patterns of EMD across the three regions are marked between the two groups. (G) The tree clustering analyses and reporting on the composition of each subgroup. ASD, autism spectrum disorder; TD, typically developed; EMD, Earth Mover’s Distance.

Figure 8
www.frontiersin.org

Figure 8. Results from the analysis of surprise micro-expressions, whereby 5 seconds of the face was captured on video upon instruction of how to make a surprised face (as in Note 1). The pipelines in Figures 2 and 3 were executed to appreciate the shifts from resting state in Figure 4. (A–C) Gamma plane scatters with insets showing the corresponding empirically estimated Gamma PDFs. (D) Gamma NSR parameter space with separated groups. (E) Gamma NSR scalar quantity vs. skewness scalar quantity of the two groups. Notice the overlap with some of the members of the TD group along both axes and the separation of most TD participants along the NSR axis due to more uniformity in facial patterns across the TD group. (F) The pairwise EMD matrix showing structure within the ASD group indicative of V1–V2 similarity vs. V1–V3 differentiation. TD group also shows high structure within the group and highest departure in distributions from the ASD group for area V3. ASD, autism spectrum disorder; TD, typically developed; EMD, Earth Mover’s Distance.

The EMD matrices and automatic cluster analyses across these emotional gestures also revealed the composition of the different subtypes amenable to visualizing their differences. In all cases, the patterns clearly showed broad ranges of variability in the Gamma NSR of the ASD participants and higher skewness ranges for ASD. The latter corresponded to higher MMS speed peak deviations from the empirical Gamma mean of the group, estimated for the anger activity. Infrequent events of faster speed peaks in ASD than those in the TD distributions showed in the ASD group, thus resulting in higher skewness. The summaries of these parameter ranges are depicted in panels D and E of each of Figures 5-8. Likewise, the EMD matrices and tree clustering revealed the differentiation and inner similarities of the groups across the face subregions under study.

3.3 Action units reveal facial emotions in ASD at the speed micro-movement level

The analysis of AUs revealed that most AUs present in TD participants are also present in ASD participants albeit with different distributions of intensities. At rest, TD did not show the lip corner puller present in ASD but showed upper lid raise absent in ASD participants. All other AUs present in TD were also present in ASD (brow lower, lid tightener, dimpler, and lip tightener). This can be seen in Figure 9 for all AUs across the V1–V2–V3 facial subregions. There we also appreciate the differences in the intensity values and their distributions.

Figure 9
www.frontiersin.org

Figure 9. Action units (AUs) engaged during resting state. Overlapping AUs between ASD and TD participants in brow lower and lid tightener show different ranges of intensity but common presence in both groups. Upper lid raiser only appeared in the TD group. These are AUs corresponding to area V1. In V2 and V3 areas, lip tightener was commonly present in both groups albeit with different distributions of intensities. The AU lip corner puller was present in ASD but absent in TD, whereas the dimpler was present only in TDs. ASD, autism spectrum disorder; TD, typically developed.

Anger (Figure 10A) reveals the presence of the brow lower for both ASD and TD, with higher intensity across ASD participants and different intensity distributions relative to TDs. The TD group showed upper lid raiser, which was absent in the ASD group. Lid tightener was also more intense in ASD than TD and had a different distribution of intensities than TD. Cheek raiser was less intense in the ASD group than in the TD group, with significantly lower intensity range and values. Inner brow raiser was present in both ASD and TD but was far more intense in ASD than in TD. ASD had an eye blink AU present, which was absent in TD.

Figure 10
www.frontiersin.org

Figure 10. Action units for anger and happiness were also commonly present in both groups, but the ranges of intensities and their distributions differed between groups. The ASD group had AU eye blink present, but this was absent in the TD group for both (A) anger and (B) happiness. Upper lid raiser was absent in anger for ASD but present for TD. In happiness, brow lower (a common AU for anger) was present in ASD but absent in TD. ASD, autism spectrum disorder; AU, action unit; TD, typically developed.

In subregions V2 and V3, anger speed motions from the facial expression revealed nose wrinkling and lips apart in ASD but not in TD. All other AUs were present in both groups, with differences in the range of intensities. Lip corner puller and lip tightener had comparable intensities in TD and ASD, but the dimpler AU was more intense in ASD than TD.

Happiness (Figure 10B) speed micro-motions revealed the presence of brow lower in ASD (commonly present in anger micro-expressions) but absence in TD. Upper lip raiser was present in both groups but expressed with higher intensity in ASD. Likewise, lid tightener was present in both but far more intense in ASD than TD, with fundamentally different distributions of intensities across the two groups’ participants. Cheek raiser, a prominent AU in micro-expression of happiness (see Note 1), was present in both groups but more intense and with broader ranges of values in TD than ASD. As with anger, blinking was present during happiness gestures in ASD but absent in TD.

Sad micro-expressions (Figure 11A) revealed in V1 the brow lower AU with much higher intensity values in ASD. Upper lid raiser was absent from ASD, but both groups had lid tightener with comparable distributions of intensity values across participants. In ASD, the cheek raiser was present with low intensity and absent in TD. Subregions V2 and V3 showed the lip corner puller and chin raiser in ASD, but these AUs were absent in TD. The outer brow raiser was absent in ASD but present in TD. In subregions V2 and V3, upper lip raiser was present but much less intense in ASD than in TD. The dimpler was more intense in ASD than in TD, and the distribution of intensities across participants was different, with a much narrower and lower value range in TD.

Figure 11
www.frontiersin.org

Figure 11. Action units (AUs) for sad (A) and surprise (B). As with the other micro-expressions and resting state, several AUs in V1 were commonly present in both groups, yet their intensity differs in ranges and distribution shapes (see text). In the sad micro-expression, the upper lid raised, and outer brow raiser was absent in ASD, but the cheek raiser (key to the happiness micro-expression) was instead present in ASD but absent in TD. For V2 and V3 subregions, in the sad micro-expression, the TD did not have AUs for the lip corner puller and chin raiser, which were both present in ASD. All others were common to both groups but differed in intensity distribution and ranges. The lip tightener common to both groups had comparable intensity. For the surprise micro-expression, more AUs were present in ASD than TD for V1 with complementary patterns. Lid tightener was present in ASD but absent in TD, while brow lower, inner brow raiser, upper lid raiser, and cheek raiser were activated in ASD but not in TD. Subregions V2 and V3 had overlapping AUs for lip corner puller, dimpler, and lip tightener, yet their distributions of intensity differ. ASD had AUs for upper lip raiser and chin raiser present, but these were not present in TD. Lastly, jaw drop, which is common in the surprise micro-expression, was present in TD but absent in ASD, which instead activated chin raiser. ASD, autism spectrum disorder; TD, typically developed.

Surprise (Figure 10B) in subregion V1 had brow lower, inner brow raiser, upper lid raiser, and cheek raiser in ASD, but these AUs were absent in TD. TD showed lid tightener, but this AU was absent in ASD. For regions V2 and V3, jaw drop (common in surprise micro-expressions) was present in TD but absent in ASD, along with upper lip raiser. Other AUs were present in both groups but with different levels and distribution ranges of intensities. The lip corner puller was far more distributed in TD than ASD, but ASD had some outliers with high-intensity values, a trend also observed in the lip tightener. The dimpler was present in both and had a comparable distribution of intensity values. Chin raiser was only present in ASD (complementing jaw drop that was only present in TD).

3.4 In non-speakers, ASD-HS with apraxia is separable from ASD-HS

Across resting, anger, happiness, sadness, and surprise, the subset of non-speaker participants with apraxia diagnosis differed significantly from the subset of non-speaker participants with the ASD diagnosis alone. This can be appreciated in Figure 12A where we depict the Gamma NSR corresponding to V1, V2, and V3 facial regions and in Figure 12B where we show the parameter plane of the Gamma NSR scalar vs. the Gamma skewness scalar obtained from the three subregions. Figure 12C shows the empirical Gamma PDFs for selected subregions in each of the facial micro-expressions. In each case, the non-speakers with ASD-HS with apraxia separate from the non-speakers with ASD-HS. Supplementary Figure 5 shows the pairwise statistical comparisons significant at the 0.01 alpha level for the Mann–Whitney test.

Figure 12
www.frontiersin.org

Figure 12. Comparison between non-speaker ASD-HS and those with the additional apraxia diagnosis. (A) The Gamma NSR parameter space for resting, anger, happiness, sadness, and surprise tested in this subset of the groups. (B) The Gamma NSR scalar quantity vs. the Gamma skewness scalar quantity parameter plane with the scatters colored coded according to the three groups. Notice the departure of the apraxia subgroup from the others. (C) Selected empirical Gamma PDFs for different subregions of the face. ASD, autism spectrum disorder; HS, high support.

3.5 Differentiation between ASD-LS and TD across the seven facial universal micro-expressions

The empirical distributions of speed MMSs derived from the positional pixel trajectories in the points of the V1, V2, and V3 regions in ASD-LS (participants requiring a lower level of support and having speaking abilities) were compared to those of TD controls close to their age. The results are shown in Figure 13A using the normalized EMD metric values for each pairwise comparison. The structure of this matrix reveals the differences between regions and participating groups. Within each group and facial region, we appreciated the structure of the matrix and saw the self-emerging boundaries of higher values of standardized EMD (denoting higher differences in distribution) with an overall higher similarity within the TD group than within the ASD-LS group (see the more prevalent blue hues denoting similarity in distributions of speed peaks in the standardized MMSs). Figure 13B shows the subtypes emerging from the 14 clusters spanned by the two groups and seven facial micro-expressions according to the cluster analysis. The composition of the 14 clusters grouping ASD or TD participants across the leaf of each tree branch was also reported according to the higher percentage of the group for each leaf.

Figure 13
www.frontiersin.org

Figure 13. Pairwise EMD for the seven recognized facial universal micro-expressions (anger, contempt, disgust, fear, happiness, sadness, and surprise) (A) showing more uniformity (similarity) of distributions in the TD groups and distribution differentiation between ASD-LS group and TD. (B) Tree clustering according to 14 subtypes (2 groups × 7 micro-expressions) with the percentage composition of each cluster. EMD, Earth Mover’s Distance; TD, typically developed; ASD, autism spectrum disorder; LS, low support.

3.6 TD composition and non-speaker ASD-HS parental differences

We compared the Gamma NSR ( ΓNSR) across each of the facial subregions in Figure 14A for all TD controls and color-coded the TD parents of non-speaker ASD-HS participants. This comparison revealed that most parents lie within a different subregion of this parameter space, away from most other TD participants of comparable age. Their speed peaks from the standardized MMSs showed lower levels of noise (lower variability), denoting some commonality among this random draw of this population. Furthermore, they self-clustered into dads, moms, and moms with neuropsychiatric conditions (MOMsM). When superimposing the non-speaker ASD-HS and non-speaker ASD-HS with apraxia on the ΓNSR vs. the Gamma Skewness parameter plane of Figure 14B, we appreciated that the NSR levels of the ASD-HS apraxia were comparable to those of their parents. There were higher skewness values for 6/9 participants in this ASD-HS apraxia group, denoting higher speed (faster) deviations from the Gamma mean, a feature that seems to be within a uniquely higher range for this subset of the group.

Figure 14
www.frontiersin.org

Figure 14. Analysis of TD group labeling the parents of ASD-HS and ASD-HS apraxia (A) On the ΓNSR space, according to moms, dads, and moms who reported autoimmune disorders and other neuropsychiatric conditions (acquired as adults). Notice the separation and clustering of each subtype departing from other TDs including those of comparable ages. (B) Parameter plane of ΓNSR scalar quantity vs. skewness quantity also shows the ASD-HS and ASD-HS apraxia relative to other TDs. Notice the uniqueness in patterns of some participants with ASD-HD apraxia. TD, typically developed; ASD, autism spectrum disorder; HS, high support.

4 Discussion

This work aimed at characterizing the nuances of facial emotional micro-movements using a brief and simple assay under the guidance of instructed emotional facial movements in micro-expressions. Using new personalized methods that do not a priori assume theoretical distributions, we wanted to better understand hidden aspects of facial motor control in autistic individuals. These participants spanned from lower to higher levels of support. The study included ASD-HS non-speakers who communicate through various augmented communication methods. To that end, we designed a research data acquisition app that requires very little effort and is brief. The acquisition step offers instructions amenable to being deployed outside the lab during natural activities. These included activities at their school, at a social event where we randomly sampled people in both the TD and ASD groups, and at various clinical settings (studios) where they received therapies.

4.1 Highly dysregulated patterns in ASD at rest

The resting state activity had elevated levels of noise across the ASD group, with increasing trends for ASD-HS non-speakers and highest ranges for non-speakers of ASD-HS with a diagnosis of apraxia. The latter is an interesting subset because most if not all ASD-HS participants have a disconnect between the movement plan and the execution of their intended motions. However, in the subgroup with the specific apraxia diagnosis, their baseline dysregulation must be visible to the extent that it reaches the level of visual detection and be diagnosed as an additional disorder. In this sense, resting state noise in facial micro-movements derived from the speed parameter seems to be informative of increasing levels of support across ASD. This result, in the realm of facial micro-movements, is congruent with prior results from our lab involving body micro-movements. Specifically, it is consistent with results concerning excessive motor noise during resting state fMRI reflected in the head motions’ variability (35), including increasing trends with the use of psychotropic medications (36) and general involuntary motions at rest in ASD (29, 38). It is also consistent with increasing random noise in the speed’s micro-movements of voluntary motions increasing with the level of clinical ASD severity in the context of pointing to communicate a decision and/or pointing to a visually prompted target (22, 29). In both bodies of work, we found that the higher the level of support needed in ASD, the higher the Gamma noise level in the micro-movements of the speed kinematic parameter, with increasing trends as the population ages (29, 38, 39).

Within the framework of facial micro-movements, this is the first time that we saw the Gamma noise profiles and can further appreciate the correlation of increasing levels of random noise in the speed motion variability with increasing levels of needed support that have visible apraxia at the highest end. Although evaluating the 3D gaze data from OpenFace outputs will be reserved for a future study, prior work in the field has revealed elevated oculomotor randomness in ASD with increasing trends in noise levels that, as in the pointing case, increase with the level of severity (68). Under these conditions, the type of eye–hand coordination required to deploy the arm linkage (with high degrees of freedom to control) forward to bring the finger to accurately point to a visually prompted target, and then backward to rest, must be very challenging for these individuals. We know, from studying goal-directed movements in the context of deafferentation, that in the absence of reafferent feedback sensation from micro-movements, the neural (surface EEG) correlates of directional motion intent reflect a much higher cognitive load than controls who have proper micro-movement reafferent feedback (69). It is likely that with excessive random noise in the reafferent feedback code, the participants with ASD also experience a higher cognitive load in that they would have to pay attention and be highly aware of activities that typically transpire largely beneath awareness. In this sense, the facial micro-movements activity may reflect the type of dysregulation that a system overburdened with such taxing states is bound to experience.

The present results were derived using the same unifying statistical platform for individualized behavioral analysis that we have used in previous work involving biorhythmic activity data from wearable devices and pose estimation using computer vision techniques (Statistical Platform for Individualized Behavioral Analysis (SPIBA); 70, 71), which have the potential to link, across the population, multiple levels of micro-movements’ noise speed with levels of facial apraxia. They can also reveal issues with eye–hand coordination during pointing behavior to communicate a decision or to point to a visually prompted target. Together, these results have implications for the design and deployment of augmented communication methods. Any communication technique developed for ASD individuals will need to consider these elevated noise levels in the speed parameter to design regulatory support aimed at dampening the motor noise during their clinical therapy and/or school-teaching sessions. Of the activities that we examined under this digital lens, we found that the resting state maximally captured disparities across the cohort. This simple assay may indeed provide the type of information that we need to estimate how regulated a system is, in the precise sense of assessing its level of volitional control. The level of control of the facial and body micro-movements at rest may be informative of individualized levels of overall motor control in flux. It may also help us derive individualized indexes of motor control reflecting the level of agreement between mental intent and physical execution of the intent—an aspect of motor control that is uniquely different in ASD.

The motor noise, which shifts dynamically at the output level as activities of daily life carry on (30, 64), may also serve as a proxy of the quality of feedback that a person’s Central Nervous System (CNS) is receiving from the Peripheral Nervous System (PNS), informing the CNS of ongoing activity, even at the sub-second time scale. As such, the level of regulation of the system and the level of smoothness in action execution (matching the intended plan of the action) may be reflected, at a micro-level, on the facial activity. At one extreme, we have levels of noise and speed MMS distribution skewness that correspond to neurotypical levels. In stark contrast at the other end, we have the largest departure from neurotypical levels on the visibly detectable apraxia. In such cases, the intended plan visibly mismatches the action execution, and even an observer, like a speech therapist giving this diagnosis of apraxia, can detect the mismatch relative to the expected neurotypical levels. Based on these results and the body of knowledge that we have accumulated over a decade of work, we posit that the stochastic signatures of the facial speed micro-movements data may indeed provide a window into the levels of feedback noise, the level of dysregulation, and the associated levels of needed support in ASD.

4.2 Presence of action units in ASD but differences in intensity ranges and distributions

Across ASD, we found AUs present underlying the speed micro-movements derived from positional trajectories of the 68 points in the grid that we extracted from videos using OpenFace algorithms. Contrary to the assumption that ASD individuals do not have emotions or lack empathy, we found that they indeed engage (on command) the universal AUs across the face, across subregions V1, V2, and V3 of the digital grid. They, however, do so with different ranges of intensity than those captured in the TD group. As such, the variations in speed amplitudes of the micro-movements from facial universal micro-expressions associated with emotions operate at unexpected stochastic ranges. We posit from these results that folks observing these ranges to screen social engagement and emotions seem to miss these ranges amid rapidly changing social dynamics. It is possible that the expected values of such ranges in neurotypicals do not overlap with those of ASD. Since our visual perception largely depends on our sensitivity levels to visual motion and is biased by that prior experience, we may fail to systematically detect such ASD facial speed ranges. In other words, the ASD facial speed ranges may not intersect with our “detection priors” for the ranges of speed that we typically expect. These results suggest that reliance on observation alone is bound to fail in capturing the emotional capacities of the ASD system and miss an opportunity to engage a person in social exchange.

Across all emotional states probed with the standardized micro-movement data type for true personalized assessment, we saw fundamental differences in the ASD facial micro-expressions but the presence of AUs, nevertheless. This suggests systematic engagement with the person providing the instructions to perform the assay and automatic recruitment of AUs on command. In all those brief 5-second tasks, the ASD participant, across all levels of support and spoken abilities, showed the potential to engage in emotional contexts as the facial system recruited relevant AUs, albeit doing so across different levels of intensity and different distribution ranges of intensity values than TD controls. As with the variations in facial speed micro-movements, under those unexpected ranges, the naked eye of an observer, trying to discern emotional states, will surely miss them. This is so because of inherent statistical learning biases and expected values acquired through interactions with other TD people who are most likely operating within those TD ranges.

4.3 Potential for discovery of social–emotional communication in ASD

Given the bidirectional nature of facial micro-expressions, namely, the type of close reafferent loops that engage emotions, it is possible that through training of TD diagnosticians, they can learn to better detect these unexpected ranges of both ASD facial speed micro-movements and ASD AUs’ intensities. It appears that clinicians can detect the two extremes, namely, neurotypical ranges and visible apraxia. They would then engage the ASD individual with greater success than currently done. The results from this investigation are indeed encouraging because they bring a new level of awareness about ASD emotions. Although these levels of intensity and noise are shifted from our perceptual radar, they are present nonetheless in the ASD face. Contrary to the current assumptions and subjective opinions, here, we clearly see that the standardized speed micro-movements present in ASD faces can serve to engage TD controls. As the ASD participants engaged the underlying universal AUs across the different regions of the face grid, we were able to capture these ranges with new analytical means. Moreover, we did so by merely using off-the-shelf instruments that most of us carry around these days.

Our work does not require training large models through machine learning algorithms. Instead, we took direct measurements of speed variability across facial regions evoked by simple, brief assays. Through these unobtrusive, highly scalable means, using commercially available tools on the go, we can further explore other avenues to engage the ASD facial system in unprecedented ways. Bringing these (up to now hidden) ranges of ASD emotions to awareness may make it easier for diagnosticians and therapists engaging with ASD fellows to detect them. Co-adapting our ranges and the ASD ranges is indeed possible now under our new personalized statistical techniques.

This approach to autism is a large departure from trying to impose our neurotypical ways on the ASD person while neglecting the capabilities that their coping systems had already developed by the time that they received the ASD diagnosis. Instead of “normalizing” the ASD individual by imposing our ranges of motion and emotions, here, we propose to expand our perceptual umwelt. By learning to identify the ASD perceptual world and augmenting ours with new expected values, we can also improve our detection systems to operate in a truly diversified humanity. Furthermore, mutual awareness of both our ranges of AUs’ intensity and theirs can bring us closer to building social rapport with the ASD person rather than stigmatizing ASD fellows as social outcasts lacking empathy by assuming that they do not have the desire to engage and communicate with others.

The flip side of augmenting our perceptual umwelt is training ASD individuals to become more aware of their own ranges of micro-motions in the first place. Doing so could help them build self-awareness of their facial micro-expressions and, in this way, own them, learn to control them, and then learn to project desirable configurations at will. Connecting the intent to move with the actual speed of micro-movements could thus become a form of therapeutic intervention mediated by the persons themselves rather than top-down imposed externally by another agent. Owning the action and its consequences can bring the autistic individual to a much-needed level of motor autonomy. Achieving this goal of motor autonomy from the bottom up in autism rather than imposing change from a form of top-down control, e.g., through the currently promoted behavioral conditioning/modification stance, would bring to the autistic person the balance between self-autonomy and self-control (8, 64, 72). In turn, this would enhance their socio-motor agency during social and communication exchanges.

This approach to ASD emotions and social potential is in tune with what autistic individuals themselves want, according to interviews conducted in our lab using the ADOS instrument (40, 73). These observational instruments (the ADOS and the DSM), which deem the ASD person incapable of social communication and boasts a deficit model of social interactions, could in fact benefit from pairing the subjective observational criteria with the type of personalized objective quantification techniques that we show here. Training the diagnosticians in ADOS and DSM settings could significantly improve the diagnostic criteria and help eliminate the stigma that subjective opinion creates across research, clinical settings, and society at large (7476).

4.4 Implications for sensory motor differences related to the brainstem in ASD

The present methods could be applied to screen for signs of brainstem neurodevelopmental differences because the facial regions (ophthalmic, maxillary, and mandibular) are innervated by the trigeminal nerves, which, in turn, connect to the brainstem. In this sense, visual, touch, and auditory stimuli could be used to probe the functioning of such regions in the processing of light, light touch/pressure/movements/pain/temperature, and sound, respectively. Indeed, with respect to sound processing, for instance, retrospective studies have suggested that prolonged auditory brainstem response (ABR) latencies are prevalent in neonates who go on to receive a diagnosis of ASD by 3–4 years of age (58). Furthermore, using clinical observation, the earliest sign predictive of autism is motor delay (77, 78) missed by traditional scales (79) but asserted in 87% of those diagnosed with ASD. These could now be probed non-invasively through the facial involuntary, reflexive, and spontaneous motions.

The facial micro-motions captured here on brief videos could help us track, from a very early age, the development of attention (80, 81), reorienting the head and body for motor control and coordination. Throughout the facial structures and functions involving sensory processing by sensory organs in the eyes, ears, nose, and mouth, we could screen the infant system for unexpected differences indicative of early derailment from the neurotypical developmental path. These sensory organs located on the head sample sensory inputs from different modalities and help form spatio-temporal maps required for the development of proper somatic-sensory-motor integration. These are needed during early dyadic social interactions that develop differently in autism (82). Starting with absent or severely distorted body maps and body schemas (83) and following with delayed milestones in reaching, grasping, postural control, gait, and overall social timing (8488), ASD, which is an umbrella term for many neurodevelopmental disorders (8992), could be screened with the help of non-invasive digital means such as those used in this study.

Another area of importance in autism is pain. States of pain could be ascertained through facial micro-movements, a feature that we have recently found present in autistic individuals at rest (52, 53). Sensory hyper- or hypo-sensitivities are common in ASD (90%) and affect all senses (93), inclusive of kinesthetic, pain and temperature reafference, proprioception (6), interoception, and vestibular integration (94, 95). Some are recognized by the diagnostic criteria, and many can be traced to malfunctioning of the brainstem (96), including sensory motor integration, centrally organized by the superior colliculus, the locus coeruleus, the raphe nucleus, and the olivary systems (97100). Prolonged sensory processing delays are consistent with the findings of extensive literature documenting ABR abnormalities in ASD at various ages. Histological study has found the auditory hindbrain (including the superior olive and inferior colliculus) to have significantly fewer neurons, while surviving neurons have smaller and dysmorphic cell bodies (101, 102). Furthermore, white matter volume in the brainstem of children with ASD is inversely proportional to motor performance (103, 104). Facial features, including the eyes’ gaze direction, head orientation, and mouth-voice analysis, are part of our current studies involving non-speakers requiring high support and deploying the means presented in this paper. These new technologies enable remote use, doing assessments from the comfort of their homes within a participant-centered paradigm. Indeed, the face can be a proxy of states of dysregulation ubiquitous in this population and help us identify ways to regulate the system and relieve it from the cognitive load of having to consciously monitor all aspects of motor control, coordination, and sequencing in activities of daily living that typically occur automatically and largely beneath awareness.

4.5 Caveats and future steps

The subset of participating parents was modest. However, their signatures across the facial universal micro-expressions probed in this study revealed unique characteristics. Their features were closer to the stochastic signatures of their offspring than to those of controls around their age. This result is consistent with prior work in motor control biomarkers involving the upper body, during voluntary pointing behavior (28, 29). In those previous studies from our lab, parents of ASD participants showed motor signatures that separated them from age- and sex-matched TD adult controls—as their standardized speed micro-movement signatures were closer to those of their offspring. As with this previous work, here, we point out the possibilities of genetic overlap and/or motor mimicry. The latter could emerge from the efforts by parents to communicate and build rapport with their children over years of interactions.

The study of 126 participants may seem adequate, but given the broad spectrum of variability in ASD—as shown here for the stochastic ranges that we unveiled—it will be important to considerably scale up this research. The present study provides proof of concept that we can do so using simple and brief apps. In future iterations of this study, we plan to gamify the interactions, so we can collect brief data samples in more naturalistic settings, particularly making it more attractive for children of school age.

Our work offers new analytics that do not require machine learning but rather sample directly the empirical distributions corresponding to a person’s individual levels of facial speed noise. Indeed, we see a whole family of PDFs for each participant, starting with the resting state and sweeping through other distributions for anger, disgust, contempt, happiness, sadness, and surprise. With the exponential rise of autism and its absorption of other developmental disorders, we may have to altogether create more emotions to cover those of the autism spectrum. The universality of those micro-expressions, which Paul Ekman helped define, stops with this side of the human spectrum. As such, we do fail to recognize autistic emotions and have profound deficits in socially engaging and communicating with autistic individuals. How can we close this gap?

5 Conclusions

Discerning such important research questions in future studies may lead us to learn more about new ways to successfully connect and co-adapt with the autistic motor system—as most parents likely have done. Such an approach to autism would also help broaden our own perceptual ranges and build the proper perceptual umwelt to communicate at a very basic level with the ASD world, particularly the world of non-speakers, which remains such a mystery to science.

The means that we introduced in this study can certainly help us deploy large-scale studies to expand on our results and to pursue new questions involving the fascinating world of ASD social communication and emotional exchange.

Note 1. Vanessa Van Edwards’s facial micro-expression description.

Note 2. GitHub site for OpenFace.

Note 3. Action units used in the study from OpenFace.

Data availability statement

The de-dentified digital data from the raw videos supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

The studies involving humans were approved by Rutgers University Institutional Review Board. The studies were conducted in accordance with the local legislation and institutional requirements. Written informed consent for participation in this study was provided by the participants’ legal guardians/next of kin. Written informed consent was obtained from the minor(s)’ legal guardian/next of kin for the publication of any potentially identifiable images or data included in this article. Written informed consent was obtained from the individual(s) for the publication of any identifiable images or data included in this article.

Author contributions

ET: Conceptualization, Data curation, Formal Analysis, Funding acquisition, Investigation, Methodology, Project administration, Resources, Software, Supervision, Validation, Visualization, Writing – original draft, Writing – review & editing. JV: Data curation, Methodology, Resources, Software, Writing – review & editing. ND: Data curation, Software, Visualization, Writing – review & editing, Resources. RR: Data curation, Resources, Writing – review & editing. TB: Investigation, Methodology, Writing – review & editing.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. The Nancy Lurie Marks Family Foundation funded this project under a Career Continuation Award to ET.

Acknowledgments

We thank the participants and their families for contributing to the research. We thank NeuroInversa LLC CEO Christopher Dudick for assisting with data collection and for providing the free face app.

Conflict of interest

ET is a Chief Scientific Officer (CSO) and holds shares in NeuroInversa LLC, a spinoff company of Rutgers University. NeuroInversa LLC provided the research app, free of cost, to collect video data of the face. The app usage and Data User Agreement between NeuroInversa LLC and the Rutgers University Sensory Motor Integration Lab RUSMIL were approved by the Rutgers IRB committee. The research received no funding from NeuroInversa LLC. COI Management plan is in place to supervise this research at the Torres RUSMIL.

The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

The author(s) declared that they were an editorial board member of Frontiers, at the time of submission. This had no impact on the peer review process and the final decision.

Generative AI statement

The author(s) declare that no Generative AI was used in the creation of this manuscript.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpsyt.2025.1559202/full#supplementary-material

References

1. Torres EB. Turning the tables: autism shows the social deficit of our society. In: Torres EB, Whyatt CP, editors. Autism: The Movement Sensing Perspective. CRC Press Taylor and Francis, Boca Raton, London, New York (2018).

Google Scholar

2. Lord C, Risi S, Lambrecht L, Cook EH Jr, Leventhal BL, Dilavore PC, et al. The autism diagnostic observation schedule-generic: a standard measure of social and communication deficits associated with the spectrum of autism. J Autism Dev Disord. (2000) 30:205–23. doi: 10.1023/A:1005592401947

PubMed Abstract | Crossref Full Text | Google Scholar

3. Lord C, Leventhal BL, Cook EH Jr. Quantifying the phenotype in autism spectrum disorders. Am J Med Genet. (2001) 105:36–8. doi: 10.1002/1096-8628(20010108)105:1<36::AID-AJMG1053>3.0.CO;2-4

PubMed Abstract | Crossref Full Text | Google Scholar

4. Constantino JN, Charman T. Diagnosis of autism spectrum disorder: reconciling the syndrome, its diverse origins, and variation in expression. Lancet Neurol. (2016) 15:279–91. doi: 10.1016/S1474-4422(15)00151-9

PubMed Abstract | Crossref Full Text | Google Scholar

5. McCarty P, R.e. F. Early detection and diagnosis of autism spectrum disorder: Why is it so difficult? Semin Pediatr Neurol. (2020) 100831:1–7. doi: 10.1016/j.spen.2020.100831

PubMed Abstract | Crossref Full Text | Google Scholar

6. Brincker M, Torres EB. Chapter 1- why study movement variability in autism. In: Torres EB, Whyatt C, editors. Autism: the movement sensing perspective. CRC Press/Taylor & Francis Group, Boca Raton (2018).

Google Scholar

7. Torres EB. Rethinking the study of volition for clinical use. Adv Exp Med Biol. (2016) 957:229–54. doi: 10.1007/978-3-319-47313-0_13

PubMed Abstract | Crossref Full Text | Google Scholar

8. Torres EB. Autism Autonomy. US: Elsevier Academic Press (2024).

Google Scholar

9. American Psychiatric Association. Diagnostic and statistical manual of mental disorders. 5th ed. Arlington, VA, USA: American Psychiatric Publishing (2013).

Google Scholar

10. Torres EB, Rai R, Mistry S, Gupta B. Hidden aspects of the research ADOS are bound to affect autism science. Neural Comput. (2020) 32:515–61. doi: 10.1162/neco_a_01263

PubMed Abstract | Crossref Full Text | Google Scholar

11. Vogeley K, Bussfeld P, Newen A, Herrmann S, Happe F, Falkai P, et al. Mind reading: neural mechanisms of theory of mind and self-perspective. Neuroimage. (2001) 14:170–81. doi: 10.1006/nimg.2001.0789

PubMed Abstract | Crossref Full Text | Google Scholar

12. Baron-Cohen S, Leslie AM, Frith U. Does the autistic child have a “theory of mind”? Cognition. (1985) 21:37–46. doi: 10.1016/0010-0277(85)90022-8

PubMed Abstract | Crossref Full Text | Google Scholar

13. Baron-Cohen S, Wheelwright S. The empathy quotient: an investigation of adults with Asperger syndrome or high functioning autism, and normal sex differences. J Autism Dev Disord. (2004) 34:163–75. doi: 10.1023/B:JADD.0000022607.19833.00

PubMed Abstract | Crossref Full Text | Google Scholar

14. Lawson J, Baron-Cohen S, Wheelwright S. Empathising and systemising in adults with and without Asperger Syndrome. J Autism Dev Disord. (2004) 34:301–10. doi: 10.1023/B:JADD.0000029552.42724.1b

PubMed Abstract | Crossref Full Text | Google Scholar

15. Charman T, Swettenham J, Baron-Cohen S, Cox A, Baird G, Drew A. Infants with autism: an investigation of empathy, pretend play, joint attention, and imitation. Dev Psychol. (1997) 33:781–9. doi: 10.1037/0012-1649.33.5.781

PubMed Abstract | Crossref Full Text | Google Scholar

16. Lawrence EJ, Shaw P, Baker D, Patel M, Sierra-Siegert M, Medford N, et al. Empathy and enduring depersonalization: the role of self-related processes. Soc Neurosci. (2007) 2:292–306. doi: 10.1080/17470910701391794

PubMed Abstract | Crossref Full Text | Google Scholar

17. Baron-Cohen S. The development of a theory of mind in autism: deviance and delay? Psychiatr Clin North Am. (1991) 14:33–51.

Google Scholar

18. Richell RA, Mitchell DG, Newman C, Leonard A, Baron-Cohen S, Blair RJ. Theory of mind and psychopathy: can psychopathic individuals read the ‘language of the eyes’? Neuropsychologia. (2003) 41:523–6. doi: 10.1016/S0028-3932(02)00175-6

PubMed Abstract | Crossref Full Text | Google Scholar

19. Baron-Cohen S. Empathizing, systemizing, and the extreme male brain theory of autism. Prog Brain Res. (2010) 186:167–75. doi: 10.1016/B978-0-444-53630-3.00011-7

PubMed Abstract | Crossref Full Text | Google Scholar

20. Sharp C, Vanwoerden S. Social cognition: empirical contribution. The developmental building blocks of psychopathic traits: revisiting the role of theory of mind. J Pers Disord. (2014) 28:78–95. doi: 10.1521/pedi.2014.28.1.78

PubMed Abstract | Crossref Full Text | Google Scholar

21. Torres EB. Objective biometric methods for the diagnosis and treatment of nervous system disorders. London: Academic Press (2018).

Google Scholar

22. Torres EB, Brincker M, Isenhower RW, Yanovich P, Stigler KA, Nurnberger JI, et al. Autism: the micro-movement perspective. Front Integr Neurosci. (2013) 7:32. doi: 10.3389/fnint.2013.00032

PubMed Abstract | Crossref Full Text | Google Scholar

23. Torres EB. Atypical signatures of motor variability found in an individual with ASD. Neurocase. (2011) 19:150–65. doi: 10.1080/13554794.2011.654224

PubMed Abstract | Crossref Full Text | Google Scholar

24. Jones W, Klaiman C, Richardson S, Aoki C, Smith C, Minjarez M, et al. Eye-tracking-based measurement of social visual engagement compared with expert clinical diagnosis of autism. JAMA. (2023) 330:854–65. doi: 10.1001/jama.2023.13295

PubMed Abstract | Crossref Full Text | Google Scholar

25. Jones W, Klaiman C, Richardson S, Lambha M, Reid M, Hamner T, et al. Development and replication of objective measurements of social visual engagement to aid in early diagnosis and assessment of autism. JAMA Netw Open. (2023) 6:e2330145. doi: 10.1001/jamanetworkopen.2023.30145

PubMed Abstract | Crossref Full Text | Google Scholar

26. Aikat V, Krishnappa Babu PR, Carpenter KLM, Di Martino JM, Espinosa S, Davis N, et al. (2024). Digital phenotyping based on A mobile app identifies distinct and overlapping features in children diagnosed with autism versus adhd, in: Uist Adjunct 2024 - Proceedings Of The 37th Annual Acm Symposium On User Interface Software and Technology.

Google Scholar

27. Torres EB, Isenhower RW, Yanovich P, Rehrig G, Stigler K, Nurnberger J, et al. Strategies to develop putative biomarkers to characterize the female phenotype with autism spectrum disorders. J Neurophysiol. (2013) 110:1646–62. doi: 10.1152/jn.00059.2013

PubMed Abstract | Crossref Full Text | Google Scholar

28. Torres EB, Isenhower RW, Nguyen J, Whyatt C, Nurnberger JI, Jose JV, et al. Toward precision psychiatry: statistical platform for the personalized characterization of natural behaviors. Front Neurol. (2016) 7:8. doi: 10.3389/fneur.2016.00008

PubMed Abstract | Crossref Full Text | Google Scholar

29. Wu D, Jose JV, Nurnberger JI, Torres EB. A biomarker characterizing neurodevelopment with applications in autism. Sci Rep. (2018) 8:614. doi: 10.1038/s41598-017-18902-w

PubMed Abstract | Crossref Full Text | Google Scholar

30. Bermperidis T, Rai R, Ryu J, Zanotto D, Agrawal SK, Lalwani AK, et al. Optimal time lags from causal prediction model help stratify and forecast nervous system pathology. Sci Rep. (2021) 11:20904. doi: 10.1038/s41598-021-00156-2

PubMed Abstract | Crossref Full Text | Google Scholar

31. An WW, Bhowmik AC, Nelson CA, Wilkinson CL. Eeg-based brain age prediction in infants-toddlers: implications for early detection of neurodevelopmental disorders. Dev Cognit Neurosci. (2024) 71:101493. doi: 10.1016/j.dcn.2024.101493

PubMed Abstract | Crossref Full Text | Google Scholar

32. Bosl W, Tierney A, Tager-Flusberg H, Nelson C. EEG complexity as a biomarker for autism spectrum disorder risk. BMC Med. (2011) 9:18. doi: 10.1186/1741-7015-9-18

PubMed Abstract | Crossref Full Text | Google Scholar

33. Carson WET, Major S, Akkineni H, Fung H, Peters E, Carpenter KLH, et al. Model selection to achieve reproducible associations between resting state EEG features and autism. Sci Rep. (2024) 14:25301. doi: 10.1038/s41598-024-76659-5

PubMed Abstract | Crossref Full Text | Google Scholar

34. Tsiaras V, Simos PG, Rezaie R, Sheth BR, Garyfallidis E, Castillo EM, et al. Extracting biomarkers of autism from MEG resting-state functional connectivity networks. Comput Biol Med. (2011) 41:1166–77. doi: 10.1016/j.compbiomed.2011.04.004

PubMed Abstract | Crossref Full Text | Google Scholar

35. Torres EB, Denisova K. Motor noise is rich signal in autism research and pharmacological treatments. Sci Rep. (2016) 6:37422. doi: 10.1038/srep37422

PubMed Abstract | Crossref Full Text | Google Scholar

36. Torres EB, Mistry S, Caballero C, Whyatt CP. Stochastic signatures of involuntary head micro-movements can be used to classify females of ABIDE into different subtypes of neurodevelopmental disorders. Front Integr Neurosci. (2017) 11:10. doi: 10.3389/fnint.2017.00010

PubMed Abstract | Crossref Full Text | Google Scholar

37. Caballero C, Mistry S, Vero J, Torres EB. Characterization of noise signatures of involuntary head motion in the autism brain imaging data exchange repository. Front Integr Neurosci. (2018) 12:7. doi: 10.3389/fnint.2018.00007

PubMed Abstract | Crossref Full Text | Google Scholar

38. Caballero C, Mistry S, Torres EB. Age-dependent statistical changes of involuntary head motion signatures across autism and controls of the ABIDE repository. Front Integr Neurosci. (2020) 14:1–14. doi: 10.3389/fnint.2020.00023

PubMed Abstract | Crossref Full Text | Google Scholar

39. Torres EB, Caballero C, Mistry S. Aging with autism departs greatly from typical aging. Sensors (Basel). (2020) 20:1–21. doi: 10.3390/s20020572

PubMed Abstract | Crossref Full Text | Google Scholar

40. Torres EB, Whyatt C. Autism: the movement sensing perspective. Boca Raton: CRC Press/Taylor & Francis Group (2018).

Google Scholar

41. Damasio AR, Maurer RG. A neurological model for childhood autism. Arch Neurol. (1978) 35:777–86. doi: 10.1001/archneur.1978.00500360001001

PubMed Abstract | Crossref Full Text | Google Scholar

42. Floris DL, Barber AD, Nebel MB, Martinelli M, Lai MC, Crocetti D, et al. Atypical lateralization of motor circuit functional connectivity in children with autism is associated with motor deficits. Mol Autism. (2016) 7:35. doi: 10.1186/s13229-016-0096-6

PubMed Abstract | Crossref Full Text | Google Scholar

43. Mosconi MW, Mohanty S, Greene RK, Cook EH, Vaillancourt DE, Sweeney JA. Feedforward and feedback motor control abnormalities implicate cerebellar dysfunctions in autism spectrum disorder. J Neurosci. (2015) 35:2015–25. doi: 10.1523/JNEUROSCI.2731-14.2015

PubMed Abstract | Crossref Full Text | Google Scholar

44. Mosconi MW, Sweeney JA. Sensorimotor dysfunctions as primary features of autism spectrum disorders. Sci China Life Sci. (2015) 58:1016–23. doi: 10.1007/s11427-015-4894-4

PubMed Abstract | Crossref Full Text | Google Scholar

45. Mostofsky SH, Burgess MP, Gidley Larson JC. Increased motor cortex white matter volume predicts motor impairment in autism. Brain. (2007) 130:2117–22. doi: 10.1093/brain/awm129

PubMed Abstract | Crossref Full Text | Google Scholar

46. Mostofsky SH, Powell SK, Simmonds DJ, Goldberg MC, Caffo B, Pekar JJ. Decreased connectivity and cerebellar activity in autism during motor task performance. Brain. (2009) 132:2413–25. doi: 10.1093/brain/awp088

PubMed Abstract | Crossref Full Text | Google Scholar

47. Muller RA, Pierce K, Ambrose JB, Allen G, Courchesne E. Atypical patterns of cerebral motor activation in autism: a functional magnetic resonance study. Biol Psychiatry. (2001) 49:665–76. doi: 10.1016/S0006-3223(00)01004-0

PubMed Abstract | Crossref Full Text | Google Scholar

48. Theoret H, Halligan E, Kobayashi M, Fregni F, Tager-Flusberg H, Pascual-Leone A. Impaired motor facilitation during action observation in individuals with autism spectrum disorder. Curr Biol. (2005) 15:R84–5. doi: 10.1016/j.cub.2005.01.022

PubMed Abstract | Crossref Full Text | Google Scholar

49. Hoffman T, Bar-Shalita T, Granovsky Y, Gal E, Kalingel-Levi M, Dori Y, et al. Indifference or hypersensitivity? Solving the riddle of the pain profile in individuals with autism. Pain. (2022) 791–803. doi: 10.1097/j.pain.0000000000002767

PubMed Abstract | Crossref Full Text | Google Scholar

50. Kalingel-Levi M, Schreuer N, Granovsky Y, Bar-Shalita T, Weissman-Fogel I, Hoffman T, et al. When I’m in pain, everything is overwhelming”: implications of pain in adults with autism on their daily living and participation. Front Psychol. (2022) 13:911756. doi: 10.3389/fpsyg.2022.911756

PubMed Abstract | Crossref Full Text | Google Scholar

51. Levi MK, Schreuer N, Granovsky Y, Bar-Shalita T, Fogel IW, Hoffman T, et al. Feeling unwanted, when nobody wants you around”: perceptions of social pain among people with autism. Am J Occup Ther. (2023) 77. doi: 10.5014/ajot.2023.050061

PubMed Abstract | Crossref Full Text | Google Scholar

52. Elsayed M, Torres EB. Exploring cardiac responses of pain and distress. In: Topics in Autonomic Nervous System. IntechOpen (2023). doi: 10.5772/intechopen.111890

Crossref Full Text | Google Scholar

53. Elsayed M. Framework of interpretable biometrics to assess internal psychophysiological states of distress in autism and the general population. PhD Thesis, (PhD Thesis). Rutgers University, New Brunswick, NJ, US (2024).

Google Scholar

54. Baltrušaitis T, Robinson P, Morency LP. (2016). Openface: an open source facial behavior analysis toolkit, in: Ieee Winter Conference On Applications Of Computer Vision (Wacv), Lake Placid, NY, USA.

Google Scholar

55. Cao Z, Hidalgo G, Simon T, Wei SE, Sheikh YA. OpenPose: realtime multi-person 2D pose estimation using part affinity fields. arXvi. (2019) 1–14. doi: 10.48550/arXiv.1812.08008

PubMed Abstract | Crossref Full Text | Google Scholar

56. Anzulewicz A, Sobota K, Delafield-Butt JT. Toward the autism motor signature: gesture patterns during smart tablet gameplay identify children with autism. Sci Rep. (2016) 6:31107. doi: 10.1038/srep31107

PubMed Abstract | Crossref Full Text | Google Scholar

57. Ekman P. Emotion in the human face. In: Editions de la Maison des Sciences de l’Homme. Cambridge University Press, Cambridge Cambridgeshire; New York, Paris (1982).

Google Scholar

58. Torres EB, Varkey H, Vero J, London E, Phan H, Kittler P, et al. Sensing echoes: temporal misalignment in auditory brainstem responses as the earliest marker of neurodevelopmental derailment. PNAS Nexus. (2023) 2:pgac315. doi: 10.1093/pnasnexus/pgac315

PubMed Abstract | Crossref Full Text | Google Scholar

59. Mosimann JE. Size allometry: size and shape variables with characterizations of the lognormal and generalized gamma distributions. J Am Stat Assoc. (1970) 65:930–45. doi: 10.1080/01621459.1970.10481136

Crossref Full Text | Google Scholar

60. Torres EB. Methods for the diagnosis and treatment of neurological disorders. US patent. Alexandria, VA: USPTO, United States Patent and Trademark Office (2018).

Google Scholar

61. Torres EB. System and method for determining amount of volition in a subject. US patent. USPTO, United States Patent and Trademark Office, Alexandria, VA (2017).

Google Scholar

62. Torres EB. System and Method for measuring physiologically relevant motion. US patent. Alexandria, VA: USPTO, United States Patent and Trademark Office (2017).

Google Scholar

63. Torres EB. Systems and Methods for Tracking Neurodevelopment Disorders. US patent. Alexandria, VA: USPTO, United States Patent and Trademark Office(2019).

Google Scholar

64. Bermperidis T, Rai R, Torres EB. Digital screener of socio-motor agency balancing motor autonomy and motor control. Front Hum Neurosci. (2024) 18:1442799. doi: 10.3389/fnhum.2024.1442799

PubMed Abstract | Crossref Full Text | Google Scholar

65. Rubner Y, Tomasi C, Guibas LJ. (1998). Metric for distributions with applications to image databases, in: Proceedings of the ICCV, Bombay, India, 4-7 January, 1998.

Google Scholar

66. Stolfi J, Guibas LJ. The earth mover’s distance as a metric for image retrieval. Int J Comput Vis. (2000) 40:99–121.

Google Scholar

67. Levenshtein VI. Binary codes capable of correcting deletions, insertions, and reversals. Soviet Phys Doklady. (1965) 10:707–10.

Google Scholar

68. Ziv I, Avni I, Dinstein I, Meiri G, Bonneh YS. Oculomotor randomness is higher in autistic children and increases with the severity of symptoms. Autism Res. (2024) 197–201. doi: 10.1002/aur.v17.2

PubMed Abstract | Crossref Full Text | Google Scholar

69. Bockadia H, Cole J, Torres EB. (2020). Neural Connectivity Evolution during Adaptive Learning with and without Proprioception, in: Proceedings of the 7th International Symposium on Movement and Computing, Jersey City, NJ. ACM.

Google Scholar

70. Torres EB, Vero J, Rai R. Statistical platform for individualized behavioral analyses using biophysical micro-movement spikes. Sensors (Basel). (2018) 18:1–20. doi: 10.3390/s18041025

PubMed Abstract | Crossref Full Text | Google Scholar

71. Ryu J, Torres EB. Motor signatures in digitized cognitive and memory tests enhances characterization of parkinson’s disease. Sensors (Basel). (2022) 22:1–15. doi: 10.3390/s22124434

PubMed Abstract | Crossref Full Text | Google Scholar

72. Bermperidis T. A unifying theoretical and empirical framework to study disorders of the nervous systems from molecules to complex social behaviors. (PhD Thesis). Rutgers University, New Brunswick, NJ, US (2024).

Google Scholar

73. Torres EB. Critical ingredients for proper social interactions: rethinking the mirror neuron system theory. In: Objective Biometric Methods for the Diagnosis and Treatment of Nervous System Disorders, 1st ed. Boca Raton, FL, USA: Elsevier (2018).

Google Scholar

74. Gurba AN, Mcnair ML, Hargreaves A, Scheerer NE, Ng CSM, Lerner MD. Editorial: Break the stigma: autism. The future of research on autism stigma - towards multilevel, contextual & global understanding. Front Psychiatry. (2024) 15:1504429. doi: 10.3389/fpsyt.2024.1504429

PubMed Abstract | Crossref Full Text | Google Scholar

75. Hotez E, Phan JM, Truong DM. Addressing stigma-related health disparities for autistic individuals through cultural competemility: insights from research and lived experience. Curr Psychiatry Rep. (2024) 761–70. doi: 10.1007/s11920-024-01551-y

PubMed Abstract | Crossref Full Text | Google Scholar

76. Scheerer NE, Ng CSM, Gurba AN, Mcnair ML, Lerner MD, Hargreaves A. Editorial: Break the stigma: autism. Front Psychiatry. (2024) 15:1513447. doi: 10.3389/fpsyt.2024.1513447

PubMed Abstract | Crossref Full Text | Google Scholar

77. LeBarton ES, Landa RJ. Infant motor skill predicts later expressive language and autism spectrum disorder diagnosis. Infant Behav Dev. (2019) 54:37–47. doi: 10.1016/j.infbeh.2018.11.003

PubMed Abstract | Crossref Full Text | Google Scholar

78. Leonard HC, Bedford R, Pickles A, Hill EL, Team B. Predicting the rate of language development from early motor skills in at-risk infants who develop autism spectrum disorder. Res Autism Spectr Disord. (2015) 13-14:15–24. doi: 10.1016/j.rasd.2014.12.012

PubMed Abstract | Crossref Full Text | Google Scholar

79. Posar A, Visconti P. Early motor signs in autism spectrum disorder. Children (Basel). (2022) 9:1–18. doi: 10.3390/children9020294

PubMed Abstract | Crossref Full Text | Google Scholar

80. Libertus K, Landa RJ, Haworth JL. Development of attention to faces during the first 3 years: influences of stimulus type. Front Psychol. (2017) 8:1976. doi: 10.3389/fpsyg.2017.01976

PubMed Abstract | Crossref Full Text | Google Scholar

81. Shephard E, Milosavljevic B, Mason L, Elsabbagh M, Tye C, Gliga T, et al. Neural and behavioural indices of face processing in siblings of children with autism spectrum disorder (ASD): A longitudinal study from infancy to mid-childhood. Cortex. (2020) 127:162–79. doi: 10.1016/j.cortex.2020.02.008

PubMed Abstract | Crossref Full Text | Google Scholar

82. Landa RJ, Haworth JL, Nebel MB. Ready, Set, Go! Low Anticipatory Response during a Dyadic Task in Infants at High Familial Risk for Autism. Front Psychol. (2016) 7:721. doi: 10.3389/fpsyg.2016.00721

PubMed Abstract | Crossref Full Text | Google Scholar

83. Bick NA, Redfern MS, Jennings JR, Eack SM, Iverson JM, Cham R. Attention and sensory integration for gait in young adults with autism spectrum disorder. Gait Posture. (2024) 112:74–80. doi: 10.1016/j.gaitpost.2024.04.035

PubMed Abstract | Crossref Full Text | Google Scholar

84. Lidstone DE, Mostofsky SH. Moving toward understanding autism: visual-motor integration, imitation, and social skill development. Pediatr Neurol. (2021) 122:98–105. doi: 10.1016/j.pediatrneurol.2021.06.010

PubMed Abstract | Crossref Full Text | Google Scholar

85. Kwakye LD, Foss-Feig JH, Cascio CJ, Stone WL, Wallace MT. Altered auditory and multisensory temporal processing in autism spectrum disorders. Front Integr Neurosci. (2011) 4:129. doi: 10.3389/fnint.2010.00129

PubMed Abstract | Crossref Full Text | Google Scholar

86. Regener P, Heffer N, Love SA, Petrini K, Pollick F. Differences in audiovisual temporal processing in autistic adults are specific to simultaneity judgments. Autism Res. (2024) 17:1041–52. doi: 10.1002/aur.v17.5

PubMed Abstract | Crossref Full Text | Google Scholar

87. Shen L, Li S, Tian Y, Wang Y, Jiang Y. Cortical tracking of hierarchical rhythms orchestrates the multisensory processing of biological motion. Elife. (2025) 13:1–28. doi: 10.7554/eLife.98701.4

PubMed Abstract | Crossref Full Text | Google Scholar

88. Suri KN, Whedon M, Lewis M. Perception of audio-visual synchrony in infants at elevated likelihood of developing autism spectrum disorder. Eur J Pediatr. (2023) 182:2105–17. doi: 10.1007/s00431-023-04871-y

PubMed Abstract | Crossref Full Text | Google Scholar

89. Bhat AN. Motor impairment increases in children with autism spectrum disorder as a function of social communication, cognitive and functional impairment, repetitive behavior severity, and comorbid diagnoses: A SPARK study report. Autism Res. (2021) 14:202–19. doi: 10.1002/aur.v14.1

PubMed Abstract | Crossref Full Text | Google Scholar

90. Franchini M, Smith IM, Sacrey L, Duku E, Brian J, Bryson SE, et al. Continuity of trajectories of autism symptom severity from infancy to childhood. J Child Psychol Psychiatry. (2023) 64:895–906. doi: 10.1111/jcpp.13744

PubMed Abstract | Crossref Full Text | Google Scholar

91. Hocking DR, Menant JC, Kirk HE, Lord S, Porter MA. Gait profiles as indicators of domain-specific impairments in executive control across neurodevelopmental disorders. Res Dev Disabil. (2014) 35:203–14. doi: 10.1016/j.ridd.2013.10.005

PubMed Abstract | Crossref Full Text | Google Scholar

92. Patterson JW, Armstrong V, Duku E, Richard A, Franchini M, Brian J, et al. Early trajectories of motor skills in infant siblings of children with autism spectrum disorder. Autism Res. (2022) 15:481–92. doi: 10.1002/aur.v15.3

PubMed Abstract | Crossref Full Text | Google Scholar

93. Robertson CE, Baron-Cohen S. Sensory perception in autism. Nat Rev Neurosci. (2017) 18:671–84. doi: 10.1038/nrn.2017.112

PubMed Abstract | Crossref Full Text | Google Scholar

94. Lim YH, Partridge K, Girdler S, Morris SL. Standing postural control in individuals with autism spectrum disorder: systematic review and meta-analysis. J Autism Dev Disord. (2017) 47:2238–53. doi: 10.1007/s10803-017-3144-y

PubMed Abstract | Crossref Full Text | Google Scholar

95. Travers BG, Mason A, Gruben KG, Dean DC 3rd, Mclaughlin K. Standing balance on unsteady surfaces in children on the autism spectrum: the effects of IQ. Res Autism Spectr Disord. (2018) 51:9–17. doi: 10.1016/j.rasd.2018.03.008

PubMed Abstract | Crossref Full Text | Google Scholar

96. Dadalko OI, Travers BG. Evidence for brainstem contributions to autism spectrum disorders. Front Integr Neurosci. (2018) 12:47. doi: 10.3389/fnint.2018.00047

PubMed Abstract | Crossref Full Text | Google Scholar

97. Jure R. Autism pathogenesis: the superior colliculus. Front Neurosci. (2018) 12:1029. doi: 10.3389/fnins.2018.01029

PubMed Abstract | Crossref Full Text | Google Scholar

98. Jure R. The “Primitive brain dysfunction” Theory of autism: the superior colliculus role. Front Integr Neurosci. (2022) 16:797391. doi: 10.3389/fnint.2022.797391

PubMed Abstract | Crossref Full Text | Google Scholar

99. Kulesza RJ Jr, Lukose R, Stevens LV. Malformation of the human superior olive in autistic spectrum disorders. Brain Res. (2011) 1367:360–71. doi: 10.1016/j.brainres.2010.10.015

PubMed Abstract | Crossref Full Text | Google Scholar

100. London EB. Neuromodulation and a reconceptualization of autism spectrum disorders: using the locus coeruleus functioning as an exemplar. Front Neurol. (2018) 9:1120. doi: 10.3389/fneur.2018.01120

PubMed Abstract | Crossref Full Text | Google Scholar

101. Mansour Y, Burchell A, Kulesza R. Abnormal vestibular brainstem structure and function in an animal model of autism spectrum disorder. Brain Res. (2022) 1793:148056. doi: 10.1016/j.brainres.2022.148056

PubMed Abstract | Crossref Full Text | Google Scholar

102. Smith A, Storti S, Lukose R, Kulesza RJ Jr. Structural and functional aberrations of the auditory brainstem in autism spectrum disorder. J Am Osteopath Assoc. (2019) 119:41–50. doi: 10.7556/jaoa.2019.007

PubMed Abstract | Crossref Full Text | Google Scholar

103. Hanaie R, Mohri I, Kagitani-Shimono K, Tachibana M, Matsuzaki J, Hirata I, et al. White matter volume in the brainstem and inferior parietal lobule is related to motor performance in children with autism spectrum disorder: A voxel-based morphometry study. Autism Res. (2016) 9:981–92. doi: 10.1002/aur.2016.9.issue-9

PubMed Abstract | Crossref Full Text | Google Scholar

104. Thomason ME, Hect J, Waller R, Manning JH, Stacks AM, Beeghly M, et al. Prenatal neural origins of infant motor development: Associations between fetal brain and infant motor development. Dev Psychopathol. (2018) 30:763–72. doi: 10.1017/S095457941800072X

PubMed Abstract | Crossref Full Text | Google Scholar

Keywords: facial micro-expressions, autism spectrum disorders, emotions, stochastic analysis, motor control, automatic screening

Citation: Torres EB, Vero J, Drain N, Rai R and Bermperidis T (2025) Hidden social and emotional competencies in autism spectrum disorders captured through the digital lens. Front. Psychiatry 16:1559202. doi: 10.3389/fpsyt.2025.1559202

Received: 12 January 2025; Accepted: 28 February 2025;
Published: 07 April 2025.

Edited by:

Oswald David Kothgassner, Medical University of Vienna, Austria

Reviewed by:

Eleni Mitsea, National Centre of Scientific Research Demokritos, Greece
Zhe Chen, Jiangnan University, China

Copyright © 2025 Torres, Vero, Drain, Rai and Bermperidis. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Elizabeth B. Torres, ZWJ0b3JyZXNAcHN5Y2gucnV0Z2Vycy5lZHU=

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.