- 1State Key Laboratory of Trauma and Chemical Poisoning, Institute of Combined Injury, Chongqing Engineering Research Center for Nanomedicine, College of Preventive Medicine, Army Medical University, Chongqing, China
- 2Department of Orthopedics, Xinqiao Hospital, Army Medical University, Chongqing, China
- 3Department of Rehabilitation Medicine, The First Affiliated Hospital, Sun Yat-sen University, Guangzhou, China
- 4School of Medicine, Sun Yat-sen University, Shenzhen, China
Background: Single-cell multi-omics technologies capture cellular heterogeneity at unprecedented resolution, yet dimensionality reduction methods face a fundamental local–global trade-off: approaches optimized for local neighborhood preservation distort global topology, while those emphasizing global coherence obscure fine-grained cell states.
Results: We introduce the Lorentz-regularized variational autoencoder (LiVAE), a dual-pathway architecture that applies hyperbolic geometry as soft regularization over standard Euclidean latent spaces. A primary encoding pathway preserves local transcriptional details for high-fidelity reconstruction, while an information bottleneck (BN) pathway extracts global hierarchical structure by filtering technical noise. Lorentzian distance constraints enforce geometric consistency between pathways in hyperbolic space, enabling LiVAE to balance local fidelity with global coherence without requiring specialized batch-correction procedures. Systematic benchmarking across 135 datasets against 21 baseline methods demonstrated that LiVAE achieves superior global topology preservation (distance correlation gains: 0.209–0.436), richer latent geometry (manifold dimensionality: 0.123–0.467; participation ratio: 0.149–0.761), and enhanced robustness (noise resilience: 0.184–0.712) while maintaining competitive local fidelity. The overall embedding quality improved by 0.051–0.284 across uniform manifold approximation and projection (UMAP) and t-distributed stochastic neighbor embedding (t-SNE) visualizations. Component-wise interpretability analysis on a Dapp1 perturbation dataset revealed biologically meaningful latent axes.
Conclusion: LiVAE provides a robust, general-purpose framework for single-cell representation learning that resolves the local–global trade-off through geometric regularization. By maintaining Euclidean latent spaces while leveraging hyperbolic priors, LiVAE enables improved developmental trajectory inference and mechanistic biological discovery without sacrificing compatibility with existing computational ecosystems.
1 Introduction
Cellular development unfolds through hierarchical differentiation programs where stem cells progressively commit to specialized fates (Trapnell et al., 2014). Single-cell multi-omics technologies now capture these hierarchies at an unprecedented resolution (Stuart et al., 2019); however, representing tree-like developmental structures computationally remains an open challenge (Luecken et al., 2022). Datasets routinely contain
This challenge manifests as a fundamental local–global trade-off: representations must capture fine-grained local neighborhoods for rare cell-type detection (Heiser and Lau, 2020; Kiselev et al., 2019) while maintaining global topology for developmental trajectory inference (Cao et al., 2019; Saelens et al., 2019). Despite advances in deep learning (Hu et al., 2021; Yuan and Kelley, 2022) and foundation models (Cui et al., 2024), most methods excel at one scale at the expense of the other. The tension reflects a geometric limitation of Euclidean spaces: methods optimized for local structure, such as t-distributed stochastic neighbor embedding (t-SNE) (Kobak and Berens, 2019) and uniform manifold approximation and projection (UMAP) (McInnes et al., 2018), distort global topology, while those prioritizing global coherence, such as principal component analysis (PCA) and diffusion maps (Moon et al., 2018; Becht et al., 2019), may obscure fine-grained cell states. Graph-based approaches (Hetzel et al., 2021; Nguyen et al., 2022) and gene co-expression modeling (Deng et al., 2025; Li et al., 2025; Song T. et al., 2022) partially address this by explicitly encoding functional relationships, yet systematic benchmarking reveals that most methods still sacrifice one scale for the other (Tian et al., 2019).
These challenges intensify across modalities: single-cell ATAC sequencing (scATAC-seq) exhibits 90%–95% zero rates versus 60%–80% in single-cell RNA sequencing (scRNA-seq) (Chen et al., 2019; Fang et al., 2021), thus requiring flexible architectures that generalize without extensive re-engineering (Song Q. et al., 2022; Zhao et al., 2024). Hyperbolic geometry offers a principled solution as its exponential volume growth naturally accommodates tree-like hierarchies common in developmental biology (Nickel and Kiela, 2017; Chami et al., 2019; Sarkar, 2012; Bronstein et al., 2021). Existing hyperbolic deep learning methods have improved visualization and captured cellular relationships (Tian et al., 2023; Klimovskaia et al., 2020), but they constrain entire latent spaces to hyperbolic manifolds (Mathieu et al., 2019; Nagano et al., 2019), thus sacrificing compatibility with standard neural architectures and downstream analytical tools. This failure stems from representational constraints: embedding
We introduce the Lorentz-regularized variational autoencoder (LiVAE), which applies hyperbolic geometry as regularization over standard Euclidean latent representations rather than constraining the latent space itself. LiVAE learns a primary embedding
We validate LiVAE on 135 datasets spanning scRNA-seq and scATAC-seq against 21 baseline methods using 12 metrics assessing embedding fidelity, manifold geometry, and robustness. LiVAE consistently achieves superior global topology preservation and noise resilience while remaining competitive on local structure metrics. Component-wise interpretability analysis demonstrates that latent dimensions decompose into biologically meaningful axes corresponding to cell cycle, immune identity, and differentiation programs (Choi et al., 2023; Madrigal et al., 2024).
Our contributions are threefold: (1) architectural innovation: a hybrid design applying hyperbolic regularization to Euclidean representations via information bottlenecks, balancing local fidelity with global coherence; (2) cross-modality generalization: a unified framework handling scRNA-seq and scATAC-seq through modality-appropriate likelihoods without architectural changes; and (3) biological interpretability: latent dimensions aligned with known biological processes that enable mechanistic hypothesis generation beyond black-box embeddings. By resolving the local–global trade-off through geometric regularization, LiVAE provides a flexible foundation for single-cell multi-omics analysis that preserves biological hierarchy without sacrificing compatibility with existing computational workflows.
2 Materials and methods
2.1 Notation
Throughout this section,
2.2 LiVAE architecture overview
LiVAE is a variational autoencoder that applies Lorentzian geometric regularization across a dual-pathway latent space architecture (Figure 1). The encoder
Figure 1. LiVAE architecture. Input
The model is trained via total loss
2.3 Model architecture
2.3.1 Encoder network
For each cell
and
2.3.2 Dual latent pathways and decoder
The bottleneck path applies two linear transformations: compression
The decoder mirrors the encoder architecture, outputting distribution parameters via linear layers followed by softmax normalization. It generates two reconstructions:
• Primary reconstruction:
• Bottleneck reconstruction:
2.4 Loss function
The total loss function (Equation 2) is a weighted sum of four components:
where
2.4.1 Reconstruction losses
The reconstruction losses (Equation 3) measure how well each pathway captures the input:
The likelihood
• scRNA-seq: negative binomial (NB) distribution with the mean
• scATAC-seq: zero-inflated negative binomial (ZINB) with additional zero-inflation probability
• Alternative: Poisson and zero-inflated Poisson (ZIP) likelihoods are also supported for datasets with minimal overdispersion.
For all likelihoods, the predicted means are obtained as
2.4.2 Kullback–Leibler divergence
The KL divergence is the standard variational autoencoder (VAE) regularizer that encourages the posterior to approximate a unit Gaussian prior (Equation 4):
2.4.3 Geometric loss
The geometric loss enforces that the bottleneck transformation preserves the hyperbolic geometric structure. Euclidean vectors
where
The geometric loss is the mean squared Lorentzian distance between paired representations (Equation 6):
where
with the Lorentzian inner product
2.5 Evaluation metrics
We assess LiVAE performance using 12 metrics organized into four categories, evaluating complementary aspects of representation quality.
2.5.1 Clustering quality metrics
We assess the biological population structure using five standard metrics and one novel metric:
Standard metrics: Normalized mutual information (NMI) and adjusted Rand index (ARI) measure agreement between the predicted clusters and ground-truth cell-type labels, with values near 1 indicating strong correspondence. Average silhouette width (ASW) and the Calinski–Harabasz index (CAL) quantify cluster cohesion and separation (higher is better), while the Davies–Bouldin index (DAV) measures the average cluster similarity (lower is better). These metrics are computed using standard implementations in scikit-learn.
Coupling degree (COR) (Equation 8): We introduce this metric to quantify preservation of interdependent biological programs:
where
2.5.2 Dimensionality reduction embedding quality metrics
We evaluate how effectively latent representations
Distance correlation
where
Local quality
where
Overall embedding quality
2.5.3 Intrinsic manifold quality metrics
We characterize geometric properties of the latent manifold through spectral analysis of the covariance matrix
Manifold dimensionality
where
Spectral decay rate
where
Participation ratio
Higher values indicate more uniform utilization of latent dimensions, thereby preventing dimension collapse.
Anisotropy score
where
Trajectory directionality
Higher values indicate a single dominant trajectory, which is characteristic of linear differentiation processes.
Noise resilience
Higher values indicate robust separation between signal and noise subspaces.
Composite scores.
We define two summary metrics: core intrinsic quality (Equation 18) integrates the fundamental geometric properties,
while overall intrinsic quality (Equation 19) incorporates task-oriented components with weights
2.5.4 Batch integration quality
The integration local inverse Simpson index (iLISI) (Equation 20) measures batch mixing quality. For each cell
where
2.6 Datasets and preprocessing
2.6.1 Dataset selection
We curated 135 single-cell datasets from public repositories (Gene Expression Omnibus, GEO): 53 scRNA-seq dataset and 82 scATAC-seq dataset samples. Raw single-cell count matrices underwent quality control and normalization prior to model training. Both modalities require raw integer counts as input as the model employs count-based likelihood functions.
2.6.2 scRNA-seq preprocessing
The top 5,000 highly variable genes (HVGs) were selected by modeling the mean–variance relationship in count data. For model input, normalized data were obtained by applying
2.6.3 scATAC-seq preprocessing
Term frequency–inverse document frequency (TF-IDF) normalization (Equation 21) was applied:
where the term frequency for cell
2.7 Model hyperparameters
LiVAE was configured with the following default hyperparameters. The encoder and decoder networks each contained a single hidden layer of dimension 128. The latent space dimension was set to
2.8 Baseline methods
We compared LiVAE against 21 methods spanning four categories:
• Classical dimensionality reduction (seven methods): PCA, kernel PCA (KPCA), factor analysis (FA), non-negative matrix factorization (NMF), independent component analysis (ICA), truncated SVD (TSVD), and dictionary learning (DICL).
• Deep generative models (eight methods): standard VAE,
• Graph and contrastive learning (three methods): contrastive learning for scRNA-seq (CLEAR), single-cell graph neural network (scGNN), and single-cell graph contrastive clustering (scGCC).
• Modality-specific methods (three methods): latent semantic indexing (LSI), peak variational inference (PeakVI), and Poisson variational inference (PoissonVI) for scATAC-seq.
2.9 Statistical analysis
We used paired experimental designs (identical datasets for all methods). For each metric, normality was assessed using the Shapiro–Wilk test
3 Results
3.1 Architectural progression from foundational VAEs yields comprehensive performance gains
We benchmarked LiVAE against its foundational predecessors—standard VAE and information bottleneck VAE (iVAE)—using 135 datasets (53 scRNA-seq and 82 scATAC-seq). LiVAE’s complete architecture established a new performance baseline, delivering statistically significant improvements across nearly all metrics for both scRNA-seq (Figure 2A) and scATAC-seq datasets (Figure 2B; Table 1).
Figure 2. Progressive architectural enhancements yield consistent performance gains. Boxplots display performance differences (
Table 1. Performance differences between LiVAE and baseline VAE models across scRNA-seq
The most striking advantage was a profound increase in model robustness. For scRNA-seq (Figure 2A), LiVAE boosted noise resilience by
This enhanced robustness stemmed from LiVAE’s geometrically expressive latent space. For scRNA-seq, the participation ratio increased by
3.2 Balanced profile of local fidelity and global structure against classical methods
We benchmarked LiVAE against seven classical algorithms on
Table 2. Performance differences between LiVAE and classical dimensionality reduction methods across
LiVAE’s UMAP local quality (Figure 3A) was statistically equivalent to those of PCA, KPCA, FA, NMF, and TSVD but slightly lower than that of dictionary learning (DICL;
Figure 3. Balanced local–global performance relative to classical dimensionality reduction. Boxplots summarize metric distributions for LiVAE and seven classical baselines across
This superior organization reflected a sophisticated latent space (Figure 3C). All four manifold metrics improved substantially: dimensionality (
3.3 Competitive edge in stability and manifold quality against state-of-the-art generative models
We assessed LiVAE against eight state-of-the-art deep generative models on
Table 3. Performance differences between LiVAE and advanced generative and scRNA-seq-specialized methods across
While trajectory-focused models such as scTour achieved UMAP distance correlation (Figure 4A) statistically equivalent to LiVAE’s (
Figure 4. Enhanced global fidelity and stability versus advanced deep generative models. Boxplots compare LiVAE with eight state-of-the-art baselines across
These strengths translated into superior geometric organization. LiVAE delivered UMAP overall quality improvements of
3.4 Implicit geometric regularization is comparable to explicit graph-based architectures
We benchmarked LiVAE against three prominent graph-aware models (CLEAR, scGNN, and scGCC) on
Table 4. Performance differences between LiVAE and graph-based deep learning methods across
Despite the graph-based models’ design for local structure, LiVAE proved highly competitive, significantly outperforming CLEAR
Figure 5. Strong embedding quality and robustness without explicit graph regularization. Boxplots compare LiVAE with three graph-based baselines across
Model stability was exceptional, with noise resilience substantially higher than that of all comparators (
3.5 Versatile and robust performance on chromatin accessibility data
We evaluated LiVAE on scATAC-seq data against three specialized methods (LSI, PeakVI, and PoissonVI) across
Table 5. Performance differences between LiVAE and scATAC-seq-specialized methods across
Against LSI and PeakVI, LiVAE was unequivocally superior across all categories. Overall intrinsic quality (Figure 6D) improved by
Figure 6. Comparison with scATAC-seq-specialized methods. Boxplots compare LiVAE with LSI, PeakVI, and PoissonVI across
The comparison against PoissonVI revealed LiVAE’s complementary strengths. LiVAE achieved higher core intrinsic quality
3.6 Dual loss pathways provide complementary representational benefits
We performed ablation studies on
Table 6. Performance differences for dual-pathway ablations under bottleneck-only (BN) and Lorentz-regularized (Lorentz) configurations across
Removing the bottleneck pathway (“w/o BN”) caused catastrophic collapse in geometric integrity. Overall intrinsic quality increased by
Figure 7. Ablation analysis of dual loss pathways. (A) Bottleneck-only configuration (BN): full model vs. w/o main and w/o BN,
Conversely, removing the main pathway (“w/o main”) produced an inverted deficiency profile. Supervised clustering accuracy collapsed (NMI/ARI
3.7 A deterministic anchor point fortifies geometric regularization
To establish the most effective method for applying Lorentz-distance regularization, we contrasted two strategies: anchoring the calculation to the deterministic information BN versus using two independently sampled posterior views (Views). Evaluation on scRNA-seq
Figure 8. Comparison of Lorentz-regularization strategies. (A) scRNA-seq
Table 7. Performance differences between bottleneck-anchored (BN) and two-view sampled (Views) Lorentz regularization across scRNA-seq
The BN approach’s most profound impact was on model robustness and manifold quality. Noise resilience increased dramatically (
These results indicate that anchoring Lorentz regularization to a fixed, deterministic reference point mitigates training instability inherent in stochastic sampling. The bottleneck provides a consistent geometric scaffold, enabling more coherent latent organization across all analysis modalities.
3.8 Optimizing data fidelity through modality-aware reconstruction
Selecting an appropriate reconstruction loss is crucial for modeling distinct single-cell assay properties. We benchmarked four likelihoods—NB, ZINB, Poisson, and ZI-Poisson—on scRNA-seq
Figure 9. Evaluation of reconstruction likelihood functions. (A) scRNA-seq
Table 8. Performance differences for reconstruction likelihood functions across scRNA-seq
For scRNA-seq, NB provided the most robust performance (Figure 9A). Its primary advantage was enhanced noise resilience (
For scATAC-seq, ZINB was superior (Figure 9B), driven by better noise resilience (
3.9 Robustness and hyperparameter stability
We evaluated LiVAE’s sensitivity to key hyperparameters on
Figure 10. Hyperparameter sensitivity analysis. (A) Lorentz-regularization weight
Table 9. Performance differences for hyperparameter ablations across
Increasing
The bottleneck dimensionality analysis revealed that
Latent dimensionality presented a clear trade-off (Figure 10C). The
These findings support the default settings of
3.10 Emergent batch correction and robust clustering performance
LiVAE’s information bottleneck and geometric regularization promote globally coherent embeddings that can disentangle biological signals from batch effects without explicit batch-correction terms. We benchmarked multi-batch scRNA-seq integration against scVI, scDHMap, scDeepCluster, scGNN, and scGCC on 21 multi-batch datasets.
UMAP visualizations of five representative datasets show well-mixed embeddings preserving biological structure (Figure 11A). Quantitative iLISI evaluation across the full set of 21 datasets with 2,000–8,000 cell subsamplings revealed that LiVAE achieves batch mixing comparable to that with specialized methods (Figure 11B).
Figure 11. Batch integration and supervised clustering across multi-batch scRNA-seq datasets. (A) Representative UMAP embeddings from LiVAE and five comparison methods across five multi-batch datasets, colored by batch. (B) iLISI evaluation across downsampled cell counts (2,000–8,000); LiVAE achieves comparable batch mixing to specialized methods across
Supervised clustering evaluation using four pipelines—combining K-means or Leiden for pre- and post-integration clustering (denoted as K–K, K–L, L–K, and L–L)—showed mixed results (Figure 11C; Table 10). LiVAE substantially outperformed scDeepCluster (ARI:
Table 10. Paired differences in supervised clustering across four clustering strategies evaluated on multi-batch single-cell datasets (
These findings demonstrate that LiVAE provides competitive batch integration and stable clustering without specialized batch parameters. Although dedicated batch-correction methods may be preferred for datasets with extreme confounding, LiVAE offers a versatile, general-purpose solution for integrated single-cell analysis.
3.11 Biological interpretability of latent components on a Dapp1 perturbation dataset
To assess LiVAE’s biological interpretability, we analyzed hematopoietic stem and progenitor cell scRNA-seq with Dapp1 knockout perturbation (GSE277292). UMAP embedding showed consistent cellular structure with minimal batch effects between wild-type and knockout conditions (Figure 12A). We annotated components by identifying genes with the highest per-cell expression–activation correlation.
Figure 12. Interpretability of LiVAE latent components in a Dapp1 perturbation scRNA-seq dataset. (A) UMAP of cells from GSE277292, colored by condition (wt: wild-type; ko: knockout), and schematic of the gene-component association workflow based on the maximum expression correlation. (B) UMAPs of selected component activation scores (top of each pair) alongside expression of the most correlated marker genes (bottom). Functional groupings include cell cycle and protein synthesis (Latent0/Cks2 and Latent2/Rps24); stress response and cellular protection (Latent1/Plac8 and Latent8/Ctla2a); transcriptional regulation (Latent4/Junb); immune identity and differentiation (Latent5/Ighm and Latent9/Irf8); and myeloid lineage commitment (Latent3/Mpo, Latent6/Ms4a3, and Latent7/Cd63). (C) Gene Ontology biological process (GOBP) enrichment for the top correlated genes with Latent1, Latent4, and Latent5. Dot size indicates the gene count; color encodes adjusted
Individual components captured distinct biological programs (Figure 12B). Cell-cycle and protein synthesis were tracked by Latent0 (Cks2) and Latent2 (Rps24). Stress-response programs aligned with Latent1 [Plac8 (Rogulski et al., 2005)] and Latent8 (Ctla2a). Transcriptional regulation was reflected in Latent4 [Junb (Santilli et al., 2021)]. Immune identity was found through Latent5 [Ighm (Dobre et al., 2021)] and Latent9 [Irf8 (Kurotaki and Tamura, 2019)], while myeloid commitment was found through Latent3 [Mpo (Lanza et al., 2001)], Latent6 [Ms4a3 (Donato et al., 2002)], and Latent7 [Cd63 (Pols and Klumperman, 2009)].
Gene Ontology biological process enrichment corroborated these assignments (Figure 12C): Latent4 enriched for “mitotic cell cycle,” Latent1 enriched for “hemopoiesis,” and Latent5 enriched for “myeloid differentiation.” These results demonstrate that LiVAE decomposes transcriptomes into disentangled, biologically meaningful axes, facilitating data-driven hypothesis generation.
4 Discussion
We introduced LiVAE, a geometrically regularized variational autoencoder that address the local–global trade-off in single-cell representation learning. Through systematic benchmarking across 135 datasets against 21 baseline methods spanning classical dimensionality reduction, deep generative models, graph-based architectures, and modality-specific approaches, we demonstrated that LiVAE achieves higher global topology preservation, richer latent manifold geometry, and enhanced robustness while maintaining competitive local structure fidelity. These technical advances translate to improved biological discovery: LiVAE embeddings better preserve developmental hierarchies, enable more accurate cell-type annotation, and provide interpretable latent dimensions aligned with known biological processes.
Unlike prior hyperbolic deep learning approaches that constrain entire latent spaces to hyperbolic manifolds (Park et al., 2021)—requiring manifold-aware operations, hyperbolic priors, and specialized reparameterization that increase computational cost and reduce flexibility (Cho et al., 2023)—LiVAE applies hyperbolic geometry only as regularization over a standard Euclidean latent space
We use the full latent vector
While not explicitly designed for batch correction, LiVAE achieves comparable iLISI scores to scVI across 21 multi-batch datasets through three mechanisms: the information bottleneck attenuates batch-specific artifacts orthogonal to biological signal (Voloshynovskiy et al., 2019), geometric loss enforces global coherence that implicitly aligns cross-batch representations, and shared decoders incentivize batch-invariant features. For datasets with severe batch confounding (e.g., cell types appearing in only one batch), scVI’s explicit batch modeling may be superior, but LiVAE’s simpler architecture—requiring no batch labels and avoiding adversarial training instabilities—offers practical advantages for routine integration.
Based on our systematic benchmarking, we recommend using LiVAE when the dataset structure is unknown and exploratory analysis is needed, global topology preservation is critical (e.g., identifying rare populations and inferring developmental hierarchies), or cross-dataset integration is required without batch labels. Alternative methods should be used when the trajectory structure is well-defined and pseudotime accuracy is paramount (prefer scDHMap and scTour), extreme sparsity (
Several limitations motivate future development. First, while our component-wise interpretability analysis demonstrates that latent dimensions capture biologically meaningful variation, LiVAE does not enforce strict disentanglement—the components may exhibit residual correlations, unlike
Beyond these immediate needs, our results establish geometric regularization—specifically Lorentzian distance constraints across information bottlenecks—as a powerful strategy for learning hierarchical representations that extend beyond transcriptomics to spatial transcriptomics (cells
Data availability statement
The original contributions presented in the study are publicly available. The source code for this research is publicly available on GitHub at https://github.com/PeterPonyu/LiVAE. The single-cell sequencing data for the Dapp1 perturbation experiments is publicly available in the Gene Expression Omnibus (GEO) under accession number GSE277292.
Author contributions
ZF: Writing – review and editing, Visualization, Funding acquisition, Conceptualization, Software, Investigation, Writing – original draft, Resources, Validation, Formal analysis, Project administration, Methodology, Data curation, Supervision. JF: Writing – review and editing, Data curation, Supervision, Formal analysis, Validation, Investigation, Resources, Visualization. CC: Resources, Validation, Formal analysis, Writing – original draft, Data curation, Visualization, Investigation. KZ: Formal analysis, Supervision, Visualization, Writing – original draft, Data curation, Resources, Validation. SW: Writing – review and editing, Data curation, Supervision, Formal analysis, Project administration, Validation, Investigation, Funding acquisition, Resources.
Funding
The authors declare that financial support was received for the research and/or publication of this article. This work was supported by the National Key R&D Program of China (Grant No. 2024YFA1107101), Ministry of Science and Technology of the People’s Republic of China (Grant No. 2024YFA1107101) and the National Key Laboratory of Trauma and Chemical Poisoning (Grant No. 2024K004).
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The authors declare that no Generative AI was used in the creation of this manuscript.
Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary material
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fgene.2025.1713727/full#supplementary-material
References
Abdelaleem, E., Abid, A., and Kording, K. (2025). Deep variational multivariate information bottleneck. J. Mach. Learn Res. 26 (19), 1–39. doi:10.48550/arXiv.2310.03311
Becht, E., McInnes, L., Healy, J., Dutertre, C. A., Kwok, I. W. H., Ng, L. G., et al. (2019). Dimensionality reduction for visualizing single-cell data using UMAP. Nat. Biotechnol. 37 (1), 38–44. doi:10.1038/nbt.4314
Bronstein, M. M., Bruna, J., Cohen, T., and Veličković, P. (2021). Geometric deep learning: grids, groups, graphs, geodesics, and gauges. arXiv Preprint arXiv:2104.13478. doi:10.48550/arXiv.2104.13478
Burgess, C. P., Higgins, I., Pal, A., Matthey, L., Watters, N., Desjardins, G., et al. (2018). Understanding disentangling in β-VAE. arXiv Preprint arXiv:1804.03599. doi:10.48550/arXiv.1804.03599
Cao, J., Spielmann, M., Qiu, X., Huang, X., Ibrahim, D. M., Hill, A. J., et al. (2019). The single-cell transcriptional landscape of Mammalian organogenesis. Nature 566 (7745), 496–502. doi:10.1038/s41586-019-0969-x
Chami, I., Ying, Z., Ré, C., and Leskovec, J. (2019). Hyperbolic graph convolutional neural networks. Adv. Neural Inf. Process Syst. 32, 4868–4879. doi:10.5555/3454287.3454725
Chen, S., Lake, B. B., and Zhang, K. (2019). High-throughput sequencing of the transcriptome and chromatin accessibility in the same cell. Nat. Biotechnol. 37 (12), 1452–1457. doi:10.1038/s41587-019-0290-0
Cho, S., Hong, S., Jeon, S., Lee, Y., Sohn, K., and Shin, J. (2023). Hyperbolic VAE via latent Gaussian distributions. Adv. Neural Inf. Process Syst. 36, 569–588. doi:10.48550/arXiv.2209.15217
Choi, J., Kim, H., Kim, C., and Sohn, I. (2023). VAE-IF: deep feature extraction with averaging for fully unsupervised artifact detection in routinely acquired ICU time-series. IEEE J. Biomed. Health Inf. 27 (6), 2777–2788.
Cui, H., Wang, C., Maan, H., Pang, K., Luo, F., Duan, N., et al. (2024). scGPT: toward building a foundation model for single-cell multi-omics using generative AI. Nat. Methods 21 (8), 1470–1480. doi:10.1038/s41592-024-02201-0
Deng, T., Huang, M., Xu, K., Lu, Y., Xu, Y., Chen, S., et al. (2025). LEGEND: identifying co-expressed genes in multimodal transcriptomic sequencing data. Genomics Proteomics Bioinforma., qzaf056. doi:10.1093/gpbjnl/qzaf056
Ding, J., Condon, A., and Shah, S. P. (2018). Interpretable dimensionality reduction of single cell transcriptome data with deep generative models. Nat. Commun. 9 (1), 2002. doi:10.1038/s41467-018-04368-5
Dobre, M., Uddin, M., and Zouali, M. (2021). Molecular checkpoints of human B-cell selection and development in the bone marrow. Immunology 164 (3), 441–451.
Donato, J. L., Kunz, G., Staf, C., Aeby, P., Kion, T., and Stamenkovic, I. (2002). Htm4, a new member of the CD20/beta-subunit of the B-cell antigen receptor-associated protein family, is a myeloid-specific, G-protein-coupled receptor. J. Immunol. 168 (10), 5037–5047.
Fang, R., Preissl, S., Li, Y., Hou, X., Lucero, J., Wang, X., et al. (2021). Comprehensive analysis of single cell ATAC-seq data with SnapATAC. Nat. Commun. 12 (1), 1337. doi:10.1038/s41467-021-21583-9
Ganea, O., Bécigneul, G., and Hofmann, T. (2018). Hyperbolic neural networks. Adv. Neural Inf. Process Syst. 31, 5345–5355. doi:10.5555/3327345.3327440
Hao, Y., Hao, S., Andersen-Nissen, E., Mauck, W. M., Zheng, S., Butler, A., et al. (2021). Integrated analysis of multimodal single-cell data. Cell 184 (13), 3573–3587.e29. doi:10.1016/j.cell.2021.04.048
Heiser, C. N., and Lau, K. S. (2020). A quantitative framework for evaluating single-cell data structure preservation by dimensionality reduction techniques. Cell Rep. 31 (5), 107576. doi:10.1016/j.celrep.2020.107576
Hetzel, L., Fischer, D. S., Günnemann, S., and Theis, F. J. (2021). Graph representation learning for single-cell biology. Curr. Opin. Syst. Biol. 28, 100347. doi:10.1016/j.coisb.2021.05.008
Hu, J., Li, X., Coleman, K., Schroeder, A., Ma, N., Irwin, D. J., et al. (2021). SpaGCN: integrating gene expression, spatial location and histology to identify spatial domains and spatially variable genes by graph convolutional network. Nat. Methods 18 (11), 1342–1351. doi:10.1038/s41592-021-01255-8
Kim, H., and Mnih, A. (2018). “Disentangling by factorising,” in 35th International Conference on Machine Learning (ICML 2018). Stockholm, Sweden: PMLR, 2649–2658.
Kiselev, V. Y., Andrews, T. S., and Hemberg, M. (2019). Challenges in unsupervised clustering of single-cell RNA-seq data. Nat. Rev. Genet. 20 (5), 273–282. doi:10.1038/s41576-018-0088-9
Klimovskaia, A., Lopez-Paz, D., Bottou, L., and Nickel, M. (2020). Poincaré maps for analyzing complex hierarchies in single-cell data. Nat. Commun. 11 (1), 2966. doi:10.1038/s41467-020-16822-4
Kobak, D., and Berens, P. (2019). The art of using t-SNE for single-cell transcriptomics. Nat. Commun. 10 (1), 5416. doi:10.1038/s41467-019-13056-x
Kurotaki, D., and Tamura, T. (2019). IRF8: the orchestrator of myeloid and lymphoid immune system. Int. Immunol. 31 (9), 569–576.
Lan, W., Yang, S., Liu, J., and Xu, J. (2023). scIAC: clustering scATAC-seq data based on Student’s t-distribution similarity imputation and zero-inflated negative binomial model. IEEE/ACM Trans. Comput. Biol. Bioinform 20 (3), 2152–2163. doi:10.1109/BIBM55620.2022.9995225
Lanza, F., Castoldi, G. L., and Castagnari, B. (2001). The flow cytometric analysis of myeloperoxidase (MPO) in the differential diagnosis of acute leukemia. Leuk. Lymphoma 42 (5), 885–896.
Li, N., Gao, M., Zheng, C., Liu, J., and Liu, Z. (2023). Hyperbolic hierarchical knowledge graph embeddings for biological entities. J. Biomed. Inf. 147, 104520. doi:10.1016/j.jbi.2023.104503
Li, W., Zhu, M., Xu, Y., Huang, M., Wang, Z., Chen, J., et al. (2025). SIGEL: a context-aware genomic representation learning framework for spatial genomics analysis. Genome Biol. 26 (1), 287. doi:10.1186/s13059-025-03748-7
Lopez, R., Regier, J., Cole, M. B., Jordan, M. I., and Yosef, N. (2018). Deep generative modeling for single-cell transcriptomics. Nat. Methods 15 (12), 1053–1058. doi:10.1038/s41592-018-0229-2
Luecken, M. D., Büttner, M., Chaichoompu, K., Danese, A., Interlandi, M., Mueller, M. F., et al. (2022). Benchmarking atlas-level data integration in single-cell genomics. Nat. Methods 19 (1), 41–50. doi:10.1038/s41592-021-01336-8
Lynch, A. W., Theodoris, C. V., Long, A., Brown, M., Liu, X. S., and Meyer, C. A. (2023). Multi-batch single-cell comparative atlas construction by generalized and supervised integration. Nat. Commun. 14 (1), 3854. doi:10.1038/s41467-023-39494-2
Madrigal, A., Brazovskaja, A., Guna, A., Treutlein, B., and Lundberg, E. (2024). A unified model for interpretable latent embedding of multi-sample, multi-condition single-cell data. Nat. Commun. 15 (1), 7279. doi:10.1038/s41467-024-50963-0
Mathieu, E., Le Lan, C., Maddison, C. J., Tomioka, R., and Teh, Y. W. (2019). Continuous hierarchical representations with Poincaré variational auto-encoders. Adv. Neural Inf. Process Syst. 32, 12565–12576. doi:10.48550/arXiv.1901.06033
McInnes, L., Healy, J., and Melville, J. (2018). UMAP: uniform manifold approximation and projection for dimension reduction. arXiv Preprint arXiv:1802.03426. doi:10.48550/arXiv.1802.03426
Moon, K. R., van Dijk, D., Wang, Z., Gigante, S., Burkhardt, D. B., Chen, W. S., et al. (2018). Manifold learning-based methods for analyzing single-cell RNA-sequencing data. Curr. Opin. Syst. Biol. 7, 36–46. doi:10.1016/j.coisb.2017.12.008
Nagano, Y., Yamaguchi, S., Fujita, Y., and Koyama, M. (2019). “A wrapped normal distribution on hyperbolic space for gradient-based learning,” in 36th International Conference on Machine Learning (ICML 2019). Long Beach, CA: PMLR, 4693–4702.
Nguyen, N. D., Huang, J., and Wang, D. (2022). A deep manifold-regularized learning model for improving phenotype prediction from multi-modal data. Nat. Comput. Sci. 2, 38–46. doi:10.1038/s43588-021-00185-x
Nickel, M., and Kiela, D. (2017). Poincaré embeddings for learning hierarchical representations. Adv. Neural Inf. Process Syst. 30, 6338–6347. doi:10.48550/arXiv.1705.08039
Park, J., Cho, J., Chang, H. J., and Choi, J. Y. (2021). “Unsupervised hyperbolic representation learning via message passing auto-encoders,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 5516–5526.
Pogány, D., Rečnik, L. M., and Žitnik, M. (2024). Towards explainable interaction prediction: embedding biological hierarchies into hyperbolic interaction space. PLOS ONE 19 (3), e0300906. doi:10.1371/journal.pone.0300906
Pols, M. S., and Klumperman, J. (2009). Trafficking and function of the tetraspanin CD63. Exp. Cell Res. 315 (9), 1584–1592. doi:10.1016/j.yexcr.2008.09.020
Rachid Zaim, S., Johnson, R., Alcaraz, N. P., Wang, B., Camino, E., Prokop, J. W., et al. (2024). MOCHA’s advanced statistical modeling of scATAC-seq data enables functional genomic inference in large human disease cohorts. Nat. Commun. 15 (1), 6405. doi:10.1038/s41467-024-50612-6
Rogulski, K., Li, N., Guerrero, C., Kim, J., Yi, D., Vence, L., et al. (2005). Onzin, a c-Myc-repressed new gene, promotes survival of myeloid cells. FASEB J. 19 (13), 1867–1869.
Saelens, W., Cannoodt, R., Todorov, H., and Saeys, Y. (2019). A comparison of single-cell trajectory inference methods. Nat. Biotechnol. 37 (5), 547–554. doi:10.1038/s41587-019-0071-9
Santilli, S., Heckl, D., and Riemke, P. (2021). The role of AP-1 transcription factor JunB in hematopoiesis and leukemogenesis. Haematologica 106 (3), 666–677.
Sarkar, R. (2012). “Low distortion Delaunay embedding of trees in hyperbolic plane,” in Graph Drawing: 19th International Symposium, GD 2011, 355–366.
Sikka, H., Zhong, W., Yin, J., and Pehlevan, C. (2019). “A closer look at disentangling in β-VAE,” in 53rd Asilomar conference on signals, systems, and computers (IEEE), 437–441.
Skopek, O., Ganea, O. E., and Bécigneul, G. (2019). Mixed-curvature variational autoencoders. arXiv Preprint arXiv:1911.08411. doi:10.48550/arXiv.1911.08411
Skopek, O., Ganea, O. E., and Bécigneul, G. (2020). “Mixed-curvature variational autoencoders,” in International conference on learning representations.
Song, T., Choi, H., Jang, A., Han, S., Roh, S., Gong, G., et al. (2022a). Detecting spatially co-expressed gene clusters with functional coherence by graph-regularized convolutional neural network. Bioinformatics 38 (5), 1344–1352. doi:10.1093/bioinformatics/btab812
Song, Q., Su, J., Zhang, W., Chen, M., and Su, J. (2022b). SMGR: a joint statistical method for integrative analysis of single-cell multi-omics data. Nar. Genom. Bioinform. 4 (3), lqac056. doi:10.1093/nargab/lqac056
Strouse, D. J., and Schwab, D. J. (2017). The deterministic information bottleneck. Neural Comput. 29 (6), 1611–1630. doi:10.1162/NECO_a_00961
Stuart, T., Butler, A., Hoffman, P., Hafemeister, C., Papalexi, E., Mauck, W. M., et al. (2019). Comprehensive integration of single-cell data. Cell 177 (7), 1888–1902.e21. doi:10.1016/j.cell.2019.05.031
Tian, L., Dong, X., Freytag, S., Le Cao, K. A., Su, S., JalalAbadi, A., et al. (2019). Benchmarking single cell RNA-sequencing analysis pipelines using mixture control experiments. Nat. Methods 16 (6), 479–487. doi:10.1038/s41592-019-0425-8
Tian, T., Zhang, Y., Huang, Z., Wan, J., Liu, F., Chen, D., et al. (2023). Complex hierarchical structures in single-cell genomics data unveiled by deep hyperbolic manifold learning. Brief. Bioinform. 24 (2), bbad086.
Tishby, N., and Zaslavsky, N. (2015). “Deep learning and the information bottleneck principle,” in 2015 IEEE information theory workshop (ITW) (IEEE), 1–5.
Trapnell, C., Cacchiarelli, D., Grimsby, J., Pokharel, S., Li, S., Morse, M., et al. (2014). The dynamics and regulators of cell fate decisions are revealed by pseudotemporal ordering of single cells. Nat. Biotechnol. 32 (4), 381–386. doi:10.1038/nbt.2859
Voloshynovskiy, S., Kondah, M., Rezaeifar, S., Taran, O., Holotyak, T., and Rezende, D. J. (2019). Information bottleneck through variational glasses. arXiv Preprint arXiv:1912.00830. doi:10.48550/arXiv.1912.00830
Yan, F., Powell, D. R., Curtis, D. J., and Wong, N. C. (2020). From reads to insight: a hitchhiker’s guide to ATAC-seq data analysis. Genome Biol. 21 (1), 22. doi:10.1186/s13059-020-1929-3
Yuan, H., and Kelley, D. R. (2022). scBasset: sequence-based modeling of single-cell ATAC-seq using convolutional neural networks. Nat. Methods 19 (9), 1088–1096. doi:10.1038/s41592-022-01562-8
Zhao, F., Kang, Y., and Wakefield, J. (2024). A novel statistical method for differential analysis of single-cell chromatin accessibility sequencing data. PLOS Comput. Biol. 20 (3), e1011854. doi:10.1371/journal.pcbi.1011854
Zuo, C., and Chen, L. (2021). Deep-joint-learning analysis model of single cell transcriptome and open chromatin accessibility data. Brief. Bioinform. 22 (4), bbaa287. doi:10.1093/bib/bbaa287
Glossary
ARI Adjusted Rand index
BN Bottleneck
CLEAR Contrastive learning for scRNA-seq
DAV Davies–Bouldin index
DIPVAE Disentangled inferred prior VAE
FactorVAE Factor variational autoencoder
HVGs Highly variable genes
ICA Independent component analysis
InfoVAE Information maximizing VAE
KPCA Kernel principal component analysis
LSI Latent semantic indexing
NB Negative binomial
NMI Normalized mutual information
PCA Principal component analysis
PoissonVI Poisson variational inference
scATAC-seq Single-cell ATAC sequencing
scDHMap Single-cell deep hyperbolic manifold learning
scGNN Single-cell graph neural network
scTour Single-cell trajectory optimization by unsupervised representation
SDR Spectral decay rate
TF-IDF Term frequency-inverse document frequency
TSVD Truncated singular value decomposition
VAE Variational autoencoder
ZIP Zero-inflated Poisson
ASW Average silhouette width
CAL Calinski–Harabasz index
COR Coupling degree
DICL Dictionary learning
FA Factor analysis
HSD Honest significant difference
HVPs Highly variable peaks
iLISI Integration local inverse Simpson’s index
KL Kullback–Leibler
LiVAE Lorentzian variational autoencoder
MD Manifold dimensionality
NMF Non-negative matrix factorization
NR Noise resilience
PeakVI Peak variational inference
PR Participation ratio
scDeepCluster Single-cell deep clustering
scGCC Single-cell graph contrastive clustering
scRNA-seq Single-cell RNA sequencing
scVI Single-cell variational inference
TD Trajectory directionality
t-SNE t-distributed stochastic neighbor embedding
UMAP Uniform manifold approximation and projection
ZINB Zero-inflated negative binomial
Keywords: single-cell multi-omics, dual-pathway c, hyperbolic geometry, information bottleneck, manifold learning, interpretable representation
Citation: Fu Z, Fu J, Chen C, Zhang K and Wang S (2026) Lorentz-regularized interpretable VAE for multi-scale single-cell transcriptomic and epigenomic embeddings. Front. Genet. 16:1713727. doi: 10.3389/fgene.2025.1713727
Received: 06 October 2025; Accepted: 20 November 2025;
Published: 05 January 2026.
Edited by:
Kenta Nakai, The University of Tokyo, JapanReviewed by:
Xiaobo Sun, Zhongnan University of Economics and Law, ChinaWeihang Zhang, Duke University, United States
Copyright © 2026 Fu, Fu, Chen, Zhang and Wang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Zeyu Fu, ZnV6ZXl1OTlAMTI2LmNvbQ==; Song Wang, c3dhbmcxOTgxQHRtbXUuZWR1LmNu
†These authors have contributed equally to this work
Song Wang1*