ORIGINAL RESEARCH article
Front. Oncol.
Sec. Cancer Imaging and Image-directed Interventions
Volume 15 - 2025 | doi: 10.3389/fonc.2025.1480384
This article is part of the Research TopicQuantitative Imaging: Revolutionizing Cancer Management with biological sensitivity, specificity, and AI integrationView all 25 articles
Leveraging a Foundation Model Zoo for Cell Similarity Search in Oncological Microscopy across Devices
Provisionally accepted- 1Collaborative Research Institute Intelligent Oncology, Freiburg, Germany
- 2Neurorobotics Lab, Department of Computer Science, University of Freiburg, Freiburg, Germany
- 3Laboratory of Applied Cellular and Molecular Biology,Institute of Veterinary Sciences of the Litoral,Universidad Nacional del Litoral (UNL)-National Scientific and Technical Research Council (CONICET), Esperanza, Argentina
- 4Department of Hematology, Oncology and Stem Cell Transplantation, Medical Center University of Freiburg, Faculty of Medicine, University of Freiburg, Freiburg, Germany
- 5Lighthouse Core Facility, Medical Center – University of Freiburg, Faculty of Medicine, University of Freiburg, Freiburg, Germany
- 6German Cancer Consortium (DKTK) and German Cancer Research Center (DKFZ), Partner Site Freiburg, Freiburg, Germany
- 7Mertelsmann Foundation, Freiburg, Germany
- 8IMBIT//BrainLinks-BrainTools, University of Freiburg, Freiburg, Germany
- 9Department of Pediatrics, Experimental Immnology and Cell Therapy, Goethe University Frankfurt, Frankfurt am Main, Germany
- 10University Cancer Center Frankfurt (UCT), Frankfurt am Main, Germany
- 11Mildred Scheel Career Center (MSNZ), Hospital of the Goethe University, Frankfurt, Hesse, Germany
- 12German Cancer Consortium (DKTK) and German Cancer Research Center (DKFZ), Partner Site Frankfurt, Frankfurt am Main, Germany
- 13Frankfurt Cancer Institute (FCI), Frankfurt, Hesse, Germany
Select one of your emails
You have multiple emails registered with Frontiers:
Notify me on publication
Please enter your email address:
If you already have an account, please login
You don't have a Frontiers account ? You can register here
Background: Cellular imaging analysis using the traditional retrospective approach is extremely time-consuming and labor-intensive. Although AI-based solutions are available, these approaches rely heavily on supervised learning techniques that require high quality, large labeled datasets from the same microscope to be reliable. In addition, primary patient samples are often heterogeneous cell populations and need to be stained to distinguish the cellular subsets. The resulting imaging data is analyzed and labeled manually by experts. Therefore, a method to distinguish cell populations across imaging devices without the need for staining and extensive manual labeling would help immensely to gain real-time insights into cell population dynamics. This especially holds true for recognizing specific cell types and states in response to treatments. We aim to develop an unsupervised approach using general vision foundation models trained on diverse and extensive imaging datasets to extract rich visual features for cell-analysis across devices, including both stained and unstained live cells. Our method, Entropy-guided Weighted Combinational FAISS (EWC-FAISS), uses these models purely in an inference-only mode without task-specific retraining on the cellular data. Combining the generated embeddings in an efficient and adaptive k-nearest neighbor search allows for automated, crossdevice identification of cell types and states, providing a strong basis for AI-assisted cancer therapy. We utilized two publicly available datasets. The WBC dataset includes 14,424 images of stained white blood cell samples from patients with acute myeloid and lymphoid leukemia, as well as those without leukemic pathology. The LISC dataset comprises 257 images of white blood cell samples from healthy individuals. We generated four in-house datasets utilizing the JIMT-1 breast cancer cell line, as well as Jurkat and K562 (leukemic cell lines). These datasets were acquired using the Nanolive 3D Cell Explorer-fluo (CX-A) holotomographic microscope and the BioTek Lionheart FX automated brightfield microscope. To generate the embeddings, we used and optimized a concatenated combination of SAM, DINO, ConvNeXT, SWIN, CLIP and ViTMAE. The combined embeddings were used as input for the adaptive k-nearest neighbor search, building an approximate Hierarchical Navigable Small World FAISS index. We compared EWC-FAISS to fully trained Classifiers with DINO, SWIN and ConvNeXT backbones, and to NMTune.
Keywords: artificial intelligence, deep learning, Foundation models, Nearest neighbor search, Cell imaging
Received: 13 Aug 2024; Accepted: 21 May 2025.
Copyright: © 2025 Kalweit, Klett, Silvestrini, Rahnfeld, Naouar, Vogt, Infante, Berger, Duque-Afonso, Hartmann, Follo, Bodurova-Spassova, Lübbert, Mertelsmann, Boedecker, Ullrich and Kalweit. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
* Correspondence: Gabriel Kalweit, Collaborative Research Institute Intelligent Oncology, Freiburg, Germany
Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.