Skip to main content

TECHNOLOGY AND CODE article

Front. Physiol., 26 May 2023
Sec. Red Blood Cell Physiology
This article is part of the Research Topic Images from Red Cells, Volume II View all 6 articles

RedTell: an AI tool for interpretable analysis of red blood cell morphology

  • 1Institute of AI for Health, Helmholtz Zentrum München—German Research Center for Environmental Health, Neuherberg, Germany
  • 2Chair for Computer Aided Medical Procedures & Augmented Reality, Technical University of Munich, Garching, Germany
  • 3Red Blood Cell Research Group, Institute of Veterinary Physiology, Vetsuisse Faculty and the Zurich Center for Integrative Human Physiology, University of Zurich, Zurich, Switzerland
  • 4Computer Aided Medical Procedures, Johns Hopkins University, Baltimore, MD, United States

Introduction: Hematologists analyze microscopic images of red blood cells to study their morphology and functionality, detect disorders and search for drugs. However, accurate analysis of a large number of red blood cells needs automated computational approaches that rely on annotated datasets, expensive computational resources, and computer science expertise. We introduce RedTell, an AI tool for the interpretable analysis of red blood cell morphology comprising four single-cell modules: segmentation, feature extraction, assistance in data annotation, and classification.

Methods: Cell segmentation is performed by a trained Mask R-CNN working robustly on a wide range of datasets requiring no or minimum fine-tuning. Over 130 features that are regularly used in research are extracted for every detected red blood cell. If required, users can train task-specific, highly accurate decision tree-based classifiers to categorize cells, requiring a minimal number of annotations and providing interpretable feature importance.

Results: We demonstrate RedTell’s applicability and power in three case studies. In the first case study we analyze the difference of the extracted features between the cells coming from patients suffering from different diseases, in the second study we use RedTell to analyze the control samples and use the extracted features to classify cells into echinocytes, discocytes and stomatocytes and finally in the last use case we distinguish sickle cells in sickle cell disease patients.

Discussion: We believe that RedTell can accelerate and standardize red blood cell research and help gain new insights into mechanisms, diagnosis, and treatment of red blood cell associated disorders.

Introduction

Hematologists analyze human red blood cells (RBCs) under the microscope to detect abnormalities in RBC morphology, and to study and diagnose disorders (Bouyer et al., 2011; Makhro et al., 2016; Fermo et al., 2017). This manual process is time-consuming, subjective and prone to variability and errors. Moreover, whenever a vast number of single cells needs to be analyzed, manual analysis is impossible, making automatic, high throughput image analysis an essential element in microscopy based RBC research.

Artificial Intelligence (AI) is emerging as one of the main technologies enabling automated processing of blood microscopy images (Deshpande et al., 2021). A typical workflow includes segmentation, feature extraction and classification based on extracted RBC features (Devi et al., 2018; Petrović et al., 2020; Wong et al., 2021). Segmentation is commonly performed through classical computer vision algorithms including operations such as thresholding, image filtering, morphological operators, etc. followed by watershed transforms (Sharif et al., 2012) and ellipse fitting (Naruenatthanaset et al., 2020) to separate single cells. Multi-purpose computational tools provide segmentation algorithms based on the mentioned methods and enable feature extraction of single cell morphology, with CellProfiler (Stirling et al., 2021) and ImageJ (Schneider et al., 2012) being the two most commonly used open-source softwares for microscopy image analysis. However, pipelines and macros require the optimization of many hyperparameters, which are specific for images and experimental setups, and thus do not generalize well to other datasets (Sinha et al., 2015). Alternatively, machine learning methods are employed. For instance, Savkare and Narote (2015) suggest segmenting cells with a k-mean clustering followed by cell separation using a watershed algorithm. In a more modern approach, Wong et al. (2021) use a U-Net (Ronneberger et al., 2015) model to segment RBCs followed by a support vector machine classifier (Suykens and Vandewalle, 1999) to separate overlapping cells, cell morphological features are extracted and used for a cell type classification task with a TabNet (Arik and Pfister, 2019) model. In one of our previous works, we suggest a fully convolutional Alexnet architecture to segment single RBCs using a sliding window approach to prevent overlapping segmentation (Sadafi et al., 2018). A more robust strategy (Dhieb et al., 2019) suggests a cell counting method based on a Mask R-CNN (He et al., 2017), an architecture able to distinguish between single cells, and thus requiring no further post-processing to separate overlapping detections.

As for the classification of the single cells, rule-based approaches, k-nearest neighbors, support vector machines or other classical machine learning methods have been employed (Akrimi et al., 2014; Piety et al., 2015). Specifically, tree-based classifiers such as random forest or gradient boosting were shown to be superior when distinguishing between normal, sickle and other abnormally shaped RBCs based on the previously extracted features (Petrović et al., 2020). Also deep learning based methods were considered for RBC classification. For instance (Alzubaidi et al., 2020; Parab and Mehendale, 2021), and (Savkare and Narote, 2015; Naruenatthanaset et al., 2020) use deep convolutional neural networks for this purpose (Sadafi et al., 2019; Song et al., 2021). apply Fast R-CNN and Faster R-CNN models to differentiate between 14 and 7 RBC subtypes, respectively. The most recent work suggests a pipeline based on deep ensemble learning for RBC morphology classification (Routt et al., 2023). Such methods result in a higher classification accuracy in multi-class settings. However, they take images as input and process them to obtain a useful representation of features, which are completely model generated and are not interpretable as opposed to extraction of hand-crafted features. Moreover, they require a large amount of annotated data to achieve a proper generalization power. As a rule, most work on RBC classification mainly focuses on fitting a classifier for a specific task and specific dataset. Extension of deep classification models to a new task or dataset requires programming skills and computer science expertise to gain a deep understanding of algorithms and adjust them properly. Additionally, their application on a new task requires understanding and running source code, which is not always given (Wagner et al., 2022).

This paper introduces RedTell, an AI tool for interpretable analysis of RBC morphology. RedTell enables single-cell segmentation, extraction of cell morphology features, and RBC classification without any prior knowledge and extensive user interaction. RedTell is a fully automated tool expecting expert input only for cell annotations. It can assist a wide range of research questions by 1) providing a robust RBC segmentation model with the possibility of adaptation to new datasets with or without fine-tuning, 2) extracting a wide range of RBC morphological features as described in the literature (Veluchamy, 2012; Devi et al., 2018) and 3) enabling task- and dataset-specific explainable classification algorithms with a low amount of annotated input images. The results of every step of the RedTell pipeline can be directly used for the downstream analysis. RBC counting, comparison of cell type and feature distribution in different experiments or sample groups, or diagnosis of diseases with irregular cell morphology (such as sickle cell disease) are just some of the possible scenarios where RedTell can be applied. We showcase the abilities of RedTell in three different case studies. First, we show that extracted features have different distributions for healthy and anemia patients. Next, we use RedTell to distinguish cells in the stomatocyte-discotye-echinocyte (SDE) sequence in control samples. Finally, we use RedTell to classify sickle cells in samples of sickle cell disease patients.

The novelty of RedTell includes feature extraction from images in fluorescent (in our case Fluo-4 was used as a fluorescent dye to detect Ca2+) channel, vesicle detection in the Fluo-4 channel, and the increased interpretability: we explicitly provide segmentation and classification results overlaid on the original images, meaningful key characteristics of RBCs widely used by hematologists and feature importance to explain predictions of classification algorithms.

Materials and methods

RedTell is an AI tool developed to facilitate the analysis of microscopic images of human RBCs. It consists of four steps:

1. Accurate cell segmentation

2. Extraction of hand-crafted morphological features from brightfield and fluorescence (Ca2+-dependent Fluo-4 signal as an example) channels including vesicle detection and counting

3. Assistance in data annotation

4. Single RBC classification

Every step can be executed by a simple command in the terminal. RedTell is a software package implemented in Python and is easily accessible through the command line under any operating system. Code and extensive documentation as well as CoMMiTMenT and MemSID datasets used in the manuscript are provided under https://github.com/marrlab/redtell. The software is distributed under MIT license without any restrictions for academic and non-academic use. An overview of RedTell’s functional pipeline is given in Figure 1A. In the following we describe each step in detail.

FIGURE 1
www.frontiersin.org

FIGURE 1. RedTell facilitates single red blood cell profiling by extracting interpretable features and enabling accurate cell classification. (A) Overview of the RedTell functional pipeline: Single red blood cells in microscopy images are segmented, 135 morphological features are extracted, and cells are classified into different cell types. (B) RedTell provides a table with extracted features, brightfield images overlayed with segmentation results, and fluorescent images with detected vesicles highlighted white boxes. (C) RedTell supports annotation of single-cells for automated supervised machine learning based on interpretable features.

Datasets

CoMMiTMenT and MemSID datasets

We use CoMMiTMenT dataset to develop the segmentation solution in RedTell as well as consider it to evaluate feature extraction and classification functionality of the tool. The dataset consists of 3 patients with Thalassemia, 9 patients with sickle cell disease, 8 patients with hereditary xerocytosis caused by mutation in PIEZO1 channel, and 13 patients with hereditary spherocytosis. 26 individuals are in a healthy control group.

An Axiocam mounted on a Axiovert 200 m Zeiss microscope with a ×100 objective was used to obtain the images. No staining or preprocessing is performed.

The CoMMiTMenT study was funded by the European Seventh Framework Program under grant agreement number 602121 (CoMMiTMenT) and from the European Union’s FP7 Programme. The study protocols were approved by the Medical Ethical Research Board of the University Medical Center Utrecht, the Netherlands, under reference code 15/426 M and by the Ethical Committee of Clinical Investigations of Hospital Clinic, Spain (IDIBAPS) under reference code 2013/8436. Additional blood samples of patients with sickle cell disease were obtained from the participants of the MemSID clinical trial performed at the University hospital Zurich (#NCT02615847 at https://clinicaltrials.gov/). The trial protocol was approved by the Ethics committee of Canton Zurich (KEK-ZH 2015-0297).

External datasets

We also consider three publically available datasets of blood smear microscopic images: ThalassemiaPBS (Tyas et al., 2022) and MP-IDB (Loddo et al., 2019). Chula-RBC-12-Dataset (Naruenatthanaset et al., 2020) on ThalassemiaPBS consists of 80 images obtained from four thalassemia patients (20 images for each patient), and covers RBCs of various morphology. The raw images have no annotations. Cropped single-cell images are assigned in a separate dataset to one of 9 morphological subtypes. MP-IDB dataset consists of 229 images of patients affected by four different kinds of malaria parasite. Each image contains RBCs with at least 1 cell hosting a parasite. Only such cells are annotated with a segmentation mask and parasite’s kind and life-cycle stage. Chula-RBC-12-Dataset covers 706 images of RBC (no further details about patient pool are given). It includes labels for every RBC in 12 different morphological subtypes in form of point annotations, no segmentation masks are provided. Due to the limited availability of relevant annotations, we use the described datasets to qualitatively assess the generalization ability of our segmentation model.

Cell segmentation

Accurate cell segmentation is the first essential step for studying morphological features of red blood cells. Single-cell segmentation masks provide exact cell localizations in the image and allow computation of precise cell features. We approach the segmentation task by using a Mask R-CNN model (He et al., 2017), an artificial neural network that extends R-CNN based models (Girshick et al., 2014; Girshick, 2015; Ren et al., 2015) consisting of two stages: 1) a region proposal network to propose and evaluate various locations of objects in the image, and 2) a three head network analyzing objects one by one for object classification and bounding box location refinement. Such architectures have been effective in various image segmentation scenarios (Chiao et al., 2019; Huang et al., 2019) and applied on red blood cell microscopy images with a large diversity of cellular morphologies (Dhieb et al., 2019; Sadafi et al., 2019; Sadafi et al., 2020; Loh et al., 2021). Mask R-CNN advances its forerunners by preserving spatial information of regions corresponding to the objects in the images and by having an additional, fully convolutional head for object segmentation.

RedTell provides a ready-to-use Mask R-CNN model trained on the brightfield images of the control samples from the CoMMiTMenT dataset annotated on the single-cell level as introduced by (Sadafi et al., 2020) as an accurate and robust segmentation solution. In the dataset, every image contains on average only 44.2 cells. Furthermore, RedTell enables automated training of new Mask R-CNN segmentation models on any custom dataset with annotated images provided.

Feature extraction

RedTell supports capturing of valuable biological insights in the analysis of RBCs by extracting interpretable hand-crafted features for every segmented cell. Every feature represents a meaningful characteristic of a cell describing its morphology. The features are extracted from two modalities: 1) brightfield microscopy and 2) fluorescence microscopy. Both image types are converted to grayscale for feature acquisition.

Feature extraction from brightfield channel

The feature set provided by the RedTell can be categorized into three different groups:

- Shape features, which describe cell morphology and include characteristics such as area, length of minor and major axes, aspect ratio and eccentricity. The shape features are widely used to distinguish between normal and irregular RBC for automated diagnosis of RBC disorders such as sickle cell disease or thalassemia (Chy and Anisur Rahaman, 2019; Das et al., 2020; Tyas et al., 2020).

- Intensity-based statistical features, which describe cell properties derived from intensity distribution of pixels corresponding to a cell such as mean intensity, standard deviation, skewness and kurtosis. Such features were shown to be useful to differentiate between healthy and sickle RBCs based on the intensities of color and grayscale images. (Akrimi et al., 2014; Tyas et al., 2020). Intensity features are particularly meaningful when extracted from the images of the fluorescence microscopy as they provide insights about cell behavior measured by the activation of the fluorescence as reviewed in (Coffman and Wu, 2012).

- Texture features, which reflect local distribution of pixel intensities and characterize homogeneity and texture of a cell. For every cell we extract the histogram, Gray Level Co-occurrence Matrix (GLCM) and Gray Level Dependence Matrix (GLDM), which represent a relationship between adjacent pixels and include features such as contrast and dissimilarity. We further extract the Gray Level Size Zone Matrix (GLSZM) and the Gray Level Run Length Matrix (GLRLM) to indicate basic structures within a cell such as the size of the largest region of the same pixel intensity. Texture features have been successfully exploited for the RBC classification in blood smear images (Veluchamy, 2012; Akrimi et al., 2014; Chy and Anisur Rahaman, 2019).

Shape features are directly obtained from segmentation masks of the cells, while intensity and texture features are calculated after co-localizing the single-cell masks in the normalized images of brightfield and fluorescent channels. One of the biggest advantages of calculating features based on single-cell masks is focusing only on pixel values that are corresponding to a cell rather than its surrounding artifacts and noise.

Feature extraction from fluorescence channel

Segmentation masks obtained from the brightfield channel provide the exact cell positions. However, the brightfield and fluorescent channel images are produced sequentially. As the microscope stage and the imaging chamber containing RBCs may move during the switch from one channel to the other or due to the tumbling of RBC membrane, slight displacement in the position of a given cell may be introduced which becomes visible why sets of brightfield and fluorescent images are overlaid. An approach to train a separate Mask R-CNN model for cell detection and segmentation in a fluorescent channel requires additional annotation of the cells in fluorescent images and introduces a non-trivial task of the assignment of detections between brightfield and fluorescent channels. Instead, we propose applying segmentation masks obtained from the brightfield channel directly to the fluorescent channel images. We approach a possible slight shift in cell position by excluding image background defined as pixels with zero intensity from the feature extraction. This approach can be justified by the fact that intensity value distribution is homogenous or symmetric with respect to the cell center, and covering only a part of the cell provides a reasonable estimate for the entire object. We further eliminate for feature extraction pixels corresponding to the intracellular artifacts such as vesicles, which we detect with an algorithm as described next.

Vesicle detection

We further detect and count vesicles, intracellular compartments of RBCs as first described by (Lew et al., 1985) in sickle cell patients. The vesicles are inside-out facing parts of the plasma membrane containing the plasma membrane Ca2+ pump (PMCA). The latter effectively sequesters Ca2+ into the vesicles preventing activation of Ca2+-dependent K+ (Gardos) channels. The measurements of the abundance, size and Ca2+ storage capacity of the vesicles in RBCs provide information on the Ca2+ leak through the membrane and the ability of the cells to avoid Ca2+ overload. These tasks are best achieved by imaging of the average levels of free Ca2+ ions in living RBCs as well as the distribution of Ca2+ ions within the cells (between the cytosol and the intracellular vesicles) performed using Fluo-4 AM fluorescent dye. This dye was proven to be the best for use in RBCs due to its high fluorescence intensity with a high degree of quenching of the fluorescent signal by hemoglobin being produced (Kaestner et al., 2006; Wang et al., 2022).

Single snapshots provide information on the current levels of Ca2+ in RBCs and their compartments. Functional tests, in which Ca2+ uptake via the Ca2+-permeable channels is stimulated mechanically or chemically, allow to monitor dynamics of Ca2+ movement across the membrane as time lapse image sequences are obtained for the same set of cells (Hänggi et al., 2014; Fermo et al., 2017; Makhro et al., 2017). These tests give indications on the abundance and/or activity of the Ca2+-permeable channels and PMCAs, that maintain the intracellular Ca2+ levels, that, in turn, are involved in regulation of RBC dehydration and deformability (Muravyov and Tikhomirova, 2012; Kaestner et al., 2020).

We can detect inside-out vesicles in the images of Fluo-4 channel as small regions with high intensity values inside the RBCs [e.g., (Hänggi et al., 2014)] as vesicles are filled with Ca2+ as it gets actively pumped into them by the plasma membrane Ca2+ ATPases against the gradient, and function as intracellular Ca2+ stores. Such intraerythrocytic compartmentalisation of Ca2+ is particularly prominent in RBCs of patients with sickle cell disease, where excessive membrane Ca2+ leak is compensated by Ca2+ sequestration into the internal stores preventing immediate terminal dehydration of the cells due to the opening of Ca2+-dependent K+ (Gardos) channels (Lew et al., 1985). Ca2+ overload in RBCs was shown to be a non-specific indicator of multiple hereditary hemolytic anemias and it is more pronounced in patients with severe disease manifestation (Hänggi et al., 2014; Hertz et al., 2017).

Therefore the number of vesicles found in every cell can be considered as a meaningful, clinically relevant feature. Moreover, vesicles having higher Ca2+ concentration can affect the value of intensity and texture features and should be ignored while calculating the features. We introduce an algorithm for vesicle detection and elimination by determining the local maxima of intensity values.

Let IRX×Y be an image represented by pixel intensities. Then

Udx,y:=x^,y^Ix^,y^x,y|1<d

defines a set of pixel intensities in a ball with radius d>0 around the pixel x,yI with respect to the L1 distance and

U¯dx,y:=1Udx,yx^,y^Udx,yIx^,y^

is the average intensity value in Udx,y. The vesicle is then defined as a region Uvx,y around the local maximum x,y)I, which satisfies

U¯vx,yt>1Ucx,y\Uvx,yx,yUcx,y\Uvx,yIx,y

where v, c and t are predefined constants. v is an estimate of vesicle radius and the threshold t corresponds to the least assumed difference in mean intensity values inside a vesicle and a region around it. To determine the local maxima of the image we apply a median filter of size m×m and a maximum filter of size v. This ensures detection of at most one maximum within a single vesicle. We calculate local maxima on the whole original image of the fluorescent channel and map a detected vesicle to a specific cell if it is inside its segmentation mask. Thus, we limit the search for local maxima only on the image section belonging to the cell. The parameters m, d, t, v and c are dependent on the dataset and should be adjusted accordingly. For our dataset, we experimentally determined parameters which provide good vesicle detection results to be =3 , d=10 for local maximum determination, v=4 and c=8 for vesicle and neighborhood definition and 1 for intensity difference threshold between vesicle and neighborhood region.

We remove a vesicle by setting Ix,y=0 for x,yUvx,y. Pixels corresponding to the vesicles are artifacts within the cell image and treated similarly to the background pixels, they should not contribute to calculation of the intensity and texture features.

In total, we extract 135 features: 14 shape features from segmentation masks for the brightfield channel, 18 intensity and 42 texture features from each brightfield and fluorescent channels, and we further obtain the number of vesicles if the fluorescent channel is Ca2+ channel. We use python libraries skimage (van der Walt et al., 2014) and pyradiomics (van Griethuysen et al., 2017) for the extraction of shape features and intensity and texture features, respectively. Shape features are biologically most relevant, while intensity and texture features are primarily useful for classification. We list all extracted features in Table 1.

TABLE 1
www.frontiersin.org

TABLE 1. RedTell extracts 135 interpretable morphological features of RBCs. The table provides a list of features for four feature groups. For every group we provide the source implementation of feature extraction in the Extracted with column. In particular, the detailed description of every feature and their computation formulas are provided in the documentation of source libraries. Column Feature prefix gives a prefix used for each feature group in the tables produced by RedTell after feature extraction.

Cell classification

For many applications such as image-based disease diagnosis there is a need for classification of cells found in the images into specific cell types. Features extracted by RedTell are a valuable resource facilitating this challenging task. These features are successfully used in some relevant works to obtain methods to distinguish different cell types of interest (Tomari et al., 2014; Chy and Anisur Rahaman, 2019). In addition to feature extraction, RedTell also provides an automated training procedure for binary or multi-class classification algorithms without requiring extensive user interaction. Experts can train their task specific and dataset specific classifier by only annotating a small subset of the cells without the need for accessing sophisticated high performance computing resources. RedTell is designed to reduce the total number of annotations necessary for training of a classifier while achieving high classification accuracy (see Figures 7, 8).

Annotation

RedTell supports annotation procedure for generating ground truth necessary for training of classification algorithms. For this, it randomly selects N segmented cells with extracted features, saves single-cell images with unique identifiers and provides a label table containing a column of unique cell identifiers and a label column. (see Figure 1C). Looking at each single-cell image an expert determines the cell label and enters it into the table in the corresponding row. Random sampling of cells for annotation should provide various cells across different images of the dataset and groups, e.g., patient, diseases or experimental settings, to cover a wider range of cells of different morphology. It also ensures that class distribution of the annotated part of the dataset properly reflects class distribution of the entire dataset. Annotated cells with non-empty labels compose a training set for supervised classification The goal is to predict a cell type for non-annotated cells and include the results in the feature output table.

Classification algorithms

For classification, we suggest using tree-based algorithms decision tree, random forest and LightGBM as 1) they do not require a large amount of data to achieve a high classification accuracy and 2) are interpretable by providing feature importance when discriminating between cell classes in every decision node. Furthermore, random forest and LightGBM are ensemble methods aggregating predictions from multiple decision trees to increase accuracy, robustness, and prevent overfitting. All three algorithms are interpretable by design and allow measurement of feature importance in classification. Decision tree is built on optimizing an information criterion (entropy or GINI). The criterion determines which features to use for a node split when building a decision tree. Features with the highest information gain or decrease in impurity in case of entropy and GINI, respectively, are considered to have the largest impact on classification results. For the random forest classifier, for each feature its importance is calculated as average information gain or impurity decrease over all decision trees. In contrast, the LightGBM model does not use any information criterion to build the decision trees and is optimized with gradient boosting. Importance for each feature is given by the number of node splits utilizing the features calculated over all decision trees.

Automated machine learning

Development of a classifier is performed by means of Automated Machine Learning (AutoML) (He et al., 2019), which automatically determines the optimal hyperparameters for every classification algorithm included in the tool and provides the best model for a given dataset and classification problem. AutoML does not require from users any technical expertise, programming skills or deep understanding of algorithms. The RedTell AutoML module includes no.

- loading the dataset from the feature and label tables,

- generating stratified, image-based or random partitioning of the dataset for 5-fold cross validation,

- training of 75 classifiers (25 steps of Bayesian hyperparameter optimization for 3 classification models),

- selecting and saving the best performing model,

- generating model evaluation and model explanation artifacts, and

- applying the model to the unlabeled observations.

For hyperparameter tuning, we use a 5-fold cross validation with average balanced accuracy over 5 folds as an optimization objective. Optimization is performed via Bayesian optimization approach (Snoek et al., 2012). A partition of the annotated dataset into the 5 folds can be chosen between being completely random or in a stratified fashion taking the images into account such that observations coming from a single image are never included in both training and validation sets. Random partition is appropriate for smaller datasets where every image contains annotated cells. In contrast, image-based partition can assess generalization ability more realistic for a larger dataset where classification of cells on previously unseen images is expected. To approach a possible class imbalance in the dataset we use the inverse of class frequency as weights while optimizing the training objective. RedTell trains a decision tree classifier, a random forest, and LightBGM models with various sets of hyperparameters. After finding the optimal model with optimal hyperparameters, the model is fitted on the whole annotated set obtaining the final classification model. This classification model is then applied to the unlabeled observations for class predictions.

Results

In this section, we first discuss and evaluate the proposed cell segmentation based on a Mask R-CNN and the vesicle detection, and then showcase the abilities of RedTell in different case studies. First, we show that extracted features are different for different RBC disorders. Next, we use RedTell to distinguish cells in the stomatocyte-discotye-echinocyte (SDE) sequence in control samples. Finally, we use RedTell to classify sickle cells in patient samples from the MemSID clinical trial.

Red blood cell segmentation

We assess the generalizability of the RBC Mask R-CNN segmentation model on a subset of control samples from the CoMMiTMenT dataset (see Materials and Methods) and investigate how the size of the training set affects model performance. Although RedTell provides a ready-to-use segmentation model, it also includes the option to train a new model for a custom dataset. Here, we are interested in estimating the number of annotated images necessary to obtain a good segmentation result. We thus consider 187 images from control samples containing 8222 annotated RBCs (73% discocytes, 21% echinocytes and 6% stomatocytes, see Figure 7A). We randomly split the data into 150 (80%) images for training and 37 (20%) for testing. From training, we further randomly sample 10, 25, 50, and 100 images to fit a Mask R-CNN model without adjusting model training hyperparameters for different sizes of the training set. We perform experiments on three different training and test splits to obtain confidence intervals for average precision (AP). As shown in Figure 2A, 50 annotated images suffice to achieve an average AP of 0.98 and 0.95 for IoU thresholds of 0.5 and 0.75, respectively. Increasing the number of training images improves AP only decently and mainly narrows the confidence intervals. Moreover, there is only a negligible visual difference between segmentation results of the models trained on 25 and 50 images (Figures 2B–D). Training on only 10 images provides relatively good quantitative results (AP = 0.963 ± 0.021 at IoU of 0.5) but visual inspection via RedTell shows that cells are either partially segmented or completely missed (Figure 2D). Apparently, 10 images are not enough to cover the full range of cell types present in our dataset. Our experiments show that at least 25-50 images are needed to observe cells of every type during training. For datasets of larger blood smear microscopy images containing hundreds of RBCs, a training set of 25 images can be considered as a good starting point.

FIGURE 2
www.frontiersin.org

FIGURE 2. 50 annotated images suffice to train a RBC Mask R-CNN segmentation model with a good performance. (A) Average precision (AP) curves for IoU of 0.5 and 0.75 calculated on three different train and test splits with a training set consistent of 10, 25, 50, 100, and 150 images, where each image contains about 44 single RBCs. (B) Sample image from the test set in one of the splits, (C) corresponding ground truth segmentation masks overlaid on raw image and (D) segmentation results achieved with models trained with datasets of 10, 25, and 50 images overlaid on raw image.

The segmentation model provided in RedTell was trained on all 187 annotated images of the control group. We assess its generalizability in two settings: 1) internally on images of sickle cell disease (SCD) from the MemSID trial and thalassemia patients from the CoMMiTMenT dataset acquired with the same microscopy and sample preparation setup as images used for training, and 2) externally on images of three different publicly available blood smear datasets with different sample preparation and acquisition techniques as introduced in Materials and Methods. Since no ground truth annotations are available, we can only investigate the model generalizability qualitatively. Figure 3 shows the results for internal assessment, where our Mask R-CNN model properly segments irregular cells, sickle cells in case of SCD, and destroyed cells in case of thalassemia without seeing such cells in the training set. In Figure 4, we showcase segmentation results on the external datasets. Our model accurately segments cells of different shapes and overlapping cells in the images of the ThalassemiaPBS and the Chula-RBC-12-Dataset datasets. For the MP-IDB dataset, it provides accurate segmentation masks for isolated cells, but does not separate rarely presented overlapping cells.

FIGURE 3
www.frontiersin.org

FIGURE 3. RedTell generalizes to patient images with unseen red blood cell types. (A) Our Mask R-CNN model was trained on red blood cell images of healthy individuals (controls) but can detect and segment different types robustly. (B) Stomatocytes, discocytes and echinocytes (from the left to the right) are segmented and show characteristic shapes. (C, D) RedTell segmentation generalizes to elongated cells appearing in SCD patients, while ignoring extra-cellular artifacts such as platelets (orange box in C). (E, F) The model also robustly segments cells of thalassemia patients with strongly irregular shapes.

FIGURE 4
www.frontiersin.org

FIGURE 4. RedTell’s segmentation model trained on the CoMMiTMenT dataset accurately segments RBCs in blood smear images from three different external datasets with strong domain shifts without need for re-training. Resized raw images (top) and segmentation results (bottom) show randomly selected images from the ThalassemiaPBS (A), Chula-RBC-12-Dataset (B), and Malaria parasite MP-IDB (C) dataset.

Vesicle detection

Vesicles are remains of cell organelles due to accelerated erythropoiesis and are indicative of anemia (Alaarg et al., 2013). The number of vesicles found thus can be considered as a meaningful single-cell feature. As discussed in Materials and methods, hyperparameters of the vesicle detection algorithm need to be set for every dataset. This process is straightforward without requiring much expertise or programming. For this case study, we annotated vesicles in 6 sample images randomly selected from the CoMMiTMenT dataset and the MemSID trial, 2 samples from each class (i.e., control, SCD and thalassemia). We use 3 images to determine the optimal parameters for the algorithm with respect to the accuracy, and the remaining 3 images serve for evaluation to assess the performance of the vesicle detection algorithm. We estimate the parameters manually by calculating detection accuracy and visualizing detection results. A vesicle is considered to be correctly detected if the predicted point has an Euclidean distance of ≤5 pixels to the annotated point. Our algorithm detects vesicles in the test set with high recall (sensitivity) of 0.89 and high precision of 0.96. We provide vesicle detection results on a test image of a SCD patient in Figure 5.

FIGURE 5
www.frontiersin.org

FIGURE 5. RedTell accurately detects vesicles in Ca2+ channel. (A, B) Test image in brightfield channel with overlaid predicted segmentation masks (A), and in Ca2+ channel (B). (C) Overlaid brightfield and Ca2+ channel with visualized vesicle detection. Most vesicles are detected correctly (white boxes), with only 3 detections being false negatives (red boxes) and no false positives.

Case study 1: extracted features differ significantly between diseases

Here, we investigate the reasonability of extracted features and whether they align with known RBC properties or provide new insights. For simplicity, we consider only shape features, mean Ca2+ intensity, and the number of vesicles extracted from 39 images of controls, 25 images of SCD and 16 images of thalassemia patients (Figure 3) from the CoMMiTMenT and MemSID datasets. We compare the distributions of extracted features for disease groups and find that they are discriminative on the image level and can potentially be used for disease diagnosis. A Kruskal-Wallis test (Kruskal and Wallis 1952) shows a significant difference (p < 10−6) between disease groups for all selected features except the solidity (H=7.16,p=0.03). For instance eccentricity, a measure of RBC cytoskeletal integrity and hemoglobin S aggregation (for SCD), of cells coming from SCD and thalassemia samples is on average higher for control group (Figure 6A). Higher values for the diseases can be explained by aggregation of Hemoglobin S known as sickling for SCD patients (Christoph et al., 2005) and by the instability of hemoglobins and premature clearance of RBCs in patients with thalassemia. Also, there is a significant difference in mean Ca2+ intensity between all three disease groups (Figure 6B) that could be explained by the hyperactivation of the plasma membrane Ca2+ ATPase resulting from its cleavage by Ca2+-dependent protease mu-calpain (Au, 1987). Moreover, in agreement with the previous observations (Bookchin et al., 1988) RBCs of SCD and thalassemia patients have more vesicles inside them to protect the cells from damage caused by enhanced Ca2+ leak through the plasma membrane (Figure 6C).

FIGURE 6
www.frontiersin.org

FIGURE 6. Case study 1—Single red blood cell features differ significantly between healthy and disease samples. Distribution of selected extracted features between SCD, thalassemia patients and healthy controls. H-statistics and p -values for the Kruskal-Wallis test are provided in the titles of the boxplots. The post hoc Dunn Bonferroni test shows significant difference (p < 0.05) between control-SCD and control-thalassemia pairs for eccentricity (A) and between all pairs for mean Ca2+ intensity (B) and number of vesicles (C).

Case study 2: discriminating cell types in control samples

We apply RedTell to classify stomatocyte, discocyte and echinocyte RBC subtypes (Figure 7A) in the samples from healthy controls. This case study also investigates how the number of annotated cells affects classification accuracy. The same CoMMiTMenT subset used in the segmentation experiments is annotated for classification. We first run RedTell to segment cells, to extract features and to compare detection labels with their corresponding groundtruth. We use bipartite graph matching (Bernardin and Stiefelhagen, 2008) to map centroids of the detected cells to the centroids of the ground truth annotations and match cells and determine their subtype label. We arrive at an unbalanced dataset, consisting of 8063 annotated cells, with 5% stomatocytes, 75% discocytes, and 20% echinocytes. We consider this a challenging problem for imbalanced multi-class classification as well as binary classification (one-versus-all for all three subtypes). We perform a random stratified imagewise 3 fold cross validation split into training (6000 cells) and validation (2063 cells) sets. Moreover, for every fold RedTell considers randomly stratified image-grouped sampled different fractions of training data to find the best model and evaluate on the test test. Figure 7B shows balanced accuracy achieved with the best classifiers from the RedTell Auto-ML module (decision tree, random forest or LightGBM) calculated on the test with confidence interval of 0.95 for multi-class and binary tasks. In both multi-class and binary settings, classification accuracy increases with the size of the training set. 200 annotated cells suffice to reach an average accuracy >70% for multi-class and >78% for binary classification. The highest classification accuracy of 0.92 is achieved for the echinocyte classification with 6000 cells in the training set. Increasing the size of the training set from 200 to 6000 images boosts multi-class accuracy from 72% to 82%, with an average improvement of 10%. For binary classification, this leads to an increase in accuracy between 3% and 5%. The difference between classification performance when training with 200 and 1000 annotated cells is less significant and falls within confidence intervals. Figure 7C shows the confusion matrix for the LightGBM model provided by RedTell’s Auto-ML module as the best classification model when training in multi-class settings with 4000 annotated cells. The model achieves 83% and 78% accuracy and balanced accuracy, respectively. Stomatocyte-discocyte-echinocyte fractions are given by 11%-75%-14% against true 6%-86%-8% in the test set, preserving the right order. Important features for the LightGBM classification model are provided in Figure 7D, ranked according to their relative importance and normalized by the value of the most important feature. Solidity and circularity from the shape feature set are most important, followed by three intensity-based features. Figure 7E shows the distribution of solidity and circularity across different cell subtypes. Both features are biologically reasonable and critical to differentiate echinocytes from discocytes and stomatocytes, as echinocytes are porous and have noticeably different shapes. Kruskal-Wallis and post hoc Dunn Bonferroni tests show significant difference insolidity between echinocytes and discocytes and between echinocytes and stomatocytes, and significant difference in circularity between all 3 cell subtypes. To differentiate stomatocytes from other cell subtypes the model relies on intensity-based features such as minimum intensity, skewness and energy.

FIGURE 7
www.frontiersin.org

FIGURE 7. Case study 2—RedTell discriminates red blood cell types with explainable classifiers. (A) Example images of cell subtypes. (B) 200 annotated cells suffice to reach >70% accuracy for supervised cell type classification. (C) Confusion matrix obtained for LightGBM model, the best classifier for multi-class classification given 2000 annotated cells. (D) Solidity is the most important feature for the LightGBM classifier. (E) Solidity and circularity is significantly reduced in echinocytes.

Case study 3: sickle cell classification

In our final case study, we use RedTell for sickle cell classification. Sickle cells have elongated forms (Figure 8A) first reported by (Herrick, 1910) and are used to diagnose SCD since then. We pick 25 images of sickle cell patients from the MemSID dataset and use RedTell to segment the cells, extract features, and randomly sample 300 cells for labeling. The class distribution in the resulting dataset is unbalanced, containing 80% normal cells and 20% sickle cells. Following the learning curve obtained for the binary classification problem in case study 2, we randomly select 200 annotated cells for training and 100 cells for testing. We use the RedTell’s Auto-ML module to train a classifier and obtain a random forest model, achieving balanced accuracy of 0.94, accuracy of 0.96 and both precision and recall of 0.90. Only 4 out of 100 cells are incorrectly classified (Figures 8C, D). One cell incorrectly predicted as normal starts to become sickle and is not elongated much, another one is on the border of the images and only partially visible Figure 8D. Cells incorrectly predicted as sickle cells are also not visible in their entirety being in the dark spots or out of focus. Partial cell visibility affects extracted features. In particular, it drastically affects the values of the classifier’s important features (Figure 8B). Random forest model uses biologically meaningful characteristics of cell morphology to differentiate between 2 cell types. The mean ranks of the two most important features, eccentricity and minor axis length, differ significantly (p=4.92*1026; p=3.68*1014 Mann-Whitney-U-Test (Mann and Whitney, 1947)) for normal and sickle cells (Figure 8E).

FIGURE 8
www.frontiersin.org

FIGURE 8. Case study—RedTell accurately detects sickle cells. (A) Examples of sickle and normal cells. (B) RedTell considers biologically meaningful features while performing classification. (C) Confusion matrix with only 4 cells being misclassified. (D) Misclassified cells are only partially visible. (E) Eccentricity and minor axis length show a significant difference between normal and sickle cells.

Discussion

We introduced RedTell, a package for automated analysis of microscopic images of RBCs. Its functionality includes RBC detection and segmentation, extraction of cell morphological properties and single cell classification. RedTell aims to accelerate research in hematology and can support diagnosis of disorders related to RBC morphologies. We showed the applicability of RedTell for three different case studies. We successfully applied it in three different case studies. First, we showed that extracted features have different distributions for healthy and anemia patients indicating feature usefulness for clinical decision making. Second, we developed a classifier to differentiate between different RBC subtypes. Finally, we used RedTell to classify sickle cells in SCD patient samples.

RedTell enables completely automated feature extraction. As for classification, it requires only expert annotation of a relatively small number of cells and no user interaction for training of the classifier. One of the main advantages of the tool is its interpretability. RedTell is interpretable tree-fold: 1) the results of cell segmentation in the brightfield channel, vesicle detection in the fluorescent channel and classification results are explicitly provided overlayed on the original images, 2) the extracted features are hand-crafted and reflect important RBCs properties, well understood and widely used by hematologists, 3) cell classification models trained within RedTell incorporate interpretability through ranking the features by their importance. Moreover, RedTell is the first RBC related work which suggests feature extraction from images of the fluorescent channels and introduces an algorithm for vesicle detection for a special case of Fluo-4 channel. In the following we will discuss single steps of the RedTell pipeline and suggest future work.

Segmentation. Mask R-CNN trained on the CoMMiTMenT dataset provides good results on the previously unseen images of three external datasets. We could not find any publicly available dataset, where our segmentation model would fail. However, to further increase its generalization ability to various datasets it would be reasonable to extend training of the Mask R-CNN with domain transfer methods (Zhang et al., 2022). Moreover, the segmentation part of RedTell can be extended by supporting other advanced machine learning methods, e.g., StarDist (Schmidt et al., 2018), providing a user a choice between different models to find an optimal one for a custom dataset.

Feature extraction. RedTell includes extraction of most of the common features widely used for the RBC analysis. Obviously, the feature set can be further extended with new features relevant to some specific user requirement, e.g., GLCM features calculated with different pixel distances and angles (Petrović et al., 2020), integral-geometry-based features (Gual-Arnau et al., 2015) or cell representation features provided by the Mask R-CNN model. Although such features are of limited interpretability, they can still improve classification accuracy.

Classification. We decided not to use Mask R-CNN for classification due to two reasons: 1) it requires a large amount of annotated data for a good performance, that is, necessarily labeled by an expert whose time is scarce and expensive, and 2) its performance is highly dependent on the domain and would work only with images generated under the same microscopic settings (Salehi, 2022). Although, we show that Mask R-CNN trained on our data qualitatively provides good segmentation results on the images from three external datasets (Figure 4), the same model would fail in the classification task since RBCs have different degrees of morphological details. We therefore include a classification module which takes the extracted features as input and thus allows building classification models for various datasets. Moreover, we decided to use decision tree classifiers as they resulted in high classification accuracy in previous work on RBC classification (Petrović et al., 2020) and are interpretable by design, ranking the features by their importance and hence allowing researchers to check the algorithmic logic. RedTell does not provide any classifier, but supports automated training of dataset and task specific classification algorithms. Due to numerous causes it is infeasible and beyond our goals to introduce a cell classifier, which works properly on all datasets. One important reason is interest in discrimination of different cell types (e.g., echinocyte versus discocyte or sickle versus normally shaped RBCs) depending on the available data and research question. Another limitation is varying morphology of the same cell type in microscopy images obtained under different acquisition settings. A good approach is development of a classifier specific to a given research question. Annotation of cell labels is a requirement for this task and RedTell facilitates annotation procedure. The tool includes tree-based classification algorithms which are interpretable by design. It is also possible to extend automated machine learning in RedTell with feature selection functionality to improve classification accuracy and include further classification algorithms, e.g., support vector machine (Devi et al., 2018) or deep learning-based method such as TabNet (Wong et al., 2021). However, such algorithms require advanced methods for interpretability (Altmann et al., 2010). Interpretability of the RedTell can further be improved by enabling local prediction explanation through SHAP (Lundberg and Lee, 2017) or LIME (Ribeiro et al., 2016), which provide information on what features and how they affect every single prediction.

The feature extraction and classification modules of RedTell can be applied to a broad variety of RBC research questions. In the future, we will apply RedTell to high throughput analyses, e.g., evaluating drug efficacy and assessing SCD progression by detecting and extracting features of sickle cells in blood samples of different patients at different timepoints, and hope that other researchers will follow. We also expect to include a graphical user interface for easier interaction with the user and extend RedTell functionality by providing automated analysis reports and data visualizations for the most common use cases as defined in consultation with RBC researchers.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author. We provide the datasets used for the software development, testing and case studies on Zenodo scientific sharing platform https://zenodo.org/record/7801430#.ZC1Ibi0Rr5k. The external datasets as described in the subsection “External datasets“ are publicly available.

Ethics statement

The studies involving human participants were reviewed and approved by the Medical Ethical Research Board of the University Medical Center Utrecht, Netherlands, under reference code 15/426M and Ethical Committee of Clinical Investigations of Hospital Clinic, Spain (IDIBAPS) under reference code 2013/8436, Ethics committee of Canton Zurich (KEK-ZH 2015-0297). The patients/participants provided their written informed consent to participate in this study.

Author contributions

AS, MB, NN, AB, and CM contributed to conception and design of the study. AM organized the database. MB implemented the software. AS, MB, and CM performed the statistical analysis. AS and MB wrote the first draft of the manuscript. AS, MB, AB, and CM wrote sections of the manuscript. All authors contributed to manuscript revision, read, and approved the submitted version.

Funding

CM has received funding from the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation programme (Grant agreement No. 866411).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Akrimi, J. A., Suliman, A., George, L. E., and Ahmad, A. R. (2014). “Classification red blood cells using support vector machine,” in Proceedings of the 6th International Conference on Information Technology and Multimedia, Putrajaya, Malaysia, 18-20 November 2014, 265–269.

CrossRef Full Text | Google Scholar

Alaarg, , Amr, A., Schiffelers, R. M., van Solinge, W. A., and Wijk, R. V. (2013). Red blood cell vesiculation in hereditary hemolytic anemia. Frontiers in Physiology 4, 365.

PubMed Abstract | CrossRef Full Text | Google Scholar

Altmann, A., Toloşi, L., Sander, O., and Lengauer, T. (2010). Permutation importance: A corrected feature importance measure. Bioinformatics 26, 1340–1347. doi:10.1093/bioinformatics/btq134

PubMed Abstract | CrossRef Full Text | Google Scholar

Alzubaidi, L., Fadhel, M. A., Al-Shamma, O., Zhang, J., and Duan, Y. (2020). Deep learning models for classification of red blood cells in microscopy images to aid in sickle cell anemia diagnosis. Electronics 9, 427. doi:10.3390/electronics9030427

CrossRef Full Text | Google Scholar

Arik, S. O., and Pfister, T. (2019). TabNet: Attentive interpretable tabular learning. arXiv [cs.LG]. Available at: http://arxiv.org/abs/1908.07442.

Google Scholar

Au, K. S. (1987). Activation of erythrocyte membrane Ca2+-ATPase by calpain. Biochim. Biophys. Acta 905, 273–278. doi:10.1016/0005-2736(87)90455-x

PubMed Abstract | CrossRef Full Text | Google Scholar

Bernardin, K., and Stiefelhagen, R. (2008). Evaluating multiple object tracking performance: The CLEAR MOT metrics. EURASIP J. Image Video Process. 2008, 1–10. doi:10.1155/2008/246309

CrossRef Full Text | Google Scholar

Bookchin, R. M., Ortiz, O. E., Shalev, O., Tsurel, S., Rachmilewitz, E. A., Hockaday, A., et al. (1988). Calcium transport and ultrastructure of red cells in beta-thalassemia intermedia. Blood 72, 1602–1607. doi:10.1182/blood.v72.5.1602.1602

PubMed Abstract | CrossRef Full Text | Google Scholar

Bouyer, G., Cueff, A., Egée, S., Kmiecik, J., Maksimova, Y., Glogowska, E., et al. (2011). Erythrocyte peripheral type benzodiazepine receptor/voltage-dependent anion channels are upregulated by Plasmodium falciparum. Blood 118, 2305–2312. doi:10.1182/blood-2011-01-329300

PubMed Abstract | CrossRef Full Text | Google Scholar

Chiao, J.-Y., Chen, K.-Y., Liao, K. Y.-K., Hsieh, P.-H., Zhang, G., and Huang, T.-C. (2019). Detection and classification the breast tumors using mask R-CNN on sonograms. Medicine 98, e15200. doi:10.1097/MD.0000000000015200

PubMed Abstract | CrossRef Full Text | Google Scholar

Christoph, G. W., Hofrichter, J., and Eaton, W. A. (2005). Understanding the shape of sickled red cells. Biophys. J. 88, 1371–1376. doi:10.1529/biophysj.104.051250

PubMed Abstract | CrossRef Full Text | Google Scholar

Chy, T. S., and Anisur Rahaman, M. (2019). “A comparative analysis by KNN, SVM & ELM classification to detect sickle cell anemia,” in 2019 International Conference on Robotics,Electrical and Signal Processing Techniques (ICREST), Dhaka, Bangladesh, 10-12 January 2019, 455–459.

Google Scholar

Coffman, V. C., and Wu, J.-Q. (2012). Counting protein molecules using quantitative fluorescence microscopy. Trends biochem. Sci. 37, 499–506. doi:10.1016/j.tibs.2012.08.002

PubMed Abstract | CrossRef Full Text | Google Scholar

Das, P. K., Meher, S., Panda, R., and Abraham, A. (2020). A review of automated methods for the detection of sickle cell disease. IEEE Rev. Biomed. Eng. 13, 309–324. doi:10.1109/RBME.2019.2917780

PubMed Abstract | CrossRef Full Text | Google Scholar

Deshpande, N. M., Gite, S., and Aluvalu, R. (2021). A review of microscopic analysis of blood cells for disease detection with AI perspective. PeerJ Comput. Sci. 7, e460. doi:10.7717/peerj-cs.460

PubMed Abstract | CrossRef Full Text | Google Scholar

Devi, S. S., Roy, A., Singha, J., Sheikh, S. A., and Laskar, R. H. (2018). Malaria infected erythrocyte classification based on a hybrid classifier using microscopic images of thin blood smear. Multimed. Tools Appl. 77, 631–660. doi:10.1007/s11042-016-4264-7

CrossRef Full Text | Google Scholar

Dhieb, N., Ghazzai, H., Besbes, H., and Massoud, Y. (2019). “An automated blood cells counting and classification Framework using mask R-CNN deep learning model,” in 2019 31st International Conference on Microelectronics (ICM), Cairo, Egypt, 15-18 December 2019, 300–303.

CrossRef Full Text | Google Scholar

Fermo, E., Bogdanova, A., Petkova-Kirova, P., Zaninoni, A., Marcello, A. P., Makhro, A., et al. (2017). Gardos channelopathy”: A variant of hereditary stomatocytosis with complex molecular regulation. Sci. Rep. 7, 1744. doi:10.1038/s41598-017-01591-w

PubMed Abstract | CrossRef Full Text | Google Scholar

Girshick, R., Donahue, J., Darrell, T., and Malik, J. (2014). Rich feature hierarchies for accurate object detection and semantic segmentation. Available at: http://openaccess.thecvf.com/content_cvpr_2014/html/Girshick_Rich_Feature_Hierarchies_2014_CVPR_paper.html.

CrossRef Full Text | Google Scholar

Girshick, R. (2015). “Fast r-cnn,” in Proceedings of the IEEE international conference on computer vision, 1440–1448.

CrossRef Full Text | Google Scholar

Gual-Arnau, X., Herold-García, S., and Simó, A. (2015). Erythrocyte shape classification using integral-geometry-based methods. Med. Biol. Eng. 53, 623–633. doi:10.1007/s11517-015-1267-x

PubMed Abstract | CrossRef Full Text | Google Scholar

Hänggi, P., Makhro, A., Gassmann, M., Schmugge, M., Goede, J. S., Speer, O., et al. (2014). Red blood cells of sickle cell disease patients exhibit abnormally high abundance of N-methyl D-aspartate receptors mediating excessive calcium uptake. Br. J. Haematol. 167, 252–264. doi:10.1111/bjh.13028

PubMed Abstract | CrossRef Full Text | Google Scholar

He, K., Gkioxari, G., Dollar, P., and Girshick, R. (2017). “Mask R-CNN,” in 2017 IEEE International Conference on Computer Vision (ICCV) (IEEE), Venice, Italy, 22-29 October 2017. doi:10.1109/iccv.2017.322

CrossRef Full Text | Google Scholar

He, X., Zhao, K., and Chu, X. (2019). AutoML: A survey of the state-of-the-art. arXiv [cs.LG]. Available at: https://www.sciencedirect.com/science/article/pii/S0950705120307516.

Google Scholar

Herrick, J. B. (1910). Peculiar elongated and sickle-shaped red blood corpuscles in A case of severe anemia. Arch. Intern. Med. VI, 517–521. doi:10.1001/archinte.1910.00050330050003

CrossRef Full Text | Google Scholar

Hertz, L., Huisjes, R., Llaudet-Planas, E., Petkova-Kirova, P., Makhro, A., Danielczok, J. G., et al. (2017). Is increased intracellular calcium in red blood cells a common component in the molecular mechanism causing anemia? Front. Physiol. 8, 673. doi:10.3389/fphys.2017.00673

PubMed Abstract | CrossRef Full Text | Google Scholar

Huang, Z., Zhong, Z., Sun, L., and Huo, Q. (2019). “Mask R-CNN with pyramid attention network for scene text detection,” in 2019 IEEE Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 07-11 January 2019, 764–772.

CrossRef Full Text | Google Scholar

Kaestner, L., Tabellion, W., Weiss, E., Bernhardt, I., and Lipp, P. (2006). Calcium imaging of individual erythrocytes: Problems and approaches. Cell Calcium 39, 13–19. doi:10.1016/j.ceca.2005.09.004

PubMed Abstract | CrossRef Full Text | Google Scholar

Kaestner, L., Bogdanova, A., and Egee, S. (2020). “Calcium channels and calcium-regulated channels in human red blood cells,” in Calcium signaling. Editor M. S. Islam (Cham: Springer International Publishing), 625–648.

CrossRef Full Text | Google Scholar

Lew, V. L., Hockaday, A., Sepulveda, M. I., Somlyo, A. P., Somlyo, A. V., Ortiz, O. E., et al. (1985). Compartmentalization of sickle-cell calcium in endocytic inside-out vesicles. Nature 315, 586–589. doi:10.1038/315586a0

PubMed Abstract | CrossRef Full Text | Google Scholar

Loddo, A., Di Ruberto, C., Kocher, M., and Prod’Hom, G. (2019). “MP-IDB: The malaria parasite image database for image processing and analysis,” in Processing and analysis of biomedical information (Cham: Springer International Publishing), 57–65.

CrossRef Full Text | Google Scholar

Loh, D. R., Yong, W. X., Yapeter, J., Subburaj, K., and Chandramohanadas, R. (2021). A deep learning approach to the screening of malaria infection: Automated and rapid cell counting, object detection and instance segmentation using Mask R-CNN. Comput. Med. Imaging Graph. 88, 101845. doi:10.1016/j.compmedimag.2020.101845

PubMed Abstract | CrossRef Full Text | Google Scholar

Lundberg, S. M., and Lee, S.-I. (2017). A unified approach to interpreting model predictions. Adv. Neural Inf. Process. Syst. Available at: https://proceedings.neurips.cc/paper/2017/hash/8a20a8621978632d76c43dfd28b67767-Abstract.html.

Google Scholar

Makhro, A., Huisjes, R., Verhagen, L. P., Mañú-Pereira, M. D. M., Llaudet-Planas, E., Petkova-Kirova, P., et al. (2016). Red cell properties after different modes of blood transportation. Front. Physiol. 7, 288. doi:10.3389/fphys.2016.00288

PubMed Abstract | CrossRef Full Text | Google Scholar

Makhro, A., Kaestner, L., and Bogdanova, A. (2017). NMDA receptor activity in circulating red blood cells: Methods of detection. Methods Mol. Biol. 1677, 265–282. doi:10.1007/978-1-4939-7321-7_15

PubMed Abstract | CrossRef Full Text | Google Scholar

Mann, H. B., and Whitney, D. R. (1947). On a test of whether one of two random variables is stochastically larger than the other. Ann. Math. Stat. 18, 50–60. doi:10.1214/aoms/1177730491

CrossRef Full Text | Google Scholar

Muravyov, A., and Tikhomirova, I. (2012). “Role Ca2+ in mechanisms of the red blood cells microrheological changes,” in Calcium signaling. Editor M. S. Islam (Dordrecht: Springer Netherlands), 1017–1038.

PubMed Abstract | CrossRef Full Text | Google Scholar

Naruenatthanaset, K., Chalidabhongse, T. H., Palasuwan, D., Anantrasirichai, N., and Palasuwan, A. (2020). Red Blood Cell Segmentation with Overlapping Cell Separation and Classification on Imbalanced Dataset. arXiv [eess.IV]. Available at: http://arxiv.org/abs/2012.01321.

Google Scholar

Parab, M. A., and Mehendale, N. D. (2021). Red blood cell classification using image processing and CNN. SN Comput. Sci. 2, 70. doi:10.1007/s42979-021-00458-2

CrossRef Full Text | Google Scholar

Petrović, N., Moyà-Alcover, G., Jaume-i-Capó, A., and González-Hidalgo, M. (2020). Sickle-cell disease diagnosis support selecting the most appropriate machine learning method: Towards a general and interpretable approach for cell morphology analysis from microscopy images. Comput. Biol. Med. 126, 104027. doi:10.1016/j.compbiomed.2020.104027

PubMed Abstract | CrossRef Full Text | Google Scholar

Piety, N. Z., Gifford, S. C., Yang, X., and Shevkoplyas, S. S. (2015). Quantifying morphological heterogeneity: A study of more than 1 000 000 individual stored red blood cells. Vox Sang. 109, 221–230. doi:10.1111/vox.12277

PubMed Abstract | CrossRef Full Text | Google Scholar

Ren, S., He, K., Girshick, R., and Sun, J. (2015). Faster r-cnn: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. Available at: https://proceedings.neurips.cc/paper/2015/hash/14bfa6bb14875e45bba028a21ed38046-Abstract.html.

Google Scholar

Ribeiro, M. T., Singh, S., and Guestrin, C. (2016). “why should I trust you?: Explaining the predictions of any classifier,” in Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, California, San Francisco, USA, August 13 - 17, 2016. doi:10.1145/2939672.2939778

CrossRef Full Text | Google Scholar

Ronneberger, O., Fischer, P., and Brox, T. (2015). “U-Net: Convolutional networks for biomedical image segmentation,” in Medical image computing and computer-assisted intervention – miccai 2015 (Cham: Springer International Publishing), 234–241.

CrossRef Full Text | Google Scholar

Routt, A. H., Yang, N., Piety, N. Z., Lu, M., and Shevkoplyas, S. S. (2023). Deep ensemble learning enables highly accurate classification of stored red blood cell morphology. Sci. Rep. 13, 3152. doi:10.1038/s41598-023-30214-w

PubMed Abstract | CrossRef Full Text | Google Scholar

Sadafi, A., Radolko, M., Serafeimidis, I., and Hadlak, S. (2018). “Red blood cells segmentation: A fully convolutional network approach,” in 2018 IEEE Intl Conf on Parallel & Distributed Processing with Applications, Ubiquitous Computing & Communications, Big Data & Cloud Computing, Social Computing & Networking, Sustainable Computing & Communications (ISPA/IUCC/BDCloud/SocialCom/SustainCom), Melbourne, VIC, Australia, 11-13 December 2018. 911–914.

CrossRef Full Text | Google Scholar

Salehi, , Raheleh, , Sadafi, A., Gruber, A., Lienemann, P., Navab, N., et al. (2022). “Unsupervised cross-domain feature extraction for single blood cell image classification,” in Medical Image Computing and Computer Assisted Intervention–MICCAI 2022: 25th International Conference, Singapore, 1September 18–22, 2022, Proceedings, Part III (Cham: Springer Nature Switzerland), 739–748.

CrossRef Full Text | Google Scholar

Sadafi, A., Koehler, N., Makhro, A., Bogdanova, A., Navab, N., Marr, C., et al. (2019). “Multiclass deep active learning for detecting red blood cell subtypes in brightfield microscopy,” in Medical image computing and computer assisted intervention – miccai 2019 (Cham: Springer International Publishing), 685–693.

CrossRef Full Text | Google Scholar

Sadafi, A., Makhro, A., Bogdanova, A., Navab, N., Peng, T., Albarqouni, S., et al. (2020). “Attention based multiple instance learning for classification of blood cell disorders,” in International Conference on Medical Image Computing and Computer-Assisted Intervention, Lima, Peru, October 4–8, 2020, 246–256.

CrossRef Full Text | Google Scholar

Savkare, S. S., and Narote, S. P. (2015). “Blood cell segmentation from microscopic blood images,” in 2015 International Conference on Information Processing (ICIP), Pune, India, 16-19 December 2015, 502–505.

CrossRef Full Text | Google Scholar

Schmidt, U., Weigert, M., Broaddus, C., and Myers, G. (2018). “Cell detection with star-convex polygons,” in Medical image computing and computer assisted intervention – miccai 2018 (Cham: Springer International Publishing), 265–273.

CrossRef Full Text | Google Scholar

Schneider, C. A., Rasband, W. S., and Eliceiri, K. W. (2012). NIH image to ImageJ: 25 years of image analysis. Nat. Methods 9, 671–675. doi:10.1038/nmeth.2089

PubMed Abstract | CrossRef Full Text | Google Scholar

Sharif, J. M., Miswan, M. F., Ngadi, M. A., Salam, M. S. H., and bin Abdul Jamil, M. M. (2012). “Red blood cell segmentation using masking and watershed algorithm: A preliminary study,” in 2012 International Conference on Biomedical Engineering (ICoBE), Penang, Malaysia, 27-28 February 2012, 258–262.

CrossRef Full Text | Google Scholar

Sinha, A., Chu, T. T. T., Dao, M., and Chandramohanadas, R. (2015). Single-cell evaluation of red blood cell bio-mechanical and nano-structural alterations upon chemically induced oxidative stress. Sci. Rep. 5, 9768. doi:10.1038/srep09768

PubMed Abstract | CrossRef Full Text | Google Scholar

Snoek, J., Larochelle, H., and Adams, R. P. (2012). Practical bayesian optimization of machine learning algorithms. Adv. Neural Inf. Process. Syst. Available at: https://proceedings.neurips.cc/paper/4522-practical-bayesian-optimization.

Google Scholar

Song, W., Huang, P., Wang, J., Shen, Y., Zhang, J., Lu, Z., et al. (2021). Red blood cell classification based on attention residual feature pyramid network. Front. Med. 8, 741407. doi:10.3389/fmed.2021.741407

PubMed Abstract | CrossRef Full Text | Google Scholar

Stirling, D. R., Swain-Bowden, M. J., Lucas, A. M., Carpenter, A. E., Cimini, B. A., and Goodman, A. (2021). CellProfiler 4: Improvements in speed, utility and usability. BMC Bioinforma. 22, 433. doi:10.1186/s12859-021-04344-9

CrossRef Full Text | Google Scholar

Suykens, J. A. K., and Vandewalle, J. (1999). Least squares support vector machine classifiers. Neural process. Lett. 9, 293–300. doi:10.1023/a:1018628609742

CrossRef Full Text | Google Scholar

Tomari, R., Zakaria, W. N. W., Jamil, M. M. A., Nor, F. M., and Fuad, N. F. N. (2014). Computer aided system for red blood cell classification in blood smear image. Procedia Comput. Sci. 42, 206–213. doi:10.1016/j.procs.2014.11.053

CrossRef Full Text | Google Scholar

Tyas, D. A., Hartati, S., Harjoko, A., and Ratnaningsih, T. (2020). Morphological, texture, and color feature analysis for erythrocyte classification in thalassemia cases. IEEE Access 8, 69849–69860. doi:10.1109/access.2020.2983155

CrossRef Full Text | Google Scholar

Tyas, D. A., Ratnaningsih, T., Harjoko, A., and Hartati, S. (2022). Erythrocyte (red blood cell) dataset in thalassemia case. Data Brief. 41, 107886. doi:10.1016/j.dib.2022.107886

PubMed Abstract | CrossRef Full Text | Google Scholar

van der Walt, S., Schönberger, J. L., Nunez-Iglesias, J., Boulogne, F., Warner, J. D., Yager, N., et al. (2014). scikit-image: image processing in Python. PeerJ 2, e453. doi:10.7717/peerj.453

PubMed Abstract | CrossRef Full Text | Google Scholar

van Griethuysen, J. J. M., Fedorov, A., Parmar, C., Hosny, A., Aucoin, N., Narayan, V., et al. (2017). Computational radiomics system to decode the radiographic phenotype. Cancer Res. 77, e104–e107. doi:10.1158/0008-5472.CAN-17-0339

PubMed Abstract | CrossRef Full Text | Google Scholar

Wagner, S. J., Matek, C., Shetab Boushehri, S., Boxberg, M., Lamm, L., Sadafi, A., et al. (2022). Make deep learning algorithms in computational pathology more reproducible and reusable. Nat. Med. 28, 1744–1746. doi:10.1038/s41591-022-01905-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Wang, H., Obeidy, P., Wang, Z., Zhao, Y., Wang, Y., Su, Q. P., et al. (2022). Fluorescence-coupled micropipette aspiration assay to examine calcium mobilization caused by red blood cell mechanosensing. Eur. Biophys. J. 51, 135–146. doi:10.1007/s00249-022-01595-z

PubMed Abstract | CrossRef Full Text | Google Scholar

Wong, A., Anantrasirichai, N., Chalidabhongse, T. H., Palasuwan, D., Palasuwan, A., and Bull, D. (2021). Analysis of vision-based abnormal red blood cell classification. arXiv [cs.CV]. Available at: http://arxiv.org/abs/2106.00389.

Google Scholar

Zhang, D., Song, Y., Zhang, F., O’Donnell, L., and Huang, H. (2022). Unsupervised instance segmentation in microscopy images via panoptic domain adaptation and task Re-weighting (supplementary material). Available at: https://openaccess.thecvf.com/content_CVPR_2020/supplemental/Liu_Unsupervised_Instance_Segmentation_CVPR_2020_supplemental.pdf.

Google Scholar

Keywords: microscopic image analysis, single red blood cells, morphological features extraction, vesicle detection, deep learning, interpretable machine learning, segmentation, classification

Citation: Sadafi A, Bordukova M, Makhro A, Navab N, Bogdanova A and Marr C (2023) RedTell: an AI tool for interpretable analysis of red blood cell morphology. Front. Physiol. 14:1058720. doi: 10.3389/fphys.2023.1058720

Received: 30 September 2022; Accepted: 13 April 2023;
Published: 26 May 2023.

Edited by:

Paulo E. Arratia, University of Pennsylvania, United States

Reviewed by:

Sergey S. Shevkoplyas, University of Houston, United States
Yuhao Qiang, Massachusetts Institute of Technology, United States

Copyright © 2023 Sadafi, Bordukova, Makhro, Navab, Bogdanova and Marr. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Carsten Marr, carsten.marr@helmholtz-munich.de

These authors have contributed equally to this work

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.