Open Access
13 November 2020 Automatic detection and characterization of quantitative phase images of thalassemic red blood cells using a mask region-based convolutional neural network
Author Affiliations +
Abstract

Significance: Label-free quantitative phase imaging is a promising technique for the automatic detection of abnormal red blood cells (RBCs) in real time. Although deep-learning techniques can accurately detect abnormal RBCs from quantitative phase images efficiently, their applications in diagnostic testing are limited by the lack of transparency. More interpretable results such as morphological and biochemical characteristics of individual RBCs are highly desirable.

Aim: An end-to-end deep-learning model was developed to efficiently discriminate thalassemic RBCs (tRBCs) from healthy RBCs (hRBCs) in quantitative phase images and segment RBCs for single-cell characterization.

Approach: Two-dimensional quantitative phase images of hRBCs and tRBCs were acquired using digital holographic microscopy. A mask region-based convolutional neural network (Mask R-CNN) model was trained to discriminate tRBCs and segment individual RBCs. Characterization of tRBCs was achieved utilizing SHapley Additive exPlanation analysis and canonical correlation analysis on automatically segmented RBC phase images.

Results: The implemented model achieved 97.8% accuracy in detecting tRBCs. Phase-shift statistics showed the highest influence on the correct classification of tRBCs. Associations between the phase-shift features and three-dimensional morphological features were revealed.

Conclusions: The implemented Mask R-CNN model accurately identified tRBCs and segmented RBCs to provide single-RBC characterization, which has the potential to aid clinical decision-making.

1.

Introduction

Information about red blood cell (RBC) morphology is crucial to reach a diagnosis of blood disorders because RBC morphologies often change due to altered membrane lipid composition, iron deficiency, or metabolic abnormalities.1,2 The current diagnostic procedure for many blood-related diseases is the blood smear test in which the morphology of stained RBCs is examined under a light microscope. Staining of the blood film is usually required to achieve adequate image contrast for proper microscopic examination. However, the staining process is time-consuming and may cause variations in the resulting contrast, both of which impede the blood smear test to be used as an automatic and high-throughput procedure.3

A promising label-free microscopy technique for high-throughput examination of blood smears is quantitative phase imaging (QPI).46 QPI is capable of mapping in two dimensions the phase shift caused by cellular constituents when a source light beam transmits through a transparent cell. The measured phase shift represents the line integral of the refractive index (RI) contrast between the specimen and its environment along the light path, which is parallel to the optical axis and corresponds to the physical thickness of the specimen. Since RBCs are mostly composed of hemoglobin and are without many organelles, the RI within an RBC is approximately constant and proportional to its hemoglobin concentration.7 Therefore, the phase shift of each pixel in an RBC phase image combines information about the hemoglobin concentration and thickness of the RBC at that position. Furthermore, integrating the phase shift over all pixels of an RBC phase image generates the so-called optical volume, which is proportional to the total dry mass of hemoglobin8 in the RBC. The optical volume has been shown to be a major feature in identifying hypochromic RBCs, as seen in iron deficiency anemia9,10 and thalassemia.10,11 In addition, phase-shift statistics and two-dimensional (2-D) morphological features, such as the projected area, perimeter, and lengths of major and minor axes, have been extracted from quantitative phase images of RBCs and have been used to classify blood disorders, including hereditary spherocytosis,4,9 malaria,1214 and sickle cell disease,1517 where RBCs possess abnormal hemoglobin content and show abnormal shape. These advances highlight a major advantage of using QPI for the diagnosis of RBC-related diseases: simultaneously quantifying diagnostically relevant hemoglobin content and morphological information without exogenous labels.

To further develop QPI to become a practical tool for automatic detection of RBC-related diseases, the conventional image processing approach18,19 consists of multiple steps including the segmentation of individual RBCs from QPI data, extraction and selection of features, and training of a classifier. For the automatic segmentation of multiple RBCs from a single quantitative phase image, Yi et al.20 have proposed a method based on marker-controlled watershed transform algorithm. This method addressed the issues of overestimation and underestimation in simple thresholding such as the Otsu method and is less susceptible to noise. However, the proposed algorithm is time-consuming and hence not suitable for real-time processing in clinical applications. As for selecting an optimal set of features to achieve accurate classification, special domain knowledge is required, and this selection is specific for each disease since different diseases present distinctive characteristics. The efficiency of both the development of an automatic tool and the inference of the results by the tool is not optimal using the conventional image processing approach.

The efficiency issues can be addressed by emerging deep-learning techniques, such as convolutional neural networks (CNNs).21,22 In addition to fast inference without time-consuming image processing, CNNs greatly simplify the development of image-based automatic diagnostic tools since feature extraction and selection are automatic and do not require any domain knowledge. Various CNN models have been applied to quantitative phase images to accurately classify 19 bacterial species,23 distinguish anthrax spores from five Bacillus species for biodefense,24 and detect breast cancer cells in whole blood samples for early cancer diagnosis and treatment response assessment.25 Other CNN models have been developed to segment healthy RBCs (hRCBs) in quantitative phase images26 and to detect RBCs with sickle cell disease2729 or RBCs infected with malaria30 based on stained blood smear images. However, no CNN model has been established to segment and detect thalassemic RBCs (tRBCs) in quantitative phase images. Additionally, the black-box nature of deep-learning techniques has limited the interpretation of results, which has impeded the adoption of CNN models in medicine. In diagnostics, a comprehensive analysis of single cells is highly desirable to provide detailed morphological and chemical characteristics, thus aiding both the diagnosis and treatment planning. Hence, single-cell segmentation is a critical step to obtain the regions of interest (ROI) and to subsequently characterize single cells. An efficient deep-learning technique, mask region-based CNN (Mask R-CNN), has been proposed to automatically achieve instance segmentation of every recognized object.31 Since Mask R-CNN can achieve detection and segmentation simultaneously, it has been applied to automatically analyze fluorescence images of immune cells32 and to detect and segment nuclei in images of hematoxylin and eosin-stained histopathology slides of multiple organs33,34 and cancer tissues.33,35

In this paper, we demonstrate the implementation of Mask R-CNN for the automatic discrimination between quantitative phase images of hRBCs and those of tRBCs. Thalassemia is an inherited anemia that has a high prevalence (3.6%)36 worldwide. Owing to the early onset and detrimental effects of thalassemia, a minimally invasive diagnostic method for infants is needed to predict the disease severity and, then, to adopt timely and suitable treatments. QPI is well suited for this application because it requires only a drop of blood sample. In the current study, single-shot digital holographic microscopy (DHM) was used to acquire 2-D quantitative phase images of RBCs collected from thalassemia-minor patients and healthy subjects. A Mask R-CNN model was then trained to perform end-to-end detection of tRBCs. The classification accuracy of the model was benchmarked against an optimal classifier, which was built with the XGBoost technique and trained with 15 single-cell features, including phase-shift statistics, 2-D morphological features, and textural features extracted from manually segmented RBC phase images. To demonstrate the advantage of the Mask R-CNN model that provides instance segmentation, we extracted single-cell features from automatically segmented RBC phase images and trained a second XGBoost classifier to discriminate tRBCs from healthy ones. The features that contributed the most to successfully distinguish tRBCs were obtained using the SHapley Additive exPlanation (SHAP) analysis,37,38 which is based on game theory. Moreover, we analyzed correlations between the 2-D QPI features and three-dimensional (3-D) morphological features obtained with optical diffraction tomography of the same RBCs. Our findings suggest the use of QPI as a useful tool to capture 3-D morphological characteristics of abnormal RBCs, such as those seen in patients with thalassemia.

2.

QPI Instrumentation and Data Acquisition

We acquired 2-D quantitative phase images of RBCs using off-axis DHM. Details of the optical setup are described in Ref. 39 and a schematic diagram is shown in Fig. 1(a). A laser beam from a 532-nm continuous-wave laser goes through an oil-immersion condenser to generate a planar wavefront, illuminating the specimen at normal incidence. An objective lens (Olympus UPLSAPO 100XO, 1.4 NA) collects the transmitted field to project a magnified image of the specimen on a CMOS camera sensor (GZL-CL-41C6M-C, Gazelle, Point Grey). A transmission grating (Edmund, 80  grooves/mm) located after the tube lens (L5) generates multiple diffracted beams to provide a uniform reference beam that goes through the same components as the sample beam, before interfering with the sample beam at the camera sensor. A window is placed at the Fourier plane of L6 to allow the full zeroth-order beam and the DC signal of the first-order beam to pass. As shown in the inset of Fig. 1(a), the field of view in the zeroth-order beam is shifted from that in the first-order beam by moving the grating slightly off the rear focal plane of L5. At the camera sensor plane, the image of the specimen in the zeroth-order beam (henceforth the sample beam) overlaps with an empty area in the first-order beam (henceforth the reference beam) to form interference images of the specimen [as shown in Fig. 1(b)], which are recorded with a transverse magnification of approximately ×85. The same optical setup has been used to acquire quantitative phase images of RBCs under various incident angles by rotating a galvanometer mirror (GM, Cambridge Technology Inc.) located at a conjugate plane of the specimen.11 3-D RI maps have been reconstructed based on optical diffraction tomography with direct interpolation in the Fourier domain and the positivity constraint.40 An example of a reconstructed phase image and RI map of an hRBC are demonstrated in Figs. 1(c) and 1(d).

Fig. 1

Schematic diagram of the DHM and imaging results of a hRBC. (a) Optical setup; upper inset: the red and the green areas depict the field of view of the first-order and the zeroth-order beam, respectively. Lower inset: the spatial frequency extent of the first-order and the zeroth-order beam is shown in red and green, respectively; dots in the center indicate the DC component. An interferogram of an hRBC is shown in (b), and the corresponding phase image and cross-sectional views of the RI tomogram are shown in (c) and (d), respectively. Po, polarizer; QWP, quarter-wave plate; OL1 and OL2, objective lens; L1 to L7, lenses; GM, galvanometer mirror.

JBO_25_11_116502_f001.png

This study was approved by the Institutional Review Board of National Taiwan University Hospital, and an informed consent was obtained from each subject. We imaged 210 hRBCs from 11 healthy subjects and 475 tRBCs from 29 adult thalassemia-minor patients. About 10 interference images were acquired from each RBC to expand the dataset. In total, 2001 hRBCs and 4268 tRBCs interference images were collected. Phase images were retrieved from the interference images by bandpass filtering in the spatial frequency domain and inverse Fourier transform without zero-padding,41 followed by a discrete cosine transform-based phase unwrapping method.42 We implemented the phase retrieval and unwrapping processes by parallel computing on a NVIDIA GTX 1080 Ti GPU and an Intel Core i7-3820 CPU, enhancing the computation speed to about 289 fps.

Since the number of tRBC images was about two times more than that of hRBC images, we adopted data augmentation by random rotation, to enlarge the hRBC dataset to balance data for XGBoost classifier training and Mask R-CNN model building. Furthermore, the operation of random rotation increased the variation in datasets, to prevent overfitting during Mask R-CNN model training.4345 The final total number of hRBC images was equal to that of tRBC.

3.

Automatic Detection and Delineation of RBCs

3.1.

Development of the Mask R-CNN Model

The Mask R-CNN model was developed using the Deep-Learning-Framework Keras (v2.3.1) and TensorFlow (v1.12.0) as the backend, and accelerated using NVIDIA CUDA (v10.0) and cuDNN (v7.5). The architecture of the Mask R-CNN model, shown in Fig. 2, can be split into two stages. Tasks of the first stage were image scanning and proposal generation. The tasks of the second stage were object classification,46 bounding box coordinates generation,46 and mask generation31 for each RBC in a quantitative phase image. The two stages were combined into a single workflow that can be trained end-to-end, achieving pipeline optimization and optimal performance.31 A 101-layer CNN backbone with a feature pyramid network, ResNet-101-FPN,46 was used to extract high-level image features via back propagation. To feed the phase images into the backbone to obtain feature maps, single-channel phase maps were replicated into three-channel images (1024×1024×3 channels). A region proposal network was then applied to predict candidate bounding boxes of objects (RBCs) from the high-level image feature maps. An RoIAlign layer scaled every candidate bounding box to the same size (7×7×3 channels) using bilinear interpolation to retain floating-point numbers for subsequent tasks. The scaled ROI feature maps were then passed to fully connected layers to infer the object class and to determine bounding box offset values. In parallel, masks (28×28×3 classes) were generated from the scaled ROI feature maps in the mask branch with multiple convolutional layers.

Fig. 2

Illustration of Mask R-CNN architecture for instance segmentation and classification. A CNN extracts image features, and a region proposal network generates ROIs containing individual RBCs. The RoIAlign layer scales every ROI to the same size and passes the ROIs to each branch to perform RBC detection and segmentation.

JBO_25_11_116502_f002.png

We partitioned the augmented dataset described in Sec. 2 into training, validation, and test datasets with a ratio of 3:1:1. The initial learning rate, batch size, and steps per epoch were 0.002, 64, and 80, respectively. The categorical cross-entropy of classification and the intersection over union (IoU)47 of masks were calculated on the validation dataset to monitor the performance of classification and segmentation, respectively.

3.2.

Development of the Benchmark Classifier Using XGBoost

To benchmark the performance of the proposed Mask R-CNN model in distinguishing tRBCs from hRBCs, we built an optimal classifier with features extracted from manually segmented RBC quantitative phase images. Table 1 lists the 15 features that were used. The optical and morphological features were readily obtained from each segmented RBC phase image, and the textural features were calculated using the gray-level co-occurrence matrix (GLCM)48,49 of each segmented RBC image to quantify its spatial hemoglobin distribution. GLCM is a statistical method to describe the textural structure of an image by statistically sampling the pattern of gray-level occurrences in relation to other gray levels. The benchmark classifier was built using XGBoost (eXtreme Gradient Boosting), an open-source library of the gradient boosted trees algorithm.50 XGBoost is a supervised machine learning technique that combines a set of classification and regression trees to predict a target variable. XGBoost uses a regularized objective function to evaluate the classification accuracy and simplification of the ensemble model and decide whether a new tree is to be ensembled to the prior model or not. XGBoost has been widely used for classification tasks and has won many machine learning competitions in Kaggle, due to its advantages in flexible model tuning, accelerated computing, and enhancements of various algorithm.51 The hRBC dataset was expanded as described in Sec. 2, and the augmented dataset was separated into a training dataset and a test dataset in a ratio of 4:1 for the XGBoost classifier.

Table 1

List of features used in the XGBoost classifier.

CategoryFeatures
Phase-shift statistics (n=5)Optical volume, mean, fifth percentile, 95th percentile, and standard deviation of phase shift
Morphological property (n=5)Projected area, perimeter, major axis length, minor axis length, and eccentricity
Textural property based on the GLCM (n=5)Contrast, dissimilarity, homogeneity, energy, and angular second moment49

3.3.

Results of RBCs Detection and Delineation

Table 2 summarizes the classification performance of Mask R-CNN model and the benchmark XGBoost classifier on the test dataset. The Mask R-CNN model achieved high accuracy (97.8%) in the test dataset, which is very close to the accuracy obtained by the benchmark XGBoost classifier (accuracy 99.9%).

Table 2

Comparison of classification performance on individual RBCs using Mask R-CNN, XGBoost with manual delineation and XGBoost with Mask R-CNN segmentation results.

MethodSensitivity (%)Specificity (%)F1-scoreAUCAccuracy (%)
Mask R-CNN97.198.50.9750.98297.8
XGBoost with MD99.91000.9920.99999.9
XGBoost with Mask R-CNNa99.91000.9920.99999.9
Note: AUC, area under the receiver operating characteristic curve; MD, manual delineation.

aXGBoost classifier trained with the segmentation results from Mask R-CNN.

Exemplary results of segmenting RBCs with the Mask R-CNN model are illustrated in Fig. 3, and corresponding ground truths by manual delineation are shown in blue lines in Fig. 3 for comparison. The segmentation performance of the Mask R-CNN model was evaluated using IoU between the model-predicted mask and the ground-truth mask. The average IoU of the test dataset was 0.945. The RBC phase images, segmented by the Mask R-CNN model, were utilized to train a second XGBoost classifier using the procedure described in Sec. 3.2. The classification performance, also listed in Table 2, was the same as that obtained by the benchmark XGBoost classifier trained with manually delineated RBC images. These results show that the Mask R-CNN model accurately segmented RBC quantitative phase images.

Fig. 3

Exemplary results of the delineation and detection by the Mask R-CNN model. Two slides with hRBCs and tRBCs are shown in (a) and (b). Blue lines are the manual delineation contours of RBCs as the ground truth. Green and red lines mark mode-predicted contours of RBCs classified as hRBCs and tRBCs, respectively. The misclassified cases in (a) and (b) are enlarged and shown in (c) and (d). The model-predicted contours match well with manual delineation contours, indicating that the misclassification is not due to failed segmentation.

JBO_25_11_116502_f003.png

4.

Characterization of Thalassemic RBC Quantitative Phase Images

To unveil the underlying characteristics of RBCs that enable the accurate classification of tRBCs and to demonstrate an advantage of Mask R-CNN’s in performing instance segmentation, we took the segmentation results of the Mask R-CNN model to train a second XGBoost classifier for discriminating tRBCs from hRBCs. Then, we performed the SHAP analysis to evaluate the relative impact of each feature on the XGBoost classifier.5255 Each dot in Fig. 4 indicates the SHAP value of an RBC in the test dataset for the feature mentioned on the left. The color of a dot indicates the measured value for the corresponding feature, with higher values in red and lower values in blue. The magnitude of the SHAP value indicates the impact level of each data point to the model. Therefore, the overall impact of each feature can be quantified by the sum of the absolute SHAP values, over all the data points. In Fig. 4, the distributions of relative feature values also indicate the feature representations of each class. For example, tRBCs demonstrated a low optical volume, mean, and fifth percentile of phase shift but a high standard deviation of phase shift. The most influential features, with a mean absolute SHAP value larger than 0.6, are all related to phase shifts, including the mean, the sum (optical volume), the fifth percentile, and the standard deviation of phase shifts, of single RBCs. On an average, the morphological features showed a moderate influence and the textural features showed the least influence, on the XGBoost classifier. The lack of significant contributions from the textural features could be attributed to approximately uniform hemoglobin distributions in both hRBCs and tRBCs. However, in patients with severe anemia or malaria, where RBCs are nucleated or infected, the texture analysis may provide important information for detecting the abnormal RBCs. SHAP analysis was also performed on the benchmark XGBoost classifier. The result is shown in Fig. 4(b) and is very similar to that of the XGBoost classifier trained with Mask R-CNN segmented RBC images shown in Fig. 4(a). These results demonstrated that there were no significant differences in the discriminant rules between the classifier trained with the Mask R-CNN segmentation and the classifier trained with the manual delineation.

Fig. 4

SHAP value distributions of the 15 selected features for XGBoost classifier trained with (a) Mask R-CNN segmentation and (b) manual segmentation. The features are sorted according to the mean absolute SHAP value that represents the relative contribution of each feature to the XGBoost classifier. The color presents the normalized feature value of each RBC in the test dataset.

JBO_25_11_116502_f004.png

5.

Correlations between the Features Extracted from 2-D Quantitative Phase Images and 3-D RI Maps of RBCs

Results shown in Table 2 indicate that 2-D QPI-extracted features are sufficient to accurately discriminate tRBCs from hRBCs. According to the SHAP analysis described in Sec. 4, the most influential features are related to the phase shifts of RBCs. Since the phase shift is the line integral of the RI contrast along the RBC thickness, it is plausible that the 2-D QPI features are correlated to the 3-D morphological information for distinguishing tRBCs from hRBCs. To elucidate correlations between the phase-shift statistics and features extracted from 3-D RI maps of the same RBCs, we applied the canonical correlations analysis (CCA) to investigate associations between the two sets of features.

5.1.

Procedure of the Canonical Correlation Analysis

CCA is a statistical method for exploring correlations between two sets of variables acquired on the same experimental units.56 Assume X and Y denote two sets of continuous variables. CCA aims to seek two vectors a and b, canonical coefficients, to maximize the correlation between linear combinations of X and Y formulated as U=aTX and W=bTY. The pair of vectors U and W is called canonical variates, and the correlation between U and W is quantified to represent the canonical correlation. After the first pair of canonical variates is obtained, a new pair of canonical variates can be found by maximizing their correlation under the condition that they are uncorrelated with previously extracted canonical variates. The process of canonical variate extraction repeats until the number of canonical variates equals to the minimum feature number between X and Y.

The top four 2-D QPI features identified by the SHAP analysis (Sec. 4) included optical volume (OV), mean (ϕmean), fifth percentile (ϕP5), and the standard deviation (ϕSD) of single-RBC phase shifts and were chosen as predictor variables (X). For selecting features of the 3-D RI maps, we chose features that demonstrated relatively good performance in discriminating tRBCs from hRBCs (the area under the receiver operating characteristic curve, AUC>0.8), as reported in Table 3 of Ref. 11. These selected 3-D features were assigned as response variables (Y) and included the volume (V), surface-area-to-volume ratio (S/V), sphericity index [SI=π1/3(6V)2/3/S], average thickness (Tavg), central thickness (T0), and the difference between the thickness at three quarters of the normalized radial position and T0 (Δ75). Each of the variables was standardized by removing the mean and scaling to unit variance before CCA was performed. Since the number of predictor variables and response variables are p=4 and q=6, respectively, there are four pairs of canonical variates [min(p,q)=4] denoted as (Ui,Wi) with i=1, 2, 3, and 4.

Since the canonical correlation is calculated between a pair of canonical variates that are linear combinations of the original variables, it is desirable to analyze associations between the canonical variates and the original variables. We followed the method described by Arifler57 that used within-set loadings to assess the associations. For example, a predictor variable is interpreted to be highly correlated with a canonical variate Ui if it has a relatively high canonical loading. Similarly, the canonical loadings of response variables can be used to interpret their correlations with Wi.

5.2.

Associations between Features Extracted from 2-D Quantitative Phase Images and 3-D RI Maps of the Same RBCs

The canonical correlations (ρ) and canonical loadings (L) of CCA are listed in Table 3. The top two canonical variate pairs showed very high correlations (ρ=0.94 and 0.87, respectively). The third and fourth pairs of canonical variates showed weak correlations (ρ=0.27 and 0.15, respectively) and will not be discussed further. The first pair of canonical variates, U1 and W1, is dominated by OV and V, respectively. Since phase shift is the product of RI and RBC thickness, and OV is the sum of the phase shift over the whole RBC image, OV is expected to be proportional to RBC volume if the average RI does not vary significantly between RBCs from the same subject.58 Both OV and V concern the volume of an RBC, not the shape of it. In contrast, an inspection of canonical loadings of the second canonical variate pair, U2 and W2, indicates associations between the other phase-shift statistics and various shape-related features. Following the same assumption that there is no significant difference in the average RI between RBCs from the same subject, the phase shift is proportional to the thickness of an RBC. Therefore, ϕmean is proportional to Tavg and ϕP5 and is closely related to T0. Figures 3 and 5 of Ref. 11 show that tRBCs tend to be shaped like doughnuts, which is associated with increase in ϕSD, S/V, and Δ75, and with a decrease in SI. The canonical loadings of U2 and W2 also indicate that ϕSD is positively correlated with S/V and Δ75 and negatively correlated with SI. The results of CCA reveal associations between the phase-shift statistics and the 3-D morphological features of RBCs and facilitate the interpretation of 2-D QPI-based features that achieved excellent accuracy in distinguishing tRBCs from hRBCs.

Table 3

CCA between phase-shift statistics extracted from 2-D quantitative phase images and features extracted from 3-D RI maps.

Pairs of canonical variatesFirstSecondThirdFourth
Canonical correlations (ρ)0.940.870.270.15
Shared variances (ρ2)0.890.760.070.02
Canonical loadings
U1U2U3U3
OV0.530.510.140.04
ϕmean0.130.600.170.06
ϕSD0.270.530.190.09
ϕP50.100.690.050.04
W1W2W3W4
V0.670.550.060.00
S/V0.260.770.060.03
SI0.180.790.100.02
Tavg0.060.810.060.03
T00.040.740.040.06
Δ750.000.810.020.05
Note: Bold values show the highest correlation in each canonical loading.

6.

Discussion and Conclusions

We successfully applied Mask R-CNN to build a classification model to accurately (accuracy=97.8%) discriminate between hRCBs and tRBCs, based on quantitative phase images of unstained RBCs. In the development phase of the Mask R-CNN model, a CNN backbone was adopted to automatically extract deep features and to replace the manual feature extraction and selection steps employed in conventional image processing-based automatic diagnostic methods. Therefore, the requirement for hematological knowledge to select optimal features is eliminated. During inference, the Mask R-CNN model can efficiently and automatically detect tRBCs. In addition, the same Mask R-CNN model also provided segmentation of RBCs for single-cell characterization of 2-D QPI features. The results shown in Sec. 4 reveal the most important 2-D QPI features of tRBCs that can separate them from hRBCs. In future clinical applications, these features can be quantified automatically and used by healthcare providers to aid their decision-making processes.

Figure 4 shows that the most influential features for the classification of tRBCs are related to phase shifts. The optical volume (i.e., the sum of phase shifts) of an RBC is proportional to the dry mass of the RBC. It is worth to note that the mean corpuscular hemoglobin obtained using a regular blood test (i.e., complete blood count) is the average mass of hemoglobin per RBC in the whole blood sample. Therefore, QPI of RBCs provides not only the average hemoglobin content per RBC but also the hemoglobin content of every RBC imaged. Results of the CCA analysis show that the other phase-shift statistics are highly associated with various 3-D morphological features of RBCs (see Table 3 in Sec. 5.2). To the best of our knowledge, correlations between phase shifts and the 3-D morphology of tRBCs have not been experimentally investigated and reported. Mugnano et al.10 characterized the phase distribution of individual RBCs by fitting quantitative phase images of RBCs with Zernike polynomials. Three Zernike coefficients, namely, the piston, the defocus, and the third-order spherical aberrations, were identified to best distinguish tRBCs from hRBCs. The piston is related to the size and the mean phase shift of an RBC, and so is the optical volume. The defocus and the third-order spherical aberration are associated with the biconcave shape of RBCs. The 3-D morphological features analyzed in the current study are more intuitive and commonly used, and thus allow for direct comparisons with existing studies in the literature.5964

To realize high-throughput screening of RBCs for clinical applications, both the hardware and the software reported in this paper need to be improved. First, the QPI instrument can be integrated with microfluidic devices to automate the dilution and pumping of blood samples for optimal acquisition speed with sufficient image quality.5,65 Second, rapid processing and inference of quantitative phase images using the Mask R-CNN method is needed for real-time RBC detection and delineation. Currently, the inference time of the Mask R-CNN model on a phase map with one million pixels is under 0.2 s using an NVIDIA GTX 1080 Ti GPU. To improve the speed of the inference, the CNN backbone can be further optimized by combining layers and optimizing the kernel selection, for instance, using inference accelerators such as the TensorRT.66 We believe that the inference frame rate of the proposed Mask R-CNN model can achieve video rate on images with one million pixels, using our current computing hardware. Furthermore, due to its excellent automatic detection and high deplorability, the proposed Mask R-CNN model is well suited for assisting diagnostics via telehealth in remote areas lacking medical resources.

In conclusion, we demonstrated simultaneously the automatic delineation of RBC phase images for single-cell analysis and highly accurate detection of tRBCs using Mask R-CNN. The developed Mask R-CNN model efficiently processed massive amounts of QPI data and greatly simplified the process of developing an automatic method for detecting RBC-related abnormalities from the QPI data. Moreover, the instance segmentation capability of the Mask R-CNN model was very useful for single-cell characterization, and combined with a classifier and the SHAP analysis, provided invaluable insights into the relationship between RBC QPI features and diseases. To help interpret the QPI features, CCA was performed to elucidate associations between the statistics of the phase shift and the more intuitive RBC 3-D morphology. We believe that Mask R-CNN has the potential to improve the efficiency of hematological examinations and achieve single-RBC characterization to aid clinical decision-making.

Disclosures

All authors state that they have no relevant financial interests in this article and no other conflicts of interest to disclose.

Acknowledgments

This study was supported by the Ministry of Science and Technology (Grant No. 108-2221-E-002-081-MY3) of Taiwan for financial support. The authors thank Dr. Shang-Ju Wu at the National Taiwan University Hospital for his help with volunteer recruitment and blood sample collection.

References

1. 

B. J. Bain, “Diagnosis from the blood smear,” N. Engl. J. Med., 353 (5), 498 –507 (2005). https://doi.org/10.1056/NEJMra043442 NEJMAG 0028-4793 Google Scholar

2. 

J. Ford, “Red blood cell morphology,” Int. J. Lab. Hematol., 35 (3), 351 –357 (2013). https://doi.org/10.1111/ijlh.12082 Google Scholar

3. 

A. Adewoyin, “Peripheral blood film-a review,” Ann. Ibadan Postgr. Med., 12 (2), 71 –79 (2014). Google Scholar

4. 

M. Ugele et al., “Label-free, high-throughput detection of P. falciparum infection in sphered erythrocytes with digital holographic microscopy,” Lab Chip, 18 (12), 1704 –1712 (2018). https://doi.org/10.1039/C8LC00350E LCAHAM 1473-0197 Google Scholar

5. 

N. Lue et al., “Live cell refractometry using microfluidic devices,” Opt. Lett., 31 (18), 2759 –2761 (2006). https://doi.org/10.1364/OL.31.002759 OPLEDP 0146-9592 Google Scholar

6. 

Y. Park, C. Depeursinge and G. Popescu, “Quantitative phase imaging in biomedicine,” Nat. Photonics, 12 (10), 578 –589 (2018). https://doi.org/10.1038/s41566-018-0253-x NPAHBY 1749-4885 Google Scholar

7. 

R. Barer, “Interference microscopy and mass determination,” Nature, 169 (4296), 366 –367 (1952). https://doi.org/10.1038/169366b0 Google Scholar

8. 

M. T. Rinehart, H. S. Park and A. Wax, “Influence of defocus on quantitative analysis of microscopic objects and individual cells with digital holography,” Biomed. Opt. Express, 6 (6), 2067 –2075 (2015). https://doi.org/10.1364/BOE.6.002067 BOEICL 2156-7085 Google Scholar

9. 

G. Kim et al., “Learning-based screening of hematologic disorders using quantitative phase imaging of individual red blood cells,” Biosens. Bioelectron., 123 69 –76 (2019). https://doi.org/10.1016/j.bios.2018.09.068 BBIOE4 0956-5663 Google Scholar

10. 

M. Mugnano et al., “Label-free optical marker for red-blood-cell phenotyping of inherited anemias,” Anal. Chem., 90 (12), 7495 –7501 (2018). https://doi.org/10.1021/acs.analchem.8b01076 ANCHAM 0003-2700 Google Scholar

11. 

Y.-H. Lin et al., “Morphometric analysis of erythrocytes from patients with thalassemia using tomographic diffractive microscopy,” J. Biomed. Opt., 22 (11), 116009 (2017). https://doi.org/10.1117/1.JBO.22.11.116009 JBOPFO 1083-3668 Google Scholar

12. 

M. T. Rinehart et al., “Hemoglobin consumption by P. falciparum in individual erythrocytes imaged via quantitative phase spectroscopy,” Sci. Rep., 6 24461 (2016). https://doi.org/10.1038/srep24461 SRCEC3 2045-2322 Google Scholar

13. 

H. S. Park et al., “Automated detection of P. falciparum using machine learning algorithms with quantitative phase images of unstained cells,” PLoS One, 11 (9), e0163045 (2016). https://doi.org/10.1371/journal.pone.0163045 POLNCL 1932-6203 Google Scholar

14. 

A. Anand et al., “Automatic identification of malaria-infected RBC with digital holographic microscopy using correlation algorithms,” IEEE Photonics J., 4 (5), 1456 –1464 (2012). https://doi.org/10.1109/JPHOT.2012.2210199 Google Scholar

15. 

B. Javidi et al., “Sickle cell disease diagnosis based on spatio-temporal cell dynamics analysis using 3D printed shearing digital holographic microscopy,” Opt. Express, 26 (10), 13614 –13627 (2018). https://doi.org/10.1364/OE.26.013614 OPEXFF 1094-4087 Google Scholar

16. 

N. T. Shaked et al., “Quantitative microscopy and nanoscopy of sickle red blood cells performed by wide field digital interferometry,” J. Biomed. Opt., 16 (3), 030506 (2011). https://doi.org/10.1117/1.3556717 JBOPFO 1083-3668 Google Scholar

17. 

J. Jung et al., “Optical characterization of red blood cells from individuals with sickle cell trait and disease in Tanzania using quantitative phase imaging,” Sci. Rep., 6 31698 (2016). https://doi.org/10.1038/srep31698 SRCEC3 2045-2322 Google Scholar

18. 

F. Yi, I. Moon and B. Javidi, “Cell morphology-based classification of red blood cells using holographic imaging informatics,” Biomed. Opt. Express, 7 (6), 2385 –2399 (2016). https://doi.org/10.1364/BOE.7.002385 BOEICL 2156-7085 Google Scholar

19. 

K. Jaferzadeh and I. Moon, “Human red blood cell recognition enhancement with three-dimensional morphological features obtained by digital holographic imaging,” J. Biomed. Opt., 21 (12), 126015 (2016). https://doi.org/10.1117/1.JBO.21.12.126015 JBOPFO 1083-3668 Google Scholar

20. 

F. Yi et al., “Automated segmentation of multiple red blood cells with digital holographic microscopy,” J. Biomed. Opt., 18 (2), 026006 (2013). https://doi.org/10.1117/1.JBO.18.2.026006 JBOPFO 1083-3668 Google Scholar

21. 

Y. Jo et al., “Quantitative phase imaging and artificial intelligence: a review,” IEEE J. Sel. Top. Quantum Electron., 25 (1), 1 –14 (2018). https://doi.org/10.1109/JSTQE.2018.2859234 IJSQEN 1077-260X Google Scholar

22. 

G. Litjens et al., “A survey on deep learning in medical image analysis,” Med. Image Anal., 42 60 –88 (2017). https://doi.org/10.1016/j.media.2017.07.005 Google Scholar

23. 

G. Kim et al., “Rapid and label-free identification of individual bacterial pathogens exploiting three-dimensional quantitative phase imaging and deep learning,” (2019). Google Scholar

24. 

Y. Jo et al., “Holographic deep learning for rapid optical screening of anthrax spores,” Sci. Adv., 3 (8), e1700606 (2017). https://doi.org/10.1126/sciadv.1700606 STAMCV 1468-6996 Google Scholar

25. 

Y. Zhang et al., “Computational cytometer based on magnetically modulated coherent imaging and deep learning,” Light: Sci. Appl., 8 (1), 1 –15 (2019). https://doi.org/10.1038/s41377-019-0203-5 Google Scholar

26. 

F. Yi, I. Moon and B. Javidi, “Automated red blood cells extraction from holographic images using fully convolutional neural networks,” Biomed. Opt. Express, 8 (10), 4466 –4479 (2017). https://doi.org/10.1364/BOE.8.004466 BOEICL 2156-7085 Google Scholar

27. 

K. de Haan et al., “Automated screening of sickle cells using a smartphone-based microscope and deep learning,” (2019). Google Scholar

28. 

M. Xu et al., “A deep convolutional neural network for classification of red blood cells in sickle cell anemia,” PLoS Comput. Biol., 13 (10), (2017). https://doi.org/10.1371/journal.pcbi.1005746 Google Scholar

29. 

L. Alzubaidi et al., “Deep learning models for classification of red blood cells in microscopy images to aid in sickle cell anemia diagnosis,” Electronics, 9 (3), 427 (2020). https://doi.org/10.3390/electronics9030427 ELECAD 0013-5070 Google Scholar

30. 

W. D. Pan, Y. Dong, D. Wu, “Classification of malaria-infected cells using deep convolutional neural networks,” Machine Learning—Advanced Techniques and Emerging Applications, 159 BoD–Books on Demand(2018). Google Scholar

31. 

K. He et al., “Mask R-CNN,” in IEEE Int. Conf. Comput. Vision, 2980 –2988 (2017). Google Scholar

32. 

M. S. Durkee et al., “Improved instance segmentation of immune cells in human lupus nephritis biopsies with Mask R-CNN,” Proc. SPIE, 11320 1132019 (2020). https://doi.org/10.1117/12.2545110 PSISDG 0277-786X Google Scholar

33. 

H. Jung, B. Lodhi and J. Kang, “An automatic nuclei segmentation method based on deep convolutional neural networks for histopathology images,” BMC Biomed. Eng., 1 (1), 24 (2019). https://doi.org/10.1186/s42490-019-0026-8 Google Scholar

34. 

N. Moshkov et al., “Test-time augmentation for deep learning-based cell segmentation on microscopy images,” Sci. Rep., 10 (1), 1 –7 (2020). https://doi.org/10.1038/s41598-020-61808-3 SRCEC3 2045-2322 Google Scholar

35. 

X. Xie et al., “Robust segmentation of nucleus in histopathology images via mask R-CNN,” in Int. MICCAI Brainlesion Workshop, 428 –436 (2018). Google Scholar

36. 

T. Vos et al., “Global, regional, and national incidence, prevalence, and years lived with disability for 310 diseases and injuries, 1990–2015: a systematic analysis for the Global Burden of Disease Study 2015,” Lancet, 388 (10053), 1545 –1602 (2016). https://doi.org/10.1016/S0140-6736(16)31678-6 LANCAO 0140-6736 Google Scholar

37. 

S. M. Lundberg et al., “From local explanations to global understanding with explainable AI for trees,” Nat. Mach. Intell., 2 (1), 56 –67 (2020). https://doi.org/10.1038/s42256-019-0138-9 Google Scholar

38. 

S. M. Lundberg and S.-I. Lee, “A unified approach to interpreting model predictions,” in Adv. Neural Inf. Process. Syst., 4765 –4774 (2017). Google Scholar

39. 

W.-C. Hsu et al., “Tomographic diffractive microscopy of living cells based on a common-path configuration,” Opt. Lett., 39 (7), 2210 –2213 (2014). https://doi.org/10.1364/OL.39.002210 OPLEDP 0146-9592 Google Scholar

40. 

J. W. Su et al., “Digital holographic microtomography for high‐resolution refractive index mapping of live cells,” J. Biophotonics, 6 (5), 416 –424 (2013). https://doi.org/10.1002/jbio.201200022 Google Scholar

41. 

P. Girshovitz and N. T. Shaked, “Real-time quantitative phase reconstruction in off-axis digital holography using multiplexing,” Opt. Lett., 39 (8), 2262 –2265 (2014). https://doi.org/10.1364/OL.39.002262 OPLEDP 0146-9592 Google Scholar

42. 

G. Dardikman et al., “Video-rate processing in tomographic phase microscopy of biological cells using CUDA,” Opt. Express, 24 (11), 11839 –11854 (2016). https://doi.org/10.1364/OE.24.011839 OPEXFF 1094-4087 Google Scholar

43. 

A. Krizhevsky, I. Sutskever and G. E. Hinton, “ImageNet classification with deep convolutional neural networks,” in Adv. Neural Inf. Process. Syst., 1097 –1105 (2012). Google Scholar

44. 

K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” (2014). Google Scholar

45. 

C. Shorten and T. M. Khoshgoftaar, “A survey on image data augmentation for deep learning,” J. Big Data, 6 (1), 60 (2019). https://doi.org/10.1186/s40537-019-0197-0 Google Scholar

46. 

T.-Y. Lin et al., “Feature pyramid networks for object detection,” in Proc. IEEE Conf Comput. Vision and Pattern Recognit., 2117 –2125 (2017). https://doi.org/10.1109/CVPR.2017.106 Google Scholar

47. 

M. A. Rahman and Y. Wang, “Optimizing intersection-over-union in deep neural networks for image segmentation,” in Int. Symp. Vis. Comput., 234 –244 (2016). Google Scholar

48. 

K. Kono et al., “Quantitative distinction of the morphological characteristic of erythrocyte precursor cells with texture analysis using gray level co‐occurrence matrix,” J. Clin. Lab. Anal., 32 (1), e22175 (2018). https://doi.org/10.1002/jcla.22175 JCANEM 1098-2825 Google Scholar

49. 

M. Hall-Beyer, “GLCM texture: a tutorial,” National Council on Geographic Information and Analysis Remote Sensing Core Curriculum, 3 PRISM(2000). Google Scholar

50. 

T. Chen and C. Guestrin, “XGBoost: a scalable tree boosting system,” in Proc. 22nd ACM SIGKDD Int. Conf. Knowl. Discov. and Data Mining, 785 –794 (2016). Google Scholar

51. 

V. Janko et al., “A new frontier for activity recognition: the Sussex-Huawei locomotion challenge,” in Proc. 2018 ACM Int. Joint Conf. and 2018 Int. Symp. Pervasive and Ubiquitous Comput. and Wearable Comput., 1511 –1520 (2018). Google Scholar

52. 

J. Suh et al., “Development and validation of explainable AI‐based decision‐supporting tool for prostate biopsy,” BJU Int., (2020). https://doi.org/10.1111/bju.15122 Google Scholar

53. 

H. Behravan et al., “Predicting breast cancer risk using interacting genetic and demographic factors and machine learning,” Sci. Rep., 10 (1), 1 –16 (2020). https://doi.org/10.1038/s41598-020-66907-9 SRCEC3 2045-2322 Google Scholar

54. 

K. S. Betts, S. Kisely and R. Alati, “Predicting postpartum psychiatric admission using a machine learning approach,” J. Psychiatr. Res., 130 35 –40 (2020). https://doi.org/10.1016/j.jpsychires.2020.07.002 JPYRA3 0022-3956 Google Scholar

55. 

P.-Y. Tseng et al., “Prediction of the development of acute kidney injury following cardiac surgery by machine learning,” Crit. Care, 24 (1), 1 –13 (2020). https://doi.org/10.1186/s13054-020-03179-9 Google Scholar

56. 

H. Hotelling, “Relations between two sets of variates,” Breakthroughs in Statistics, 162 –190 Springer(1992). Google Scholar

57. 

D. Arifler, “Sensitivity of spatially resolved reflectance signals to coincident variations in tissue optical properties,” Appl. Opt., 49 (22), 4310 –4320 (2010). https://doi.org/10.1364/AO.49.004310 APOPAI 0003-6935 Google Scholar

58. 

Y. Park et al., “Refractive index maps and membrane dynamics of human red blood cells parasitized by Plasmodium falciparum,” Proc. Natl. Acad. Sci. U. S. A., 105 (37), 13730 –13735 (2008). https://doi.org/10.1073/pnas.0806100105 Google Scholar

59. 

H. Park et al., “Measuring cell surface area and deformability of individual human red blood cells over blood storage using quantitative phase imaging,” Sci. Rep., 6 34257 (2016). https://doi.org/10.1038/srep34257 SRCEC3 2045-2322 Google Scholar

60. 

S. Y. Lee et al., “The effects of ethanol on the morphological and biochemical properties of individual human red blood cells,” PLoS One, 10 (12), e0145327 (2015). https://doi.org/10.1371/journal.pone.0145327 POLNCL 1932-6203 Google Scholar

61. 

H. Park et al., “Three-dimensional refractive index tomograms and deformability of individual human red blood cells from cord blood of newborn infants and maternal blood,” J. Biomed. Opt., 20 (11), 111208 (2015). https://doi.org/10.1117/1.JBO.20.11.111208 JBOPFO 1083-3668 Google Scholar

62. 

Y. Kim et al., “Profiling individual human red blood cells using common-path diffraction optical tomography,” Sci. Rep., 4 6659 (2014). https://doi.org/10.1038/srep06659 SRCEC3 2045-2322 Google Scholar

63. 

K. Kim et al., “High-resolution three-dimensional imaging of red blood cells parasitized by Plasmodium falciparum and in situ hemozoin crystals using optical diffraction tomography,” J. Biomed. Opt., 19 (1), 011005 (2013). https://doi.org/10.1117/1.JBO.19.1.011005 JBOPFO 1083-3668 Google Scholar

64. 

P. Memmolo et al., “3D morphometry of red blood cells by digital holography,” Cytometry Part A, 85 (12), 1030 –1036 (2014). https://doi.org/10.1002/cyto.a.22570 1552-4922 Google Scholar

65. 

F. Merola et al., “Tomographic flow cytometry by digital holography,” Light: Sci. Appl., 6 (4), e16241 (2017). https://doi.org/10.1038/lsa.2016.241 Google Scholar

Biography

Yang-Hsien Lin received his MS degree in biomedical imaging and radiological science from National Yang-Ming University in 2013, and graduated with a PhD in biomedical electronics and bioinformatics from National Taiwan University in 2020. His research interests include developments and applications of the optimization of digital holographic microscopy and tomographic phase microscopy.

Ken Y.-K. Liao graduated with a bachelor’s degree in information technology from Kwantlen Polytechnic University, Canada, and earned his master’s degree in biomedical engineering and biomedical informatics from Feng Chia University, Taiwan. He is currently a PhD candidate studying in the PhD Program of Electrical and Communications Engineering at Feng-Chia University, Taichung, Taiwan.

Kung-Bin Sung received his MS and PhD degrees in biomedical engineering from the University of Texas at Austin in 1999 and 2003, respectively. He worked as a research scientist at Intel Corporation from 2003 to 2006. He joined the Department of Electrical Engineering and Graduate Institute of Biomedical Electronics and Bioinformatics at National Taiwan University in 2006. His research focuses on developing optical spectroscopy and microscopy techniques for early diagnosis of cancer and precancer.

CC BY: © The Authors. Published by SPIE under a Creative Commons Attribution 4.0 Unported License. Distribution or reproduction of this work in whole or in part requires full attribution of the original publication, including its DOI.
Yang-Hsien Lin, Ken Y.-K. Liao, and Kung-Bin Sung "Automatic detection and characterization of quantitative phase images of thalassemic red blood cells using a mask region-based convolutional neural network," Journal of Biomedical Optics 25(11), 116502 (13 November 2020). https://doi.org/10.1117/1.JBO.25.11.116502
Received: 30 June 2020; Accepted: 26 October 2020; Published: 13 November 2020
Lens.org Logo
CITATIONS
Cited by 17 scholarly publications.
Advertisement
Advertisement
KEYWORDS
Image segmentation

Blood

Phase shifts

Feature extraction

Performance modeling

Data modeling

3D image processing

Back to Top