Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

White blood cell classification via a discriminative region detection assisted feature aggregation network

Open Access Open Access

Abstract

White blood cell (WBC) classification plays an important role in human pathological diagnosis since WBCs will show different appearance when they fight with various disease pathogens. Although many previous white blood cell classification have been proposed and earned great success, their classification accuracy is still significantly affected by some practical issues such as uneven staining, boundary blur and nuclear intra-class variability. In this paper, we propose a deep neural network for WBC classification via discriminative region detection assisted feature aggregation (DRFA-Net), which can accurately locate the WBC area to boost final classification performance. Specifically, DRFA-Net uses an adaptive feature enhancement module to refine multi-level deep features in a bilateral manner for efficiently capturing both high-level semantic information and low-level details of WBC images. Considering the fact that background areas could inevitably produce interference, we design a network branch to detect the WBC area with the supervision of segmented ground truth. The bilaterally refined features obtained from two directions are finally aggregated for final classification, and the detected WBC area is utilized to highlight the features of discriminative regions by an attention mechanism. Extensive experiments on several public datasets are conducted to validate that our proposed DRFA-Net can obtain higher accuracies when compared with other state-of-the-art WBC classification methods.

© 2022 Optica Publishing Group under the terms of the Optica Open Access Publishing Agreement

1. Introduction

As a synonym of leukocytes, white blood cells (WBCs) make up 1% of the total blood volume in a healthy body and play an important role for defending the human body against infectious agents such as virus, bacteria, foreign invaders and infected or cancerous cells [1,2]. Therefore, observing the changes of WBCs in the human blood volume is a valid indicator for disease diagnosis. With the rapid development of microscope imaging techniques [36], there are many WBC images captured from peripheral blood smear being obtained [2,711], which can be directly used for WBC detection and classification. Since all WBCs are over 5 micrometers in diameter, of which the size is large enough to be seen using a typical optical microscope (compound microscope). Stained with Leishman’s stain, different types of leukocytes can be easily identified, and it is also easy to count them. There are various cellular entities in human blood, which are responsible for different functions of body such as carrying oxygen, regeneration, clotting and immunity. Blood mainly consists of four components, including plasma, red blood cells (RBC), platelets, and WBC. Of the four components, WBCs play important role in protecting the body from infection through fighting with foreign pathogens [1,1214]. Therefore, WBC classification is important and beneficial for clinical medical disease detection. WBCs are composed of granulocytes (neutrophils, eosinophils, and basophils) and non-granulocytes (lymphocytes and monocytes), which can be categorized into five types, i.e., Monocyte, Eosinophil, Basophil, Lymphocytes and Neutrophil [7,15] (As shown in Fig. 1). All of these types of WBC have a certain function for human body in the fight against kinds of infections such as virus, bacteria and fungus. In a normal human body, the ratio of RBC and WBC is 600:1, which makes the identification of WBC type in a blood smear sample a difficult task. When some diseases occur, it will cause changes in the number and shape of WBCs. For example, when the human body is invaded by influenza virus, in order to resist the damage of the virus to the human body, the number of white blood cells will increase at this time. In addition, if the human body develops malignant tumors and other diseases, not only the number but also the proportion of different types of WBCs will change obviously. Furthermore, WBCs will also deform significantly. Therefore, when doctors need to check whether a patient suffers from a certain disease, the most intuitive way is to detect whether WBCs in the human body have undergone quantitative changes or lesions [16,17]. In clinical practice, WBCs are generally identified by manual microscopy [1820], which is time-consuming and labour-intensive. In addition, human errors are inevitable during the identifying process due to some human subjective factors. As a result, it is necessary to develop intelligent WBC classification methods based on advanced artificial intelligence techniques. During the past decades, machine learning-based methods are widely used to reduce the workload of human beings for WBC classification and have earned great success. In this paper, we also focus on designing an effective method for automatically classifying different types of WBCs based on machine learning techniques.

 figure: Fig. 1.

Fig. 1. Five different types of white blood Cells [21].

Download Full Size | PDF

As to machine learning-based WBC classification methods, it is a requisite to collect an appropriate dataset by taking quality, variety, and size into account. In the early days, the lack of high quality dataset is the major challenge and hinders the classification accuracies of different methods [22]. With sufficient datasets, diverse machine learning techniques can be utilized for WBC classification. In general, previous machine learning based methods can be roughly classified into two categories, i.e., traditional hand-crafted feature extraction based methods and deep learning driven frameworks. In traditional methods, appropriate hand-crafted features are firstly extracted from cell images, and then, different WBC types are classified by certain one or an ensemble of several classifiers. Therefore, it is very critical to extract discriminative features for final classification, and various features have been proposed, such as hough transform [23], local binary pattern [24] and shape [25]. Different to traditional methods, deep learning based framework automatically extract features from original cell images by means of deep neural networks [2628], which are powerful for feature representation learning. The most commonly used networks for image classification are convolutional neural networks (CNNs). In order to obtain good classification results, deep CNNs with numerous parameters are often required and plenty of data are needed to train such deep neural networks. However, unlike natural images, it is often not easy to obtain enough medical data for deep neural network training. Therefore, pre-trained networks from natural images are usually used for feature extraction or fine-tuning on small-scale medical dataset.

Although great success has been achieved in previous WBC classification methods, some practical issues such as uneven staining, boundary blur and nuclear intra-class variability are still hindering the final classification accuracy. In this work, we propose a deep neural network for WBC classification via discriminative region detection assisted feature aggregation, referred as DRFA-Net, which can accurately locate the WBC area from cell images to boost final classification performance. In detail, DRFA-Net uses an adaptive feature enhancement module to refine multi-level deep features in a bilateral manner for efficiently capturing both high-level semantic information and low-level details of cell images. In order to suppress the interference caused by background areas, we design a network branch to detect the WBC area supervised by segmented ground truth. The bilaterally refined features are finally aggregated for final classification, and the detected WBC area is utilized to highlight the features of discriminative regions by an attention mechanism. In a nutshell, we summarize the main contributions of this paper as follows:

  • • We propose a deep neural network for WBC classification via discriminative region detection assisted feature aggregation (DRFA-Net), which can accurately locate the WBC area to boost final classification performance;
  • • In order to capture both high-level semantic information and low-level details of WBCs, we design an adaptive feature enhancement module to refine multi-level deep features in a bilateral manner and aggregate the enhanced features for classification;
  • • Extensive experiments on several public datasets are conducted to demonstrate that our proposed DRFA-Net outperforms other well-known techniques.
The rest of this paper is structured as follows. Section 2. reviews some related research works for WBC classification from blood smear images. In Section 3, we introduce the proposed network step by step and give the details about the network training. Experimental results and analysis are given in Section 4. Finally, concluding remarks are drawn in Section 5.

2. Related Work

In the past few years, many intelligent approaches have been proposed for WBC classification. In order to improve the classification accuracy, digital image technology has become one of the main directions to combine WBC detection and classification into a unified task. By taking a blood smear image of cells under a microscope, cell images can be easily obtained. Then, various techniques in digital image processing community can be used to classify the WBCs. By observing the classified WBC images, professional doctors can easily find the lesions of WBCs. In such a manner, the error caused by personnel operation can be effectively reduced, and the efficiency of disease diagnosis can be well improved. Generally, previous computer aided intelligent methods can be roughly classified into two categories, i.e., traditional hand-crafted feature extraction based methods and deep learning based frameworks.

As to hand-crafted feature extraction based WBC classification methods, how to extract effective features is critical. In previous researches, many kinds of traditional methods are put-forward. Since the white blood cell density is very high in the bone area, Nipon et al. [29] proposed morphological granulometric features of nucleus for bone marrow WBC classification. Mishra et al. [30] proposed a gray level co-occurrence matrix based feature extraction method and a probabilistic PCA based feature section model, which are followed by random forest based classification techniques for acute lymphoblastic leukemia detection and classification. By extracting hybrid features such as texture, shape, color, and performing data normalization, Ahmed et al. [31] successfully used traditional classifiers including Naive Bayes, Decision Tree, support vector machine (SVM) and $k$-nearest neighbor (KNN) to classify different acute lymphoblastic leukemia from WBC images. In [32], Das et al. first extracted gray-level-run-length matrix and gray-level co-occurrence matrix-based texture features together with color and shape, then PCA and SVM are utilized for efficient feature selection and cell classification, respectively. Mishra et al. [33] performed linear discriminant analysis (LDA) on the features extracted by a newly introduced discrete orthogonal S-transform to learn more discriminative low-dimensional features and then employed the Adaboost algorithm as well as random forest for leukemia classification. Considering that the shape changes of WBCs are very complicated, Lamberti [25] proposed a kind of interpretable metrics based shape feature for WBCs classification. Although above mentioned methods can work efficiently on small-scale dataset without needing of time-consuming model training, they need to accurately segment the area of WBCs from blood smear images, of which itself is also a challenging task.

Due to the powerful ability for learning to extract hierarchical features [27,34,35], deep convolution neural networks (CNNs) have been widely used for image processing and refreshed the records of many visual tasks [27,3642], including WBC classification [17,4346]. Since training a deep neural network from scratch needs large-scale datasets, while it is not easy to capture medical datasets with plentiful enough training samples. To this end, pre-trained models are often used. Given a certain pre-trained model, one way is to use it for extracting features as the input of traditional classifiers such as SVM, KNN, etc. Another way is to transfer the pre-trained model to small dataset by fine-tuning. For feature extraction with pre-trained models, many previous works have also been proposed for WBC classification [4750]. In [49], three different CNN architectures including AlexNet, GoogleNet and ResNet-50 are used for feature extraction, then the features are merged and discriminative features are selected as the input of quadratic discriminant analysis model for final WBC classification. With features extracted from pre-trained CNN model, Sahlol et al. [50] employed a statistically enhanced salp swarm algorithm and SVM for feature selection and classification, respectively. In addition, there are also many researches that focus on directly training deep neural network to classify WBCs [5156]. Without using pre-trained model, even better classification results can be reached [52]. In [57], Jung et al. introduced a new CNN architecture (referred as W-Net) to classify WBCs. Other advanced network architectures are also used for this task, such as capsule networks [55] and long short term memory network [51]. Compared with traditional hand-crafted feature based methods, deep learning based methods often obtain better results due to the strong representation capability of deep features. However, their performance is still degenerated by some issues such as noisy background, uneven staining, boundary blur and nuclear intra-class variability.

3. Proposed DRFA-Net

In this section, we give the details of our proposed DRFA-Net, which takes a WBC image as input and outputs its corresponding type. Overall, DRFA-Net consists of three major parts: discriminative region detection module (DRDM), attention-wise feature aggregation module (AFAM) and classification module (CM). In detail, given an input WBC image, we first use previous pre-trained deep neural network model to extract hierarchical deep features for locating the WBC area. Then the detected WBC area is embedded to highlight discriminative features from deep feature maps with an attention mechanism, which can effectively suppress noisy information for boosting final classification performance. Figure 2 gives a simple architecture of the proposed DRFA-Net.

 figure: Fig. 2.

Fig. 2. The pipeline of our DRFA-Net.

Download Full Size | PDF

It should be noted that there are many network structures can be used for our task, such as Densenet, VGG16, etc. However, the performance of VGG16 usually cannot match Resnet structure, while training a network with Densenet structure needs more computational cost than Resnet structure. As to Resnet structure, it can be easily trained as well as obtaining good performance. Therefore, in order to take a compromise between computational cost and efficacy, we choose the ResNet structure [58,59] as the backbone for feature extraction and deploy the pre-trained ResNeXt model [60] for network initialization. As a result, there are five basic feature extraction layers: conv1, conv2_x, conv3_x, conv4_x, and conv5_x. Firstly, we use the backbone network to extract a series of hierarchical deep feature maps which encode the low-level details and high-level semantic information at different scales of an blood smear image. On the one hand, the fine details of original input image are inevitably gradually lost due to a series of spatial pooling and convolution operations, which is harmful for distinguishing the WBC boundary from background. On the other hand, the features extracted by deep layers are with sufficient high-level semantic information, which can help accurately locate WBC areas. Therefore, it is necessary to exploit the complementary information extracted from both shallow and deep layers for the detection of WBC regions.

3.1 DRDM

During the blood smear imaging process, many interference factors including dyeing impurities and illumination change could induce various challenges such as cytoplasm with low image contrast, variable appearances under different staining conditions, which makes it difficult to separate WBCs from blood smear images. Therefore, in our proposed DRFA-Net, we first design a discriminative region detection module to highlight the WBC area for assisting following classification. For WBC detection, we should consider two aspects. On one hand, high-level semantic information encoded in deep layers should be used to locate the WBC area. One the other hand, low-level details encoded in shallow layers should be used to distinguish the boundary between WBC area and background. In order to capture both deep and shallow features, we design a discriminative region detection module to detect WBC by adaptively refining multi-level deep features in a bilateral manner. Since there is little semantic information in deep features extracted from shallow layers, we design a branch for feature enhancement from deep to shallow layers by embedding multiple deep to shallow modules (DTSMs). Similarly, in order to remedy the fine details missed in the features extracted from deep layers, we design another branch for feature enhancement from shallow to deep layers by embedding a series of shallow to deep modules (STDMs).

With the pre-trained ResNeXt model, we denote the extracted feature maps as $\textbf {F}_1, \textbf {F}_2, \textbf {F}_3, \textbf {F}_4, \textbf {F}_5$ from shallow layers to the deep layers. In order to ensure that the final WBC detection map is the same as original input, we first upsample the five layers of feature maps to the size of input image. For the shallow to deep pathway, we embed a series of STDMs to enhance the details in deep layers by refining features from the shallow layers to deep ones in a hierarchical manner. For each STDM (as shown in the top part of Fig. 3), the input consists of two parts, including the output of last STDM and the features extracted from the corresponding layer. For the sake of simple presentation, we denote $\textbf {F}_L$ and $\textbf {F}_H$ as the output of last STDM and the corresponding input deep features, respectively. Then the output of STDM can be formulated as:

$${\textbf{F}_o^{STDM}} = \Phi (Cat({\textbf{F}_H'},\Phi ({\textbf{F}_L}))),$$
where $\Phi$ represents a mapping function that consists of a series of convolution and ReLU operations, and $\textbf {F}_H'$ denotes the refined feature enhanced by $\textbf {F}_L$. As to $\textbf {F}_L$, we first let it pass a $1 \times 1$ convolution operation, and then average all channels to get a weight map $W$ which will be used to refine $\textbf {F}_H$ by a multiplication. Therefore, $\textbf {F}_H'$ can be mathematically written as:
$$\textbf{F}_H' = \Omega (\Phi ({\textbf{F}_L})) \otimes \Phi ({\textbf{F}_H}) + \Phi ({\textbf{F}_H}),$$
where $\Omega$ represents the mapping function to get the weight map.

 figure: Fig. 3.

Fig. 3. The brief structure of our designed STDM (top) and DTSM (bottom).

Download Full Size | PDF

Similarly, For the deep to shallow pathway, we embed a series of DTSMs to enhance the details in deep layers by refining features from the shallow layers to deep ones. For each DTSMs (as shown in the bottom part of Fig. 3), the input also consists of two parts, including the output of last DTSM and the features extracted from the corresponding layer. We denote $\textbf {F}_H$ and $\textbf {F}_L$ as the output of last DTSM and the corresponding input deep features, respectively. Then the output of DTSM can be formulated as:

$${\textbf{F}_o^{DTSM}} = \Phi (Cat({\textbf{F}_L'},\Phi ({\textbf{F}_H}))),$$
where $\textbf {F}_L'$ denotes the refined feature enhanced by $\textbf {F}_H$. For $\textbf {F}_H$, we also let it pass a $1 \times 1$ convolution operation, and then average all channels to get the corresponding weight map which will be used to refine $\textbf {F}_L$ by a multiplication. As a result, $\textbf {F}_L'$ can be mathematically written as:
$$\textbf{F}_L' = \Omega (\Phi ({\textbf{F}_H})) \otimes \Phi ({\textbf{F}_L}) + \Phi ({\textbf{F}_L}),$$
where $\Omega$ represents the mapping function to get the weight map.

After we get the outputs from the two pathways (denoted as $\textbf {F}^{STDM}$ and $\textbf {F}^{DTSM}$, respectively), we concat them to learn a WBC score map $\textbf {S}$ which highlights the area of WBC from background as well as suppress background as much as possible. The learning process can be mathematically defined as follows:

$$\textbf{S} = ReLU(\textbf{W}_l * Cat(\textbf{F}^{STDM},\textbf{F}^{DTSM}) + \textbf{b}_l),$$
where $*$ represents convolution operation; $\textbf {W}_l$ and $\textbf {b}_l$ are the weights and bias of the convolution that need to be learned during training and ReLU is the ReLU activation function.

3.2 AFAM

With the WBC score map and a series of hierarchical deep feature maps, we design a attention-wise feature aggregation module to learn discriminative features for final classification. During the feature learning process, we should consider two points. Firstly, the global context information of original blood smear image should be aggregated to capture global features. Secondly, features of the most important WBC area should be emphasised. To this end, original deep features from different layers are weighted by the WBC sore map obtained from previous DRDM, then the weighted feature maps are concatenated and aggregated by our designed AFAM. Specifically, the learned discriminative features can be obtained as follows:

$$\textbf{F}_D = ReLU(\textbf{W}_2 * Cat(\textbf{F}_1 \otimes \textbf{S},\textbf{F}_2 \otimes \textbf{S},\textbf{F}_3 \otimes \textbf{S},\textbf{F}_4 \otimes \textbf{S},\textbf{F}_5 \otimes \textbf{S}) + \textbf{b}_2),$$
where $\textbf {W}_2$ and $\textbf {b}_2$ are the weights and bias of the convolution that need to be learned.

Finally, the learned discriminative features $\textbf {F}_D$ are used to classify the types of original WBC images by the classification module.

3.3 Model training and testing

In this work, we use the ResNet architecture [58] as the backbone for feature extraction and implement the proposed DRFA-Net by using the PyTorch framework. The well-trained ResNeXt network on ImageNet [60] consists of 101 basic convolution layers, we use it to initialize the parameters of the feature extraction network. As a result, we have five feature extraction layers as mentioned in previous sections [58,59]. Training: Different to many previous works that only use the classification loss to train networks, we add a WBC detection loss in the proposed DRFA-Net since the manually annotated WBC regions can be used to train the network for WBC detection. Therefore, the total loss of DRFA-Net consists of two parts, i.e., classification loss $L_C$ and detection loss $L_D$.

For the output of WBC detection block, the cross-entropy loss is used for training the network to detect WBC regions. Specifically, the pixel-wise cross-entropy loss between $\textbf {S}$ and the ground-truth WBC mask $\textbf {G}$ is calculated as:

$$L_D(\boldsymbol{\theta} ) ={-} \sum_{x = 1}^W {\sum_{y = 1}^H {\sum_{l \in \{ 0,1\} } {\left\{ {_{{\cdot} \textbf{1}(\textbf{G}(x,y) = l)}^{\log \Pr (\textbf{S}(x, y) = l|\boldsymbol{\theta} )}} \right\}} } } ,$$
where $\textbf {1}(\cdot )$ denotes the indicator function. The notation $l\in \{0, 1\}$ represents the label of the pixel at location $(x,y)$ that dose not belong to WBC area or belongs to WBC area, respectively. $Pr (\textbf {O}(x,y) = l|\boldsymbol{\theta } )$ represents the probability of being predicted as a pixel that belongs to WBC area. $\boldsymbol{\theta }$ contains the network parameters which need to be learned during training.

As to the classification loss, since the task is a multi-class classification problem, we use the multi-class cross-entropy to train the network for learning discriminative features, which can be formulated as follows:

$${L_C}(\boldsymbol{\theta}) ={-} {1 \over N}\sum_i {\sum_{c = 1}^M {1({y_i} = c)\log \Pr ({y_i} = c|\boldsymbol{\theta} )} } ,$$
where $N$ represents the total number of training samples, $M$ is the number of classes and $y_i$ represents the output class of the $i$-th sample.

Finally, the total loss of our DRFA-Net can be written as follows:

$$L = {L_D}(\boldsymbol{\theta}) + \lambda {L_C}(\boldsymbol{\theta}),$$
where $\lambda$ is a parameter to balance the two losses. In our experiments, during the training process in 5000 iteration times, the numerical range of $L_D$ and $L_C$ in the final loss function are $[1.23, 20.91]$ and $[0.85, 5.27]$, respectively, which are in the same scale space. The total loss function values with varying iterative times are shown in Fig. 5 in the experimental section. Therefore, we empirically set the balance parameter $\lambda$ to 1 for experiments.

Our DRFA-Net is initialized by the well-trained ResNeXt network on ImageNet [60]. In order to learn more discriminative features for WBC classification, we fine-tune the pre-trained model on blood smear image dataset. Specificity, we train DRFA-Net on BCISC dataset [2], which consists of five types of WBCs, i.e., Monocyte, Eosinophil, Basophil, Lymphocytes and Neutrophil. For each type, there are 50+ samples with manually annotated WBC mask labelled by pathologists. The statistic details of this training dataset are summarized in Table 1. The total number of samples is 268. In addition, there are also 268 sub-images that cut from original images and these sub-images focus more on the local information of WBC. Since the number of training samples in BCISC is not enough for training a deep neural network, we perform data augmentation by randomly rotating, resizing and horizontally flipping all of the images and their corresponding ground masks, and finally, the training set is enlarged to 8,576 images. We fine-tune the newly designed network by using this enlarged dataset. The proposed DRFA-Net is trained on a machine equipped with an Intel 3.4 GHz CPU with 32 GB memory and a Nvidia Titan V GPU. The stochastic gradient descent (SGD) algorithm with a momentum of 0.9 and a weight decay of 0.0005 is used to optimize the whole network. The learning rate is adjusted by the “poly" policy with a power of 0.9. The training batch size is set to 4, and we stop the whole learning process after 5,000 iterations. The training process is completed after approximately 0.5 hours. Testing: For each input image in the testing phase, we feed it into the proposed network and obtain its corresponding predicted types. Only approximately 0.037s is needed for type prediction for an input image of size $320 \times 320$ pixels by using a single Nvidia Titan Xp GPU, which is very efficient.

Tables Icon

Table 1. Statistic details of the training dataset BCISC

4. Experiments

4.1 Datasets

In our experiments, three different datasets are used for classification test, including Raabin-WBC [10], LISC [7], and BCCD [8].

Raabin-WBC is a large-scale dataset which was released in 2021. There are three sets of cropped sub-images for classifcation, i.e., Train, Test-A, and Test-B. As to all WBCs in Train and Test-A sets, two experts have separately labeled the types of each image. Yet, images in Test-B have not been labeled thoroughly. As a result, we only used Train and Test-A sets in this study. The two sets were collected from 56 normal peripheral blood smears (for lymphocyte, monocyte, neutrophil, and eosinophil) and one chronic myeloid leukemia (CML) case (for basophil) and there are totally 14,514 WBC images. Giemsa technique was used to capture all the films, and the normal peripheral blood smears have been taken using the camera phone of Samsung Galaxy S5 and the microscope of Olympus CX18. As to the CML slide, it has been imaged utilizing an LG G3 camera phone along with a microscope of Zeiss brand. In addition, all of the images were taken with a magnifcation of 100.

LISC contains 257 WBCs from peripheral blood, which have been labeled by only one expert. The images in this dataset were acquired from peripheral blood smear and stained through Gismo-right technique, and they were taken at a magnification of 100 by using a light microscope (Microscope-Axioskope 40) and a digital camera (Sony Model No. SSCDC50AP).

BCCD was taken from the peripheral blood and there are totally 349 WBCs labeled by one expert. The Gismo-right technique has been employed for staining the blood smears. The images in this dataset were also captured at a magnification of 100 by using a regular light microscope together with a CCD color camera. In addition, similar to [61], we correct the label of one image which has been incorrectly labeled.

Since the WBC detection module of the proposed DRFA-Net needs the segmented WBC mask for training, while there is no manually annotated WBC mask in Raabin-WBC and BCCD datasets, we fix the WBC detection module and just fine-tune the classification parameters for the three testing datasets. Then, the fine-tuned model will be used for final classification on the test subsets of different datasets.

4.2 Evaluation metrics

In order to evaluate the classification performance of our proposed network, we use four metrics including Precision, Sensitivity, F1-score (F1), and Accuracy (Acc) in our experiments. With following definition:

  • • True positive (TP): Number of samples that are positive while classification results are positive;
  • • False positive (FP): Number of samples that are negative while classification results are positive;
  • • False negative (FN): Number of samples that are positive while classification results are negative;
  • • True negative (TN): Number of samples that are negative while classification results are negative.
Then the four metrics are calculated as follows:
$$Precision = {{TP} \over {TP + FP}}$$
$$Sensitivity = {{TP} \over {TP + FN}}$$
$$F1 = {{2 \times Precision \times Sensitivity }\over {Precision + Sensitivity} }$$
$$Accuracy = {{TP + TN} \over {TP + FP + TN + FN}}$$

4.3 Experimental results

4.3.1 Comparison with other state-of-the-art methods

In order to verify the superiority of our proposed network, we compare it with some previous WBC classification methods, which include both deep learning based methods and traditional hand-crafted feature based methods. The details of the compared methods are as follows:

  • (1) CNN-RNN [51], which combines the CNN and RNN to deepen the understanding of WBC image content and learn the structured features of images;
  • (2) FusedCNN [62], which combines shallow and deep layer features for accurate and highly efficient classifier of biomedical images;
  • (3) NucSegNet [56], which is a new CNN model that combines the concept of fusing the features of first and last convolutional layers, and propagating the input image to the convolutional layer for Nucleus segmentation;
  • (4) ShapeColorClassifier [61], which is a three-step method for the classification of white blood cells using image processing techniques and machine learning methods. It consists of three steps: detecting the nucleus and cytoplasm, extracting features, and classification. Both color and shape features are extracted for WBC segmentation.
  • (5) WBCaps [55], which is a capsule network based method for white blood cells.
  • (6) CNNDC [17], which is an efficient deep convolutional neural network based method for Acute Lymphoblastic Leukemia detection and classification.
In Table 2, we report the classification performance comparison of our proposed method with other competitors in terms of four metrics on three different datasets. As can be seen, our proposed DRFA-Net consistently outperforms other methods on all of the datasets. Specifically, the average classification accuracies of different types of WBCs on three datasets are 95.17%, 93.21% and 95.87%, respectively, which is of great clinical significance.

Tables Icon

Table 2. Classification performance comparison of our proposed method with other competitors in terms of four metrics

4.3.2 Ablation analysis

As shown the Fig. 2, there are three major elements designed in our proposed DRFA-Net, i.e., WBC detection module, DTSM and STDM. In order to validate the efficacy of different modules, we remove each module individually from the whole network and obtain final classification performance. Firstly, we remove the WBC detection module and append the enhanced features from the two feature enhancing pathways, i.e., $\textbf {F}^{STDM}$ and $\textbf {F}^{DTSM}$, to the concatenated up-sampled deep features extracted from original five feature extraction layers. Therefore, we have another version of DRFA-Net without the attention mechanism, referred as DRFA-Net_noAtt. Secondly, we remove DTSM and STDM from DRFA-Net, respectively and obtain another two versions of DRFA-Net, i.e., DRFA-Net_noDTSM and DRFA-Net_noSTDM. In Table 3, we report the classification results on different datasets obtained by different versions of DRFA-Net. As can be seen, our designed different modules are important and can effectively improve the final classification accuracy. In addition, in order to give an intuitive validation of the WBC detection module, we show some WBC detection results in Fig. 4. As can be seen, the proposed WBC detection module can accurately locate the WBC areas, which is very important for final classification. Since this work does not focus on accurate WBC segmentation, we just use the WBC detection module to locate the discriminative region for final classification. Therefore, we do not use any other post-processing technique to refine the detection results.

 figure: Fig. 4.

Fig. 4. An intuitive showing of the WBC region detection results of the proposed method. The top row represents original WBC images, the bottom row denotes the WBC area detection results of the proposed network.

Download Full Size | PDF

 figure: Fig. 5.

Fig. 5. The training loss and testing accuracies of different datasets with varying iteration times.

Download Full Size | PDF

Tables Icon

Table 3. WBC classification results of the proposed methods on different datasets

4.3.3 Convergence property analysis

In order to intuitively show the convergence property of the proposed network, we plot the training loss and ACC values ($\times 100$) obtained with varying iteration times in Fig. 5. As can be seen, the optimization process is with fast convergence property and the testing accuracies of different datasets increase fast and keep stable when the whole training process achieves convergence.

5. Conclusion

In this paper, we introduce a deep neural network for white blood cell classification via discriminative region detection assisted feature aggregation. The proposed network can accurately locate the WBC area for boosting final classification performance. An adaptive feature enhancement module is designed to refine multi-level deep features in a bilateral manner for efficiently capturing both high-level semantic information and low-level details of WBC images. Since background areas could inevitably produce interference in the WBC images, we design a network branch to detect the WBC area with the supervision of segmented ground truth, then the detected WBC area is utilized to highlight the features of discriminative regions by an attention mechanism. Experiments on three public datasets are conducted to validate the efficacy of the proposed method in terms of different classification metrics.

Funding

National Natural Science Foundation of China (62076228).

Disclosures

The authors declare that there are no conflicts of interest related to this article.

Data availability

Data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.

References

1. T. Terwilliger and M. Abdul-Hay, “Acute lymphoblastic leukemia: a comprehensive review and 2017 update,” Blood Cancer J. 7(6), e577 (2017). [CrossRef]  

2. H. Fan, F. Zhang, L. Xi, Z. Li, G. Liu, and Y. Xu, “Leukocytemask: An automated localization and segmentation method for leukocyte in blood smear images using deep neural networks,” J. Biophotonics 12(7), e201800488 (2019). [CrossRef]  

3. A. Anand, I. Moon, and B. Javidi, “Automated disease identification with 3-d optical imaging: a medical diagnostic tool,” Proc. IEEE 12(5), 924–946 (2019). [CrossRef]  

4. B. Javidi, A. Carnicer, J. Arai, T. Fujii, H. Hua, H. Liao, M. Martínez-Corral, F. Pla, A. Stern, L. Waller, Q.-H. Wang, G. Wetzstein, M. Yamaguchi, and H. Yamamoto, “Roadmap on 3d integral imaging: sensing, processing, and display,” Opt. Express 28(22), 32266–32293 (2020). [CrossRef]  

5. T. O’Connor, A. Anand, B. Andemariam, and B. Javidi, “Deep learning-based cell identification and disease diagnosis using spatio-temporal cellular dynamics in compact digital holographic microscopy,” Biomed. Opt. Express 11(8), 4491–4508 (2020). [CrossRef]  

6. T. O’Connor, J.-B. Shen, B. T. Liang, and B. Javidi, “Digital holographic deep learning of red blood cells for field-portable, rapid covid-19 screening,” Opt. Lett. 46(10), 2344–2347 (2021). [CrossRef]  

7. S. H. Rezatofighi and H. Soltanian-Zadeh, “Automatic recognition of five types of white blood cells in peripheral blood,” Comput. Med. Imaging and Graph. 35(4), 333–343 (2011). [CrossRef]  

8. M. Mohamed, B. Far, and A. Guaily, “An efficient technique for white blood cells nuclei automatic segmentation,” in 2012 IEEE International Conference on Systems, Man, and Cybernetics (SMC), (IEEE, 2012), pp. 220–225.

9. P. Ledwig, M. Sghayyer, J. Kurtzberg, and F. E. Robles, “Dual-wavelength oblique back-illumination microscopy for the non-invasive imaging and quantification of blood in collection and storage bags,” Biomed. Opt. Express 9(6), 2743–2754 (2018). [CrossRef]  

10. Z. M. Kouzehkanan, S. Saghari, E. Tavakoli, P. Rostami, M. Abaszadeh, F. Mirzadeh, E. S. Satlsar, M. Gheidishahran, F. Gorgi, S. Mohammadi, and R. Hosseini, “Raabin-wbc: a large free access dataset of white blood cells from normal peripheral blood,” bioRxiv (2021).

11. G. N. McKay, R. C. Niemeier, C. Castro-González, and N. J. Durr, “Scattering oblique plane microscopy for in-vivo blood cell imaging,” Biomed. Opt. Express 12(5), 2575–2585 (2021). [CrossRef]  

12. J. L. Burnett, J. L. Carns, and R. Richards-Kortum, “Towards a needle-free diagnosis of malaria: in vivo identification and classification of red and white blood cells containing haemozoin,” Malaria J. 16(1), 447 (2017). [CrossRef]  

13. S. Camon, C. Quiros, N. Saubi, A. Moreno, M. A. Marcos, Y. Eto, S. Rofael, E. Monclus, J. Brown, T. D. McHugh, J. Mallolas, and R. Perello, “Full blood count values as a predictor of poor outcome of pneumonia among hiv-infected patients,” BMC Infect. Dis. 18(1), 189 (2018). [CrossRef]  

14. Y. Li, W. Wu, T. Yang, W. Zhou, Y. Fu, Q. Feng, and J. Ye, “Characteristics of peripheral blood leukocyte differential counts in patients with covid-19,” Zhonghua nei ke za zhi 59(5), 372–374 (2020). [CrossRef]  

15. A. Bøyum, “Separation of white blood cells,” Nature 204(4960), 793–794 (1964). [CrossRef]  

16. P. Chitra, M. Jebarani, P. Kavipriya, K. Srilatha, M. Sumathi, and S. Lakshmi, “Detection of aml in blood microscopic images using local binary pattern and supervised classifier,” Res. J. Pharm. Technol. 12(4), 1717–1720 (2019). [CrossRef]  

17. P. K. Das and S. Meher, “An efficient deep convolutional neural network based detection and classification of acute lymphoblastic leukemia,” Expert Syst. Appl. 183, 115311 (2021). [CrossRef]  

18. A. Forcucci, M. E. Pawlowski, C. Majors, R. Richards-Kortum, and T. S. Tkaczyk, “All-plastic, miniature, digital fluorescence microscope for three part white blood cell differential measurements at the point of care,” Biomed. Opt. Express 6(11), 4433–4446 (2015). [CrossRef]  

19. K. Dong, Y. Feng, K. M. Jacobs, J. Q. Lu, R. S. Brock, L. V. Yang, F. E. Bertrand, M. A. Farwell, and X.-H. Hu, “Label-free classification of cultured cells through diffraction imaging,” Biomed. Opt. Express 2(6), 1717–1726 (2011). [CrossRef]  

20. R. Qian, R. P. McNabb, K. C. Zhou, H. M. Mousa, D. R. Saban, V. L. Perez, A. N. Kuo, and J. A. Izatt, “In vivo quantitative analysis of anterior chamber white blood cell mixture composition using spectroscopic optical coherence tomography,” Biomed. Opt. Express 12(4), 2134–2148 (2021). [CrossRef]  

21. H. Moawad, “Leukopenia vs. neutropenia: What are the differences?” https://www.verywellhealth.com/leukopenia-vs-neutopenia-5218935. Accessed April 10, 2022.

22. A.-D. Khamael, J. Banks, K. Nugyen, A. Al-Sabaawi, I. Tomeo-Reyes, and V. Chandran, “Segmentation of white blood cell, nucleus and cytoplasm in digital haematology microscope images: A review–challenges, current and future potential techniques,” IEEE Rev. Biomed. Eng. 14, 290–306 (2021). [CrossRef]  

23. V. Acharya and P. Kumar, “Identification and red blood cell automated counting from blood smear images using computer-aided system,” Med. Biolog. Eng. Comput. 56(3), 483–489 (2018). [CrossRef]  

24. V. Singhal and P. Singh, “Local binary pattern for automatic detection of acute lymphoblastic leukemia,” in 2014 Twentieth National Conference on Communications (NCC), (IEEE, 2014), pp. 1–5.

25. W. F. Lamberti, “Blood cell classification using interpretable shape features: A comparative study of svm models and cnn-based approaches,” Comput. Methods Prog. Biomed. Update 1, 100023 (2021). [CrossRef]  

26. C. Szegedy, A. Toshev, and D. Erhan, “Deep neural networks for object detection,” Adv. Neural Inform. Process. Syst. 26 (2013).

27. K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” in International Conference on Learning Representations, (2015).

28. C. Tang, L. Xinwang, X. Zheng, W. Li, J. Xiong, L. Wang, A. Zomaya, and A. Longo, “Defusionnet: Defocus blur detection via recurrently fusing and refining discriminative multi-scale deep features,” IEEE Trans. Pattern Anal. Mach. Intell. 44(2), 955–968 (2022). [CrossRef]  

29. N. Theera-Umpon and S. Dhompongsa, “Morphological granulometric features of nucleus in automatic bone marrow white blood cell classification,” IEEE Trans. Inf. Technol. Biomed. 11(3), 353–359 (2007). [CrossRef]  

30. S. Mishra, B. Majhi, P. K. Sa, and L. Sharma, “Gray level co-occurrence matrix and random forest based acute lymphoblastic leukemia detection,” Biomed. Signal Process. Control 33, 272–280 (2017). [CrossRef]  

31. A. M. Abdeldaim, A. T. Sahlol, M. Elhoseny, and A. E. Hassanien, “Computer-aided acute lymphoblastic leukemia diagnosis system based on image analysis,” in Advances in Soft Computing and Machine Learning in Image Processing, (Springer, 2018), pp. 131–147.

32. P. K. Das, P. Jadoun, and S. Meher, “Detection and classification of acute lymphocytic leukemia,” in 2020 IEEE-HYDCON, (IEEE, 2020), pp. 1–5.

33. S. Mishra, B. Majhi, and P. K. Sa, “Texture feature based classification on microscopic blood smear for acute lymphoblastic leukemia detection,” Biomed. Signal Process. Control 47, 303–311 (2019). [CrossRef]  

34. C. Tang, X. Zhu, X. Liu, M. Li, P. Wang, C. Zhang, and L. Wang, “Learning a joint affinity graph for multiview subspace clustering,” IEEE Trans. on Multimed. 21(7), 1724–1736 (2019). [CrossRef]  

35. C. Tang, Z. Li, J. Wang, X. Liu, W. Zhang, and E. Zhu, “Unified one-step multi-view spectral clustering,” IEEE Trans. on Knowledge and Data Eng. (2022).

36. K. Kang, W. Ouyang, H. Li, and X. Wang, “Object detection from video tubelets with convolutional neural networks,” in IEEE Conference on Computer Vision and Pattern Recognition, (2016), pp. 817–825.

37. C. Affonso, A. L. D. Rossi, F. H. A. Vieira, and A. C. P. de Leon Ferreira, “Deep learning for biological image classification,” Expert Systems Appl. 85, 114–122 (2017). [CrossRef]  

38. P. Li, D. Wang, L. Wang, and H. Lu, “Deep visual tracking: Review and experimental comparison,” Pattern Recognit. 76, 323–338 (2018). [CrossRef]  

39. H. Zhao, J. Shi, X. Qi, X. Wang, and J. Jia, “Pyramid scene parsing network,” in IEEE conference on computer vision and pattern recognition, (2017), pp. 2881–2890.

40. E. Korot, Z. Guan, D. Ferraz, S. K. Wagner, G. Zhang, X. Liu, L. Faes, N. Pontikos, S. G. Finlayson, H. Khalid, G. Moraes, K. Balaskas, A. K. Denniston, and P. A. Keane, “Code-free deep learning for multi-modality medical image classification,” Nat. Mach. Intell. 3(4), 288–298 (2021). [CrossRef]  

41. X. Gong, X. Xia, W. Zhu, B. Zhang, D. Doermann, and L. Zhuo, “Deformable gabor feature networks for biomedical image classification,” in Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, (2021), pp. 4004–4012.

42. A. Das, S. K. Mohapatra, and M. N. Mohanty, “Design of deep ensemble classifier with fuzzy decision method for biomedical image classification,” Appl. Soft Comput. 115, 108178 (2022). [CrossRef]  

43. M. J. Macawile, V. V. Qui nones, A. Ballado, J. D. Cruz, and M. V. Caya, “White blood cell classification and counting using convolutional neural network,” in 2018 3rd International conference on control and robotics engineering (ICCRE), (IEEE, 2018), pp. 259–263.

44. A. Khan, A. Eker, A. Chefranov, and H. Demirel, “White blood cell type identification using multi-layer convolutional features with an extreme-learning machine,” Biomed. Signal Process. Control 69, 102932 (2021). [CrossRef]  

45. B. Leng, M. Leng, M. Ge, and W. Dong, “Knowledge distillation-based deep learning classification network for peripheral blood leukocytes,” Biomed. Signal Process. Control 75, 103590 (2022). [CrossRef]  

46. A. Girdhar, H. Kapur, and V. Kumar, “Classification of white blood cell using convolution neural network,” Biomed. Signal Process. Control 71, 103156 (2022). [CrossRef]  

47. A. Rehman, N. Abbas, T. Saba, S. I. u. Rahman, Z. Mehmood, and H. Kolivand, “Classification of acute lymphoblastic leukemia using deep learning,” Microsc. Res. Tech. 81(11), 1310–1317 (2018). [CrossRef]  

48. M. Togacar, B. Ergen, and M. E. Sertkaya, “Subclass separation of white blood cell images using convolutional neural network models,” Elektronika ir Elektrotechnika 25(5), 63–68 (2019). [CrossRef]  

49. M. Toğaçar, B. Ergen, and Z. Cömert, “Classification of white blood cells using deep features obtained from convolutional neural network models based on the combination of feature selection methods,” Appl. Soft Comput. 97, 106810 (2020). [CrossRef]  

50. A. T. Sahlol, P. Kollmannsberger, and A. A. Ewees, “Efficient classification of white blood cell leukemia with improved swarm optimization of deep features,” Sci. Rep. 10(1), 2536 (2020). [CrossRef]  

51. G. Liang, H. Hong, W. Xie, and L. Zheng, “Combining convolutional neural network with recursive neural network for blood cell image classification,” IEEE Access 6, 36188–36197 (2018). [CrossRef]  

52. R. B. Hegde, K. Prasad, H. Hebbar, and B. M. K. Singh, “Comparison of traditional image processing and deep learning approaches for classification of white blood cells in peripheral blood smear images,” Biocybernet. Biomed. Eng. 39(2), 382–392 (2019). [CrossRef]  

53. A. I. Shahin, Y. Guo, K. M. Amin, and A. A. Sharawi, “White blood cells identification system based on convolutional deep neural learning networks,” Comput. Methods Programs Biomed. 168, 69–80 (2019). [CrossRef]  

54. P. P. Banik, R. Saha, and K.-D. Kim, “Fused convolutional neural network for white blood cell image classification,” in 2019 International Conference on Artificial Intelligence in Information and Communication (ICAIIC), (IEEE, 2019), pp. 238–240.

55. Y. Y. Baydilli and Ü. Atila, “Classification of white blood cells using capsule networks,” Comput. Med. Imaging and Graphics 80, 101699 (2020). [CrossRef]  

56. P. P. Banik, R. Saha, and K.-D. Kim, “An automatic nucleus segmentation and cnn model based classification method of white blood cell,” Expert Syst. Appl. 149, 113211 (2020). [CrossRef]  

57. C. Jung, M. Abuhamad, J. Alikhanov, A. Mohaisen, K. Han, and D. Nyang, “W-net: a cnn-based architecture for white blood cells image classification,” arXiv preprint arXiv:1910.01091 (2019).

58. K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, (2016), pp. 770–778.

59. S. Xie, R. Girshick, P. Dollár, Z. Tu, and K. He, “Aggregated residual transformations for deep neural networks,” in Proceedings of the IEEE conference on computer vision and pattern recognition, (2017), pp. 1492–1500.

60. S. Xie, R. Girshick, P. Dollar, Z. Tu, and K. He, “Aggregated residual transformations for deep neural networks,” in IEEE Conference on Computer Vision and Pattern Recognition, (2016), pp. 1492–1500.

61. S. Tavakoli, A. Ghaffari, Z. M. Kouzehkanan, and R. Hosseini, “New segmentation and feature extraction algorithm for classification of white blood cells in peripheral smear images,” Sci. Rep. 11(1), 19428 (2021). [CrossRef]  

62. S. Pang, A. Du, M. A. Orgun, and Z. Yu, “A novel fused convolutional neural network for biomedical image classification,” Med. Biol. Eng. Comput. 57(1), 107–121 (2019). [CrossRef]  

Data availability

Data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (5)

Fig. 1.
Fig. 1. Five different types of white blood Cells [21].
Fig. 2.
Fig. 2. The pipeline of our DRFA-Net.
Fig. 3.
Fig. 3. The brief structure of our designed STDM (top) and DTSM (bottom).
Fig. 4.
Fig. 4. An intuitive showing of the WBC region detection results of the proposed method. The top row represents original WBC images, the bottom row denotes the WBC area detection results of the proposed network.
Fig. 5.
Fig. 5. The training loss and testing accuracies of different datasets with varying iteration times.

Tables (3)

Tables Icon

Table 1. Statistic details of the training dataset BCISC

Tables Icon

Table 2. Classification performance comparison of our proposed method with other competitors in terms of four metrics

Tables Icon

Table 3. WBC classification results of the proposed methods on different datasets

Equations (13)

Equations on this page are rendered with MathJax. Learn more.

F o S T D M = Φ ( C a t ( F H , Φ ( F L ) ) ) ,
F H = Ω ( Φ ( F L ) ) Φ ( F H ) + Φ ( F H ) ,
F o D T S M = Φ ( C a t ( F L , Φ ( F H ) ) ) ,
F L = Ω ( Φ ( F H ) ) Φ ( F L ) + Φ ( F L ) ,
S = R e L U ( W l C a t ( F S T D M , F D T S M ) + b l ) ,
F D = R e L U ( W 2 C a t ( F 1 S , F 2 S , F 3 S , F 4 S , F 5 S ) + b 2 ) ,
L D ( θ ) = x = 1 W y = 1 H l { 0 , 1 } { 1 ( G ( x , y ) = l ) log Pr ( S ( x , y ) = l | θ ) } ,
L C ( θ ) = 1 N i c = 1 M 1 ( y i = c ) log Pr ( y i = c | θ ) ,
L = L D ( θ ) + λ L C ( θ ) ,
P r e c i s i o n = T P T P + F P
S e n s i t i v i t y = T P T P + F N
F 1 = 2 × P r e c i s i o n × S e n s i t i v i t y P r e c i s i o n + S e n s i t i v i t y
A c c u r a c y = T P + T N T P + F P + T N + F N
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.