Keywords
Microscopic Images, White Blood Cells, Image Processing, Image Segmentation, Image Classification, Convolutional Neural Network
This article is included in the Research Synergy Foundation gateway.
Microscopic Images, White Blood Cells, Image Processing, Image Segmentation, Image Classification, Convolutional Neural Network
According to,1 peripheral blood (PB) or whole blood is the circulating fluid through the entire human body. PB delivers oxygen and nutrients to all the body cells, tissues, and organs, and removes the carbon dioxide and other waste products. It consists of erythrocytes (red blood cells, RBCs), leukocytes (white blood cells, WBCs) and thrombocytes (platelets). RBCs transport oxygen from the lungs to all the body tissues; WBCs fight against the harmful bacteria, parasitic and fungal infections; while platelets clot the blood in wounds on surfaces of the tissue layers. WBCs with a single granulocyte are the monocyte and lymphocyte; while basophil, eosinophil, neutrophil are the WBCs with more than one granulocyte. Lymphocyte can be affected by acute lymphoblastic leukemia (ALL).
Leukemia is a disease formed in tissues that produce large portions of malfunctional and abnormal WBCs that spreads from bone marrow. Based on findings of,2 chronic lymphoblastic leukemia (CLL) (35%) and acute myelogenous leukemia (AML) (32%) are the most common leukemias for adults, while ALL (75%) affects children and teens the most. The World Health Organization (WHO) once stated that ALL is one of the six cancers for children that requires extra attention.3
ALL, AML, CLL and chronic myelogenous leukemia (CML) are leukemia subtypes. Fast-growing cancer in lymphoid cells results in the formation of ALL,2 as opposed to CML; while fast-growing cancer in myeloid cells resulted in the formation of AML, as opposed to CLL. Those over 50 years old and children below 5 years old are the main affected populations of ALL, and the disease can be fatal if not treated earlier.4
PB smear analysis can detect potential disorders and inform health status, while pathology tests can help in tracking the ongoing status of infections, allergies, cancers etc.5 A WBC test is one sub-element of a complete blood count (CBC), one of the pathology tests that helps doctors discovers the unexposed infections. The University of Roschester Medical Center (UMRC) has declared that the normal range of WBCs per microliter of blood is 4,000 to 11,000; if WBCs exceed 11,000 per microliter, it is termed as leukocytosis.
Traditional PB smear analysis and CBCs are based on the human inspection. It is laboriously suffered from the intra-observer variability and is not time-efficient or cost efficient. Today, computer-aided diagnosis (CADx) systems are employed and contain four main steps: preprocessing, segmentation, feature extraction, and classification.
Previous work gained 99.14% and 94.12% accuracy for the nucleus segmentation and cell segmentation respectively, while classification accuracy was over 90%. Cell segmentation and binary classification works can be improved further.
Studies by6,7 summarized WBC segmentation works into pattern recognition-based, deformable model-based, threshold-based, morphological operations-based, and clustering-based segmentation.7 suggested a combination of dual-threshold and morphological operations which achieved 97.85% accuracy. Dual-threshold, binarization, and morphological opening were applied on both preprocessed contrast-stretched grey and H components images. A threshold method was also proposed by other researchers.8–12 recommended a watershed-based and Otsu threshold-based segmentation which resulted in 99.3% and 93.3% accuracy, respectively.
K-means clustering was another famous segmentation approach.13–16,20,32 applied K-means clustering-based segmentation on the G component of RGB image for two datasets and gained 99.51% and 99.74% accuracy respectively; when applied to a CMYK image, 98.89% accuracy was obtained.16
Deep learning (DL) performed object class prediction by recognizing and learning patterns in visual inputs, making it the state-of-the-art method today. Region of interest (ROI) and neural networks were two other parts of machine learning that were popular nowadays.17 Recommended semantic segmentation that performed whole-slide WBCs segmentation and received 93.34% accuracy.18 Used four-moment statistical features and artificial neural networks (ANN) to segment based on local pixel information, and the overall accuracy was 97%. Work done by19 included WBC localization and Grabcut to perform WBCs segmentation. Edge density (ED) and color contrast (CC) were measured. DL gains high segmentation accuracy; however, it has complex architecture, making it challenging to be a robust and generalized DL model.
For the WBCs classification works,20–22 employed a deep convolutional neural network (DCNN); 29 suggested two methods, neural network (NN) combined with the autoencoders, and a convolutional neural network (CNN); 23–25 also proposed CNN; 26,27 suggested the support vector machine (SVM); 28 suggested the combination of K-means neighbours and social spider optimization. Works of23,29,24,21 yielded great outputs.24 used DCNN to classify WBCs into monocytes, neutrophils, eosinophils, and lymphocytes with accuracy of 92.14%, 94.72%, 91.25%, and 94.61% respectively.
This research project is a continuing work that aims to gain improved results for both the WBCs segmentation and classification works. The cell segmentation from the previous work has achieved high accuracy for the nucleus segmentation.
The input dataset was the ALL-IDB2 dataset provided by the Department of Information Technology at University degli Studi di Milano.30 ALL-IDB2 contains a total of 260 normal and blasts WBCs images collected from blood samples of ALL patients which were designed to test the cell classification efficiency. Classification and labelling process to get the ground truth images was done by the expert oncologists. This dataset was captured using an optical laboratory microscope coupled with a Canon PowerShot G5 camera, saved as JPG format.
Figure 1 shows the overall flowchart of the proposed cell segmentation. H component is extracted from the transformed HSV formatted microscopic image (Figure 2 and Figure 3). Subsequently, Gaussian filtering and Otsu thresholding31,11,32 are applied three times to remove RBCs, RBCs boundaries, and to segment the WBCs. Results for each round can be viewed in Figure 4. Then, a 13 × 13 morphological opening is employed to remove small objects from the foreground to get better results as shown in Figure 5. Finally, cell masks are collected to detect the cell contour from the original RGB formatted image as shown in Figure 6.
Figure 7 shows the flowchart of the formation of the proposed binary WBCs classifier model. Characteristics and conditions of WBCs are crucial for this work.
A pre-trained InceptionV3 model is employed.33–35 It is used to extract features in a data preprocessing step which was inspired by.12 Also, the transfer learning technique is adopted. Dataset is divided into training, validation and testing sets, 208 images for training set, while validation and test sets each contains 26 images. All the input images are re-sized to 150 × 150 pixels and are normalized to (−1, 1) to fit into InceptionV3 model.
First, a new classifier is built and its architecture is shown in Figure 8. It has 2049 trainable parameters, global average pooling that forms a feature map to prevent overfitting, and a dense layer that applies sigmoid activation to do the binary classification. The pre-trained InceptionV3 model extracts all the features from the training and validation sets. Next, the model is trained for 300 epochs. The training learning curves graph (accuracy and loss) of the newly defined classifier are shown in Figure 9 and Figure 10. The point of interception has the highest validation accuracy and smaller differences between training and validation accuracy. Hence, the epoch value along with the lead of intersection between the two curve lines is the best epoch chosen as a final classifier. New classifier will then be combined with the pre-trained InceptionV3 to form the final classifier. As shown in Figure 11, out of 21,804,833 parameters, the final classifier had a total of 6,321,857 trainable parameters.
Data augmentation increases the diversity of data using various techniques such as flipping, rotation, translation etc. The benefit that comes along with this technique is reducing the model bias tendency towards a particular class of data. Thus, it is applied to both the training and validation sets to allow the model to learn better and to reduce the overfitting consequences. The training learning curves graph (accuracy and loss) of the fine-tuned classifier model can be viewed in Figure 12 and Figure 13. Epoch with the highest validation accuracy and lowest validation loss is chosen as the final classifier.
Figure 14 shows some results of the proposed cell segmentation that attained 90.45%, 93.91%, 76.25% and 0.0186 seconds for the average accuracy, structural similarity, dice similarity coefficient and computational time respectively. The final results achieve high accuracy and similarity, over the average overlapping between the ground truth and the original images with short computation time. In short, the proposed cell segmentation method makes a performance.
The sample results shown in Figure 14 demostrate that the first row yields an excellent segmentation outcomes while the following rows of images weren’t. This may be due to the undesirable illuminations and blur levels. In addition, the inconspicuous difference between the foreground (WBC) and the background caused by the uneven acquisition can be another factor that affects the final results.
Table 1 and the bar chart in Figure 15 show the comparison between the K-means clustering and the proposed triple thresholding method. This work obtained higher accuracy in a shorter time.
K-means clustering | Triple thresholding | |
---|---|---|
Accuracy | 0.8905 | 0.9045 |
Structural similarity | 0.8005 | 0.8381 |
Dice similarity coefficient | 0.7438 | 0.7625 |
Computational time | 1.1935 | 0.0186 |
Table 2 shows the performance of the newly defined classifier; while Table 3 shows the performance of the fine-tuned classifier.
It is observed that the accuracy, precision, recall, true positive and true negative for all the three sets of data increased while loss, false positive and false negative results decreased after the fine-tuned stage. This indicates that the fine-tuned classifier is a robust and good fitted classifier that can accurately predict and make fewer errors.
Figure 16 and Figure 17 are the samples of the binary classification results. If the prediction value is over 0.5, they are considered as healthy WBCs; otherwise, they are ALL WBCs. Results of the predicted label by the classifier and the actual label are the same. Both the sample images are classified correctly.
Comparisons between the previous works and the proposed work for the binary classification are shown in Table 4 and the bar chart in Figure 18. The proposed binary classification work outperforms previous works with the highest accuracy, 96.15%. The proposed work is able to classify the WBCs into healthy and malignant groups accurately.
Proposed method | Accuracy (%) | |
---|---|---|
[29] | Feature extraction using SSOA and KNN | 95.23 |
[36] | VGGNet features extractor and Improved Swarm optimization (SESSA) features selector | 96.11 |
[12] | Transfer learning on MobileNetV2 classifier | 92.31 |
Proposed work | Transfer learning on pre-trained InceptionV3 CNN classifier | 96.15 |
The proposed method in this study shows a new possible approach or direction for the future research work on WBC cell segmentation and binary classification.
In the comparison table between the previous works and the proposed work for the cell segmentation, it is proven that the threshold-based segmentation can outperform the K-means clustering when dealing with a small dataset. Otsu thresholding uses the global thresholding while the K-means clustering uses the local thresholding to perform the segmentation. Otsu thresholding calculates the optimum thresholding after distributing the pixels into foreground and background classes. K-means clustering determines the k centroids, evaluates pixels and groups the similar pixels into the same cluster. Most images of the ALL-IDB2 contain large foreground (WBC) size leading to a good bimodal distribution of the histogram where Otsu thresholding is more capable and expertise than K-means clustering to perform segmentation due to the thresholding technique used. The pre-processing and post-processing are crucial as they can very much affect the final output of the Otsu thresholding.
On the other hand, the proposed binary WBC classification that uses the pre-trained InceptionV3 and transfer learning technique yields 96% of accuracy. The pre-trained classifier model is trained with a large dataset, hence, it eliminates the need to build a classifier model from scratch. This makes it able to extract features better than VGGNet or SSOA in previous works. Transfer learning technique reuses pre-trained model on a different but similar problem is a noble approach. This gives the model a higher learning rate during the training. It is not only more intelligent but also able to accomplish works in higher performance. Using pre-trained and transfer learning approaches, the binary classifier model is more brilliant and flexible than the previous research. It can adapt quickly and use the best-fitted methods to perform classification on the dataset. Also, the InceptionV3 model requires fewer generated parameters of networks as compared with VGGNet. Hence, the final results are better than the previous works when combined with the binary WBCs classification. From Table 4, it is observed that the proposed work has a potential to be considered for the future application.
This research focuses on improving the WBC cell segmentation and binary classification works. A public dataset, ALL-IDB2 is used to perform the cell segmentation and binary classification. Triple thresholding method is proposed to achieve the first objective of this research and the results are 90.45%, 83.81%, 76.25%, and 0.0186 seconds for the accuracy, structural similarity index, dice similarity coefficient, and computation time respectively. Combining the pre-trained InceptionV3 model and the transfer learning technique produce over 96% accuracy and the precision with a lower loss value was suggested to accomplish the second objective of this research. The overall performance for both the segmentation and classification works have been improved.
LamXH, NgKW, YoongYJ and NgSB conceived the presented idea. LamXH carried out the experiment and wrote the manuscript. NgKW, YoongYJ and NgSB supervised the project and provided feedback.
Ethical Approval Body: Research Ethic Committee 2021, Multimedia University
Ethical Approval Number: EA1552021
Data were obtained from Acute Lymphoblastic Leukemia Image Database for Image Processing (ALL-IDB) (https://homes.di.unimi.it/scotti/all/).
This dataset was not generated nor is it owned by the authors of this article; the listed owners is the Department of Computer Science - Università degli Studi di Milano. Therefore, neither the authors nor F1000Research are responsible for the content of this dataset and cannot provide information about data collection.
We thank the anonymous reviewers for their careful reading of our manuscript and their insightful comments and suggestions.
Views | Downloads | |
---|---|---|
F1000Research | - | - |
PubMed Central
Data from PMC are received and updated monthly.
|
- | - |
Is the work clearly and accurately presented and does it cite the current literature?
Partly
Is the study design appropriate and is the work technically sound?
Yes
Are sufficient details of methods and analysis provided to allow replication by others?
Yes
If applicable, is the statistical analysis and its interpretation appropriate?
Yes
Are all the source data underlying the results available to ensure full reproducibility?
Yes
Are the conclusions drawn adequately supported by the results?
Yes
References
1. Pałczyński K, Śmigiel S, Gackowska M, Ledziński D, et al.: IoT Application of Transfer Learning in Hybrid Artificial Intelligence Systems for Acute Lymphoblastic Leukemia Classification.Sensors (Basel). 2021; 21 (23). PubMed Abstract | Publisher Full TextCompeting Interests: No competing interests were disclosed.
Reviewer Expertise: Deep Learning, Machine Learning
Is the work clearly and accurately presented and does it cite the current literature?
Yes
Is the study design appropriate and is the work technically sound?
Yes
Are sufficient details of methods and analysis provided to allow replication by others?
Yes
If applicable, is the statistical analysis and its interpretation appropriate?
Yes
Are all the source data underlying the results available to ensure full reproducibility?
Yes
Are the conclusions drawn adequately supported by the results?
Yes
References
1. Hegde R, Prasad K, Hebbar H, Singh B: Comparison of traditional image processing and deep learning approaches for classification of white blood cells in peripheral blood smear images. Biocybernetics and Biomedical Engineering. 2019; 39 (2): 382-392 Publisher Full TextCompeting Interests: No competing interests were disclosed.
Reviewer Expertise: Myself working in the area of machine learning, deep learning and made good number of publications in bioinformatics, machine learning.
Alongside their report, reviewers assign a status to the article:
Invited Reviewers | ||
---|---|---|
1 | 2 | |
Version 1 17 Nov 21 |
read | read |
Provide sufficient details of any financial or non-financial competing interests to enable users to assess whether your comments might lead a reasonable person to question your impartiality. Consider the following examples, but note that this is not an exhaustive list:
Sign up for content alerts and receive a weekly or monthly email with all newly published articles
Already registered? Sign in
The email address should be the one you originally registered with F1000.
You registered with F1000 via Google, so we cannot reset your password.
To sign in, please click here.
If you still need help with your Google account password, please click here.
You registered with F1000 via Facebook, so we cannot reset your password.
To sign in, please click here.
If you still need help with your Facebook account password, please click here.
If your email address is registered with us, we will email you instructions to reset your password.
If you think you should have received this email but it has not arrived, please check your spam filters and/or contact for further assistance.
Comments on this article Comments (0)