Beruflich Dokumente
Kultur Dokumente
PII: S0167-8655(19)30105-9
DOI: https://doi.org/10.1016/j.patrec.2019.03.022
Reference: PATREC 7476
Please cite this article as: SanaUllah Khan, Naveed Islam, Zahoor Jan, Ikram Ud Din,
Joel J. P. C Rodrigues, A Novel Deep Learning based Framework for the Detection and Clas-
sification of Breast Cancer Using Transfer Learning, Pattern Recognition Letters (2019), doi:
https://doi.org/10.1016/j.patrec.2019.03.022
This is a PDF file of an unedited manuscript that has been accepted for publication. As a service
to our customers we are providing this early version of the manuscript. The manuscript will undergo
copyediting, typesetting, and review of the resulting proof before it is published in its final form. Please
note that during the production process errors may be discovered which could affect the content, and
all legal disclaimers that apply to the journal pertain.
ACCEPTED MANUSCRIPT
1
Highlights
T
IP
CR
US
AN
M
ED
PT
CE
AC
ACCEPTED MANUSCRIPT
2
A Novel Deep Learning based Framework for the Detection and Classification of
Breast Cancer Using Transfer Learning
SanaUllah Khana , Naveed Islama , Zahoor Jana , Ikram Ud Dinb , Joel J. P. C Rodriguesc,∗∗
T
a IslamiaCollege University Peshawar, Pakistan
IP
b The University of Haripur, Pakistan.
c National Institute of Telecommunications (Inatel), Santa Rita do Sapucaı́, MG, Brazil; Instituto de Telecomunicações, Portugal; Federal
CR
ABSTRACT
Breast cancer is among the leading cause of mortality among women in developing as well
US
as under-developing countries. The detection and classification of breast cancer in the early
stages of its development may allow patients to have proper treatment. In this article, we
proposed a novel deep learning framework for the detection and classification of breast cancer
in breast cytology images using the concept of transfer learning. In general, deep learning
AN
architectures are modeled to be problem specific and is performed in isolation. Contrary
to classical learning paradigms, which develop and yield in isolation, transfer learning is
aimed to utilize the gained knowledge during the solution of one problem into another related
problem. In the proposed framework, features from images are extracted using pre-trained
M
CNN architectures, namely, GoogLeNet, Visual Geometry Group Network (VGGNet) and
Residual Networks (ResNet), which are fed into a fully connected layer for classification of
malignant and benign cells using average pooling classification. To evaluate the performance of
the proposed framework, experiments are performed on standard benchmark data sets. It has
ED
been observed that the proposed framework outclass all the other deep learning architectures
in terms of accuracy in detection and classification of breast tumor in cytology images.
c 2019 Elsevier Ltd. All rights reserved.
PT
In biomedical research, analysis of microscopic images early cancer symptoms in women by the radiologist. It
representing different human organs and tissues play an has been observed that due to the use of mammography
important role in the understanding of different biologi- for cancer detection, the death ratio has decreased [3].
AC
cal activities. Among microscopic image examination as- A biopsy is another well efficient and accurate diagnosis
signments, classification of images (tissues, organs etc) is method for breast cancer detection. In this approach, a
one of great significance. Different applications identified tissue sample from an affected region of the breast is ana-
with microscopic image classification have been developed. lyzed under a microscope by a pathologist for the detection
Breast cancer is the most common and a leading cause of and classification of the tumor. Currently, biopsy plays a
death all over the world in women aged between 20 to 59 vital role in breast cancer as well as in other types of cancer
years [1]. If diagnosed in early stages, the survival rate diagnosis [4]. Through biopsy, pathologist can determine
from breast cancer may be increased up to 80% [2]. The two types of lesion: benign and malignant. The benign
two common diagnosing methods used for breast cancer lesion is not cancerous; it is indeed the abnormalities in
the epithelial cells, and most of these abnormalities are
unable to become a source of breast cancer. The malig-
∗∗ Correspondingauthor: nant or cancerous cells are those types of cells, which start
e-mail: joeljr@ieee.org (Joel J. P. C Rodrigues ) divisions abnormally and grows irregularly. It is a very
ACCEPTED MANUSCRIPT
3
complex and challenging task to analyze the microscopic accurate model for cell based image classification [27, 28].
images manually due to the irregular appearance of benign In the proposed framework, transfer learning has been
and malignant cells [5, 6]. exploited to overcome the deficiencies in existing systems
for the detection and classification of cancer tumor. The
In the past few decades, numerous researchers have pro-
main contribution in this paper can be summarized in the
posed different solutions for automated cells classifications
following:
for cancer detection in breast cytology images. In this
regards, some researchers have worked on nuclei analysis • To provide a framework based on deep learning archi-
by extracting features from nuclei to provide significant tecture for the detection and classification of breast
information for cell classification into benign and malig- cancer.
nant [7]. Similarly, clustering based algorithms along with
circular Hough Transform and various statistical features • To analyze the concept of transfer learning on three
are also exploited for nuclei segmentation and classifica- different deep learning architectures.
tion [8, 9, 10]. In the medical image analysis, algorithms
T
• To provide a comparative analysis of each deep learn-
for histopathological images are developing rapidly but
ing architecture with respect to accuracy in the con-
still, it is highly demanded to have an automatic system to
IP
text of transfer learning.
get efficient and highly accurate results [11, 12, 13]. There-
fore, such types of techniques are required that gives the The rest of the paper is organized as follows: Section. 2
CR
right direction towards qualitative products for diagnoses, provides a detailed analysis of the proposed approach
to provide uniformity in the results during the observation which include subsections like data pre-processing and
process and improve the objectivity. The complex nature data augmentation, pre-trained CNN architectures and
of tasks like pre-processing, segmentation, feature extrac- transfer learning. Similarly, Section. 3 discuss the ex-
tion, etc in classical machine learning approaches degrades
the performance of the system regarding efficiency and ac-
curacy.
To overcome the problems of traditional machine learn-
US perimental results obtained after applying the proposed
approach along with its performance evaluation. Finally,
Section. 4 gives the conclusion of the paper and provide
future directions.
AN
ing techniques, the concept of deep learning has been
introduced to extract the relevant information from the 2. Proposed Method
raw images and use it efficiently for classifications pro-
cess [14, 15]. In deep learning, features are not adjusted In this section, the proposed framework based on CNN
architecture is explained for the detection and classifica-
M
in the field of biomedical image analysis like detection of of GoogLeNet, VGGNet, and ResNet. The combined fea-
mitosis cells from microscopic images [16, 17] tumor de- tures are fed into a fully connected layer for the classifica-
tection [18], segmentation of neural membranes [19] skin tion task, as given in the block diagram shown in Fig. 1.
disease and its classification [20], detection and classifi- The details about each step of the proposed architecture
PT
cation of immune cells [21] and quantization of mass in is given in the following subsections.
mammograms [22]. Although, the CNN application works
very well on large data sets, yet on small data sets it fails 2.1. Data Pre-Processing and Augmentation Processing
to achieve significant gains. In order to achieve higher The pre-processing step is essential in tissue images to
CE
recognition accuracy and reduce the computational costs, remove different types of noises. In the proposed approach,
the concept of transfer learning can be exploited to im- the microscopic H&E stain tissue images are normalized
prove the performance of individual CNN architectures by using the method proposed in [29]. To achieve higher per-
AC
combining their knowledge [23, 24]. In this regards, the formance in accuracy, CNN requires large data sets. More-
set of features is extracted from generic image data sets over, the performance of CNN deteriorates with small data
using pre-trained deep CNN and then directly applied for sets due to over-fitting. It means that the network per-
domain specific and smaller data sets [25]. The concept forms very well on training data but under-perform on
of context based learning gives a new direction to transfer test data. In the proposed framework, data augmenta-
learning in which CNN is trained in two phases both for tion technique is applied to increase the data set and re-
single and overlapping patches and performed very well in duce the over-fitting problems [30, 31]. In the data aug-
breast cancer detection and classification [26]. The com- mentation method, the number of samples is increased
bination of multiple CNN architecture boosts up the per- by applying geometric transformations to the image data
formance of transfer learning and may replace the use of sets using simple image processing techniques. In this re-
traditional single model CNN architecture. Similarly, the gards, the image data set is increased by color processing,
combination of ResNet50, InceptionV2 and InceptionV3 transformation (translating, scaling, and rotation), flip-
are pre-trained on ImageNet which produced a fast and ping and noise perturbation. Since the microscopic images
ACCEPTED MANUSCRIPT
4
T
IP
CR
US
Fig. 1. Block Diagram of the Proposed Deep Learning Framework
for the classification of breast cancer in cytology images. convolution layers. VGGNet consists of 13 convolution,
These architectures are pre-trained for various generic im- rectification, pooling and 3 fully connected layers [34]. The
age descriptors, followed by relevant feature extraction convolution network uses 3×3 windows size filter and 2×2
from microscopic images on the basis of transfer learn- pooling network. VGGNet performs better as compared
AC
ing theory [36]. The basic structure of each adopted CNN to AlexNet due to its simple architecture. The underlying
architectures are described in the following sub-sections. architecture of VGGNet is illustrated in the Fig. 3.
2.2.1. GoogLeNet
It is a small network consisting of three convolution lay-
ers, rectified linear operation layers, pooling layers, and 2.2.3. ResNet
two fully connected layers. Using the architecture of ResNet is a very deep residual network and it achieves
GoogLeNet, we proposed a model which combines various good results in classification task on the ImageNet [37].
convolution filters of different sizes into a new single filter, ResNet combined multiple sized convolution filters which
which not only reduces the number of perimeters but also manage the degradation problem and reduces the training
minimizes the computational complexity. The underlying time that occurs due to its deep structures. The underly-
architecture of GoogLeNet is illustrated in the Fig. 2. ing architecture of ResNet is illustrated in the Fig. 4.
ACCEPTED MANUSCRIPT
5
3.1. Dataset
To evaluate the performance of the proposed framework
Fig. 3. Basic Architecture of VGGNet [34]
two breast microscopic image data sets are used: the first
one is a standard benchmark data set [39], and the other
T
is developed locally at LRH hospital Peshawar, Pakistan.
For both data sets, first augmentation technique is ap-
IP
plied by scaling, rotation, translation and colors modeling
to produce a total of 8000 images. In these 8000 images,
CR
6000 images are used for training the architecture while
2000 images are used for testing the trained model. In
both data sets, images are captured by microscope with
various magnifications (the enlargement process of images
seen by microscopic lens known as lens magnification). In
Fig. 4. Basic Architecture of ResNet [35]
US the proposed framework various magnified images (100X,
140X, 200X and 500X) are used for accurate evaluation.
During the execution of the proposed framework, 75% of
AN
2.3. Transfer Learning data set is used for training purpose while 25% of data
set is used for testing the accuracy of the proposed archi-
In practices, large size of data is required to train a CNN
tecture. Moreover, to control the over-fitting issues, the
from scratch but in some cases, it is very difficult to ar-
initial stopping criteria is designed which is based on the
range a big data set of relevant problems. Contrary to an
performance validation, i.e., to stop the training process
M
task on their relevant data set and then transfer to the data into combined features extraction using transfer
target task trained by target data set [38]. The transfer learning. The obtained results from single CNN is com-
learning process can be divided into two main steps: selec- pared with combined features set along with different ex-
tion of the pre-trained model, problem size and similarity. isting techniques. Table. 1 show the result of each archi-
CE
The selection of the pre-trained model is made on the basis tecture individually in different magnification sizes as well
of the associated problem which is relevant to the target as the proposed transfer learning approach.
problem. If the size of the target data set is smaller (i.e.,
AC
less then 1000 images) and similar to the source training Table 1. Magnification based Comparative Analysis of the
data set (medical data sets, hand-written character data Proposed Framework with other CNN Architectures
CNN Architectures Lens Magnification Average Classification Accuracies
sets, vehicles data sets or biometric related data sets etc.) 100X 140X 200X 500X
then the chance of over fitting is high. Similarly, if the size GoogLeNet 90.4 93.7 95.3 94.6 93.5%
VGGNet 90.8 94.8 96.7 94.2 94.15%
of the target data is larger and similar to the source data ResNet 91.5 93.3 95.4 97.2 94.35%
sets then the chance of over fitting is low and it requires Proposed Framework 96.8 96.9 97.8 98.6 97.525%
and classification of breast cancer tumor as compared to the UID/EEA/50008/2019 Project; by RNP, with re-
the other three architectures. sources from MCTIC, Grant No. 01250.075413/2018-04,
Moreover, during experimentation of the proposed ap- under the Centro de Referência em Radiocomunicações
proach, the data is splatted into training and testing data - CRR project of the Instituto Nacional de Telecomu-
sets. The splatting is performed using three different nicações (Inatel), Brazil; and by Brazilian National Coun-
procedures:90%-10%, 80%-20%, 70%-30%. The 90%-10% cil for Research and Development (CNPq) via Grant No.
splatting means that 90% data is used for training while 309335/2017-5.
the rest of 10% is used for testing the CNN architectures.
A comparative analysis of the proposed approach based
References
on data splatting is performed with other CNN architec-
tures given in Table. 2. In Table. 2, ’Class Type’ represent References
the type of cancer (B or M), where B represent benign
and M represent malignant and their respective Precision, [1] F. Bray, J. Ferlay, I. Soerjomataram, R. L. Siegel, L. A. Torre,
A. Jemal, Global cancer statistics 2018: GLOBOCAN estimates
T
Recall, F1 Score and Accuracy. It also provides an aver- of incidence and mortality worldwide for 36 cancers in 185 coun-
age accuracy of each architecture on the basis of splatting tries, CA: a cancer journal for clinicians 68 (6) (2018) 394–424.
IP
procedures. It can be noted that the proposed frame- [2] W. H. Organization, et al., WHO position paper on mammog-
raphy screening, World Health Organization, 2014.
work gives higher accuracy in the classification of cancer
[3] A. Chetlen, J. Mack, T. Chan, Breast cancer screening con-
cells in breast cytology images as compare to individual troversies: who, when, why, and how?, Clinical imaging 40 (2)
CR
architectures. (2016) 279–282.
[4] A. Chekkoury, P. Khurd, J. Ni, C. Bahlmann, A. Kamen, A. Pa-
3.3. Comparative Analysis of Accuracy with other Methods tel, L. Grady, M. Singh, M. Groher, N. Navab, et al., Automated
malignancy detection in breast histopathological images, in:
Similarly, a comparative analysis of the results obtained Medical Imaging 2012: Computer-Aided Diagnosis, vol. 8315,
using the proposed framework with four well-known meth-
ods is carried out to relates the strength of the proposed
architecture as given in Table. 3. It can be observed from
Table. 3 that the methods in [25, 26, 27, 28] give an ac-
US International Society for Optics and Photonics, 831515, 2012.
[5] C. López, M. Lejeune, R. Bosch, A. Korzynska, M. Garcı́a-Rojo,
M.-T. Salvadó, T. Álvaro, C. Callau, A. Roso, J. Jaén, Digital
image analysis in breast cancer: an example of an automated
methodology and the effects of image compression., Studies in
AN
curacy of 92.63%, 90.0%, 97.0%, and 97.5% respectively, health technology and informatics 179 (2012) 155–171.
whereas, the results obtained using the proposed frame- [6] I. Pöllänen, B. Braithwaite, T. Ikonen, H. Niska, K. Haataja,
P. Toivanen, T. Tolonen, Computer-aided breast cancer
work gives an accuracy of 97.52%, which is higher than
histopathological diagnosis: Comparative analysis of three
all the four methods. These results show the strength in DTOCS-based features: SW-DTOCS, SW-WDTOCS and SW-
M
terms of accuracy of the proposed approach as compare to 3-4-DTOCS, in: Image Processing Theory, Tools and Applica-
other similar methods. tions (IPTA), 2014 4th International Conference on, IEEE, 1–6,
2014.
[7] Z. JAN, S. KHAN, N. ISLAM, M. ANSARI, B. BALOCH,
Automated Detection of Malignant Cells Based on Structural
ED
4. Conclusion
Analysis and Naive Bayes Classifier, Sindh University Research
In this article, we proposed a novel deep learning frame- Journal-SURJ (Science Series) 48 (2).
[8] M. Kowal, P. Filipczuk, A. Obuchowicz, J. Korbicz, R. Mon-
work for the detection and classification of breast cancer czak, Computer-aided diagnosis of breast cancer based on fine
using the concept of transfer learning. In this framework, needle biopsy microscopic images, Computers in biology and
PT
features are extracting from breast cytology images us- medicine 43 (10) (2013) 1563–1572.
ing three different CNN architectures (GoogLeNet, VG- [9] P. Filipczuk, T. Fevens, A. Krzyzak, R. Monczak, Computer-
Aided Breast Cancer Diagnosis Based on the Analysis of Cy-
GNet, and ResNet) which are combined using the concept tological Images of Fine Needle Biopsies., IEEE Trans. Med.
CE
of transfer learning for improving the accuracy of classi- Imaging 32 (12) (2013) 2169–2178.
fication. Similarly, we also proposed the concept of data [10] Y. M. George, H. H. Zayed, M. I. Roushdy, B. M. Elbagoury,
augmentation to increase the size of a data set to improve Remote computer-aided breast cancer detection and diagnosis
system based on cytological images, IEEE Systems Journal 8 (3)
the efficiency of CNN structure. Finally, the performance (2014) 949–964.
AC
of the proposed framework is compared with different CNN [11] H. Irshad, A. Veillard, L. Roux, D. Racoceanu, Methods for
architectures independently and also compared with other nuclei detection, segmentation, and classification in digital
existent methods. It has been observed that the proposed histopathology: a reviewcurrent status and future potential,
IEEE reviews in biomedical engineering 7 (2014) 97–114.
framework gives excellent results regarding accuracy with- [12] M. Veta, J. P. Pluim, P. J. Van Diest, M. A. Viergever, Breast
out training from scratch which improves classification ef- cancer histopathology image analysis: A review, IEEE Trans-
ficiency. In future, both hand-crafted features along with actions on Biomedical Engineering 61 (5) (2014) 1400–1411.
[13] M. T. McCann, J. A. Ozolek, C. A. Castro, B. Parvin, J. Ko-
CNN features will be used to further improve the classifi-
vacevic, Automated histology analysis: Opportunities for signal
cation accuracy. processing, IEEE Signal Processing Magazine 32 (1) (2015) 78–
87.
[14] Y. LeCun, Y. Bengio, G. Hinton, Deep learning, nature
Acknowledgments 521 (7553) (2015) 436.
[15] Y. Bengio, A. Courville, P. Vincent, Representation learning:
This work is supported by National Funding from the A review and new perspectives, IEEE transactions on pattern
FCT - Fundação para a Ciência e a Tecnologia, through analysis and machine intelligence 35 (8) (2013) 1798–1828.
ACCEPTED MANUSCRIPT
7
Table 2. Splatting based Comparative Analysis of Proposed Framework with other CNN Architecture
Class Average
Classifier Training-Testing Precision Recall F1 Score Accuracy
Type Accuracy
Data Splitting
GoogLeNet 90%-10% B 0.93 0.94 0.94 93.67% 93.22%
M 0.96 0.94 0.95
80%-20% B 0.93 0.93 0.93 93.00%
M 0.93 0.94 0.93
70%-30% B 0.96 0.9 0.93 93.00%
M 0.92 0.98 0.95
VGGNet 90%-10% B 0.9 0.97 0.94 93.67% 94.00%
M 0.9 0.91 0.92
80%-20% B 0.97 0.96 0.95 96.00%
T
M 0.95 0.93 0.92
70%-30% B 0.91 0.92 0.94 92.33%
IP
M 0.9 0.99 0.96
ResNet 90%-10% B 0.97 0.98 0.99 98.00% 94.89%
CR
M 0.99 0.98 0.99
80%-20% B 0.99 0.9 0.99 96.00%
M 0.9 0.98 0.99
70%-30% B 0.9 0.92 0.9 90.67%
80%-20%
M
B
M
B
US
0.91
0.96
0.95
0.97
0.98
0.97
0.96
0.99
0.99
0.98
0.98
0.97
97.00%
97.67%
97.67%
AN
M 0.96 0.97 0.98
70%-30% B 0.98 0.98 0.99 98.33%
M 0.97 0.96 0.98
M
Nguyen [25] 92.63% [21] T. Chen, C. ChefdHotel, Deep learning based automatic im-
mune cell detection for immunohistochemistry images, in: In-
Awan [26] 90.00% ternational Workshop on Machine Learning in Medical Imaging,
Springer, 17–24, 2014.
Kensert [27] 97.00% [22] N. Dhungel, G. Carneiro, A. P. Bradley, Deep learning and
PT
of pathology informatics 4.
for generic visual recognition. arXiv preprint, arXiv preprint
[17] A. Cruz-Roa, A. Basavanhally, F. González, H. Gilmore,
arXiv:1310.1531 .
M. Feldman, S. Ganesan, N. Shih, J. Tomaszewski, A. Mad-
[25] L. D. Nguyen, D. Lin, Z. Lin, J. Cao, Deep CNNs for micro-
abhushi, Automatic detection of invasive ductal carcinoma in
scopic image classification by exploiting transfer learning and
whole slide images with convolutional neural networks, in: Med-
feature concatenation, in: Circuits and Systems (ISCAS), 2018
ical Imaging 2014: Digital Pathology, vol. 9041, International
IEEE International Symposium on, IEEE, 1–5, 2018.
Society for Optics and Photonics, 904103, 2014.
[26] R. Awan, N. A. Koohbanani, M. Shaban, A. Lisowska, N. Ra-
[18] A. A. Cruz-Roa, J. E. A. Ovalle, A. Madabhushi, F. A. G. Os-
jpoot, Context-Aware Learning Using Transferable Features for
orio, A deep learning architecture for image representation, vi-
Classification of Breast Cancer Histology Images, in: Interna-
sual interpretability and automated basal-cell carcinoma cancer
tional Conference Image Analysis and Recognition, Springer,
detection, in: International Conference on Medical Image Com-
788–795, 2018.
puting and Computer-Assisted Intervention, Springer, 403–410,
[27] A. Kensert, P. J. Harrison, O. Spjuth, Transfer learning with
2013.
deep convolutional neural network for classifying cellular mor-
[19] D. Ciresan, A. Giusti, L. M. Gambardella, J. Schmidhuber,
phological changes, bioRxiv (2018) 345728.
Deep neural networks segment neuronal membranes in electron
[28] S. Vesal, N. Ravikumar, A. Davari, S. Ellmann, A. Maier, Clas-
microscopy images, in: Advances in neural information process-
ACCEPTED MANUSCRIPT
8
T
Computing and Computer-assisted Intervention, Springer, 411–
418, 2013.
IP
[33] C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov,
D. Erhan, V. Vanhoucke, A. Rabinovich, Going deeper with
convolutions, in: Proceedings of the IEEE conference on com-
CR
puter vision and pattern recognition, 1–9, 2015.
[34] K. Simonyan, A. Zisserman, Very deep convolutional net-
works for large-scale image recognition, arXiv preprint
arXiv:1409.1556 .
[35] K. He, X. Zhang, S. Ren, J. Sun, Deep residual learning for
US
image recognition, in: Proceedings of the IEEE conference on
computer vision and pattern recognition, 770–778, 2016.
[36] J. Yosinski, J. Clune, Y. Bengio, H. Lipson, How transferable
are features in deep neural networks?, in: Advances in neural
information processing systems, 3320–3328, 2014.
AN
[37] Y. Yu, H. Lin, Q. Yu, J. Meng, Z. Zhao, Y. Li, L. Zuo, Modality
classification for medical images using multiple deep convolu-
tional neural networks, J Comput Inf Syst 11 (2015) 5403–13.
[38] L. Yang, S. Hanneke, J. Carbonell, A theory of transfer learning
with applications to active learning, Machine learning 90 (2)
(2013) 161–189.
M