Sie sind auf Seite 1von 8

ORIGINAL ARTICLE

Diagnostic outcomes of esophageal cancer by artificial


intelligence using convolutional neural networks
Yoshimasa Horie, MD,1,2 Toshiyuki Yoshio, MD,1,3 Kazuharu Aoyama, MM,4 Shoichi Yoshimizu, MD,1
Yusuke Horiuchi, MD,1 Akiyoshi Ishiyama, MD,1 Toshiaki Hirasawa, MD,1,3 Tomohiro Tsuchida, MD,1
Tsuyoshi Ozawa, MD,3,5 Soichiro Ishihara, MD,3,5 Youichi Kumagai, MD,6 Mitsuhiro Fujishiro, MD,7
Iruru Maetani, MD,2 Junko Fujisaki, MD,1 Tomohiro Tada, MD3,4,8
Tokyo, Saitama, Japan

Background and Aims: The prognosis of esophageal cancer is relatively poor. Patients are usually diagnosed at
an advanced stage when it is often too late for effective treatment. Recently, artificial intelligence (AI) using deep
learning has made remarkable progress in medicine. However, there are no reports on its application for diag-
nosing esophageal cancer. Here, we demonstrate the diagnostic ability of AI to detect esophageal cancer
including squamous cell carcinoma and adenocarcinoma.
Methods: We retrospectively collected 8428 training images of esophageal cancer from 384 patients at the Cancer
Institute Hospital, Japan. Using these, we developed deep learning through convolutional neural networks
(CNNs). We also prepared 1118 test images for 47 patients with 49 esophageal cancers and 50 patients without
esophageal cancer to evaluate the diagnostic accuracy.
Results: The CNN took 27 seconds to analyze 1118 test images and correctly detected esophageal cancer cases with a
sensitivity of 98%. CNN could detect all 7 small cancer lesions less than 10 mm in size. Although the positive predictive
value for each image was 40%, misdiagnosing shadows and normal structures led to a negative predictive value of 95%.
The CNN could distinguish superficial esophageal cancer from advanced cancer with an accuracy of 98%.
Conclusions: The constructed CNN system for detecting esophageal cancer can analyze stored endoscopic im-
ages in a short time with high sensitivity. However, more training would lead to higher diagnostic accuracy. This
system can facilitate early detection in practice, leading to a better prognosis in the near future. (Gastrointest
Endosc 2018;-:1-8.)

Esophageal cancer is the eighth most common cancer cell carcinoma (ESCC) remains the most common
worldwide, with an estimated 456,000 new cases each histologic type of esophageal cancer in Asia and in Japan.
year. It is the sixth most common cause of cancer-related When esophageal cancer is diagnosed at an advanced
mortality with an estimated 400,000 deaths per year.1 stage, it requires a highly invasive treatment, and its
Although the incidence of esophageal adenocarcinoma prognosis is poor. Therefore, early detection is of great
(EAC) has been increasing globally, esophageal squamous importance.

Abbreviations: AI, artificial intelligence; CNN, convolutional neural Division of Gastroenterology and Hepatology, Department of Internal
network; EAC, esophageal adenocarcinoma; ESCC, esophageal squa- Medicine, Toho University Ohashi Medical Center, Tokyo, Japan (2), Tada
mous cell carcinoma; NBI, narrow-band imaging; PPV, positive predic- Tomohiro Institute of Gastroenterology and Proctology, Saitama, Japan
tive value; NPV, negative predictive value; WLI, white-light imaging. (3), AI Medical Service Inc, Tokyo, Japan (4), Surgery Department, Sanno
Hospital, International University of Health and Welfare, Tokyo, Japan (5),
DISCLOSURE: All authors disclosed no financial relationships relevant
Department of Digestive Tract and General Surgery, Saitama Medical
to this publication.
Center, Saitama Medical University, Saitama, Japan (6), Department of
Copyright ª 2018 by the American Society for Gastrointestinal Endoscopy Gastroenterology, Graduate School of Medicine (7), Department of
0016-5107/$36.00 Surgical Oncology, Graduate School of Medicine (8), The University of
https://doi.org/10.1016/j.gie.2018.07.037 Tokyo, Tokyo, Japan.
Received April 16, 2018. Accepted July 29, 2018. Reprint requests: Toshiyuki Yoshio, MD, PhD, Department of
Gastroenterology Cancer Institute Hospital, 3-8-31, Ariake, Koto-ku,
Current affiliations: Department of Gastroenterology, Cancer Institute
Tokyo 135-8550, Japan.
Hospital, Japanese Foundation for Cancer Research, Tokyo, Japan (1),

www.giejournal.org Volume -, No. - : 2018 GASTROINTESTINAL ENDOSCOPY 1


AI for diagnosing esophageal cancer Horie et al

It is difficult to make an endoscopic diagnosis of ESCC TABLE 1. Patient and lesion characteristics in the test image sets
during the early stages with white-light imaging (WLI)
alone. Iodine staining has been used as the preferred Patient characteristics (n [ 47) Value
method of screening in high-risk patients. However, the Sex, male/female 41/6
use of iodine staining is associated with problems, such Median age, y (range) 70 (48-81)
as chest pain or discomfort and an increased procedural
Lesion characteristics (n [ 49)
time. Narrow-band imaging (NBI) is a revolutionary tech-
Median tumor size, mm (range) 20 (5-700)
nology of optical image-enhanced endoscopy that facili-
tates the detection of ESCC as a brownish area without Tumor location, Ce/Ut/Mt/Lt/Ae 0/8/23/10/8
using iodine staining,2-4 and the results are available in a Macroscopic type
moment by simply pressing a button. NBI is considered su- Superficial cancer, 0-I/0-IIa/0-IIb/0-IIc 1/6/13/23
perior to iodine for its ease of use and because it causes Advanced cancer 1/2/3/4 0/3/3/0
less discomfort for patients. However, NBI has demon- Depth of tumor, T1a/T1b/T2-4 40/2/7
strated an insufficient sensitivity of 53% for detecting
Histopathology, ESCC/EAC 41/8
ESCC when used by inexperienced endoscopists.5
ESCC, Esophageal squamous cell carcinoma; EAC, esophageal adenocarcinoma; Ut,
Therefore, a real-time diagnosis supporting system may upper thoracic esophagus; Mt, middle thoracic esophagus; Lt, lower thoracic
be of great help for endoscopists in the detection of super- esophagus; Ae, abdominal esophagus; Ce, cervical esophagus.
ficial ESCC.
In recent years, artificial intelligence (AI) using deep
learning has made remarkable progress in various medical EVIS LUCERA ELITE CV-290/CLV-290SL; Olympus Medical
fields, especially as a system to screen medical images. These System).
fields include diagnosis in radiation oncology,6 skin cancer We collected 8428 training images of esophageal cancer
classification,7 diabetic retinopathy,8 histologic classification lesions that were histologically proven to be SCC or adeno-
of gastric biopsy specimens,9 and characterization of carcinoma in 384 patients. The training esophageal cancer
colorectal lesions using endocytoscopy.10 We have also images included 397 lesions of ESCC that consisted of 332
reported on AI diagnoses using endoscopic images of lesions of superficial cancer and 65 lesions of advanced
Helicobacter pylori infections11 and of gastric cancer cancer. A total of 32 images of EAC were also included
detection.12 However, no reports have assessed the for training and consisted of 19 lesions of superficial can-
usefulness of AI in the diagnosis of esophageal cancer. cers and 13 lesions of advanced cancers. Multiple images
In the context of medical imaging, deep learning can were prepared for each lesion by controlling the air vol-
become a powerful supportive tool that can interpret med- ume in the esophagus to distend the esophageal wall
ical images based on an accumulated set of unique algo- and control the distance to the lesions. The training images
rithms. Deep learning allows computational models that included standard WLI and NBI, excluding poor quality im-
are composed of multiple processing layers to learn repre- ages resulting from halation, blur, defocus, mucus, and
sentations of data with multiple levels of abstraction.13 poor insufflation of the air. Magnified images by magni-
To develop deep learning through a convolutional neu- fying endoscopy were also excluded from this study. All im-
ral network (CNN) for the detection of esophageal cancer, ages of esophageal cancer lesions were marked manually
we constructed an AI-based diagnostic system that was by 1 author (Y.H.) who is a well-experienced endoscopist,
trained using a large number of EGD images. Here we with more than 5 years of experience and having per-
demonstrate the diagnostic ability of this system to detect formed 3000 examinations in a high-volume cancer center.
esophageal cancer including ESCC and EAC. To evaluate the diagnostic accuracy of the constructed
CNN, an independent set of test images were prepared
from 47 patients with 49 esophageal cancer lesions,
METHODS including 41 ESCCs and 8 EACs (162 images of esophageal
cancer and 376 images without cancer) and from 50 patients
Preparation of training and test image sets with no esophageal cancer (573 images of the noncancerous
For this single-center retrospective study, we obtained part of the esophagus). Of the 47 patients with esophageal
EGD images from February 2016 to April 2017 from the cancer lesions, 45 had 1 lesion each, whereas 2 of them
Cancer Institute Hospital, Tokyo, Japan, to develop an al- had 2 lesions each. There was a median number of 2 images
gorithm for the detection of esophageal cancer. The (range, 1-6) for every lesion with WLI and 1 image per lesion
EGD was performed for screening or preoperative exami- (range, 0-4) with NBI. During the procedure, the entire
nations in daily clinical practice, and the images were esophagus was observed, and images of every part of the
captured using high-resolution endoscopes (GIF-H290Z, esophagus were captured using WLI and NBI.
GIF-H290, GIF-XP290N, GIF-H260Z, GIF-H260; Olympus All cases of esophageal cancer were confirmed to have
Medical Systems, Co, Ltd, Tokyo, Japan) and standard no other cancer using WLI, NBI, iodine staining, and
endoscope video systems (EVIS LUCERA CV-260/CLV-260, follow-up endoscopy after the treatment.

2 GASTROINTESTINAL ENDOSCOPY Volume -, No. - : 2018 www.giejournal.org


Horie et al AI for diagnosing esophageal cancer

Figure 1. Examples of AI diagnosed images. A and B, For WLI and NBI, the CNN recognized esophageal cancer (white square) and was matched with a
green square that we used to identify the cancer. C and D, The CNN could not recognize esophageal cancer (green square) in WLI but could recognize it
in NBI (white and green squares), and it was matched with the correct diagnosis (green square). All recognized lesions were diagnosed as superficial
cancers. WLI, White-light imaging; NBI, narrow-band imaging; CNN, convolutional neural networks; AI, artificial intelligence.

Constructing a CNN algorithm cancer from the input data of the test images, a disease
To construct an AI-based diagnostic system, we used a name (superficial or advanced esophageal cancer) was as-
deep neural network architecture called Single Shot Multi- signed and a rectangular frame was displayed in the endo-
Box Detector (https://arxiv.org/abs/1512.02325), without scopic image so as to surround the lesion of interest.
altering its algorithm as described previously.12 The Several criteria were chosen to evaluate the diagnostic per-
Single Shot MultiBox Detector is a deep CNN that formance of the CNN for the detection of esophageal cancer
consists of 16 layers or more. A Caffe deep learning and included the following. When the CNN could recognize
framework, one of the most widely used frameworks even a part of the cancer, the CNN was considered as correct,
originally developed at the Berkeley Vision and Learning because it is sometimes difficult to identify the whole bound-
Center, was used to train, validate, and test the CNN. ary of cancer in 1 image. However, even when there was can-
All layers of the CNN were fine-tuned using stochastic cer in the frame judged by the CNN, it was considered
gradient descent with a global learning rate of .0001. incorrect if it contained wide noncancerous sites occupying
Each image was resized to 300  300 pixels; the bounding more than 80% of the frame. In the image with findings
box was also resized accordingly. These values were set up that indicated cancer, when the CNN recognized noncan-
by trial and error to ensure all data were compatible with cerous sites as cancerous, it was judged as a false-positive
Single Shot MultiBox Detector. result. When the CNN diagnosed 2 noncancerous sites as can-
cers in 1 image without cancer, it was counted as 1 false pos-
Outcome measures of AI diagnosis itive. In the image without cancer, when the CNN recognized
After constructing the CNN using the training image set, a noncancerous site as cancer, it was considered a false pos-
performance evaluation was performed using independent itive. In the case with 2 lesions, only when the CNN could
test images. When the trained CNN detected esophageal recognize both lesions was it considered a correct diagnosis.

www.giejournal.org Volume -, No. - : 2018 GASTROINTESTINAL ENDOSCOPY 3


AI for diagnosing esophageal cancer Horie et al

TABLE 2. Detailed results of AI diagnosis for each case P <.05

Results of biopsy 100


specimen
Cancer* Noncancer
80

AI diagnosis by WLI 60

Sensitivity
Cancer 38 35
40
Noncancer 9 15
AI diagnosis by NBI* 20
Cancer 41 28
Noncancer 5 22 0
Comprehensive AI diagnosisy A WLI NBI Comprehensive
diagnosis
Cancer 46 42
100
Noncancer 1 8 P = .32
AI, Artificial intelligence; WLI, white-light imaging; NBI, narrow-band imaging. 80
*Because there was 1 case without NBI images, a total number of cases diagnosed by

Sensitivity
NBI were 46 cases.
yIn the comprehensive CNN diagnosis, when diagnosed with either WLI or NBI, we
60
defined the CNN had been correctly diagnosed.
40

Definition of the analysis 20


The sensitivity for the CNN’s ability to detect esophageal
0
cancer for each case was calculated as follows: number of cases B WLI NBI
in which the CNN correctly diagnosed esophageal cancer O
all cases of esophageal cancer. In a comprehensive analysis, Figure 2. The sensitivity of the AI diagnosis for each case and for each
when the CNN detected the esophageal cancer with either image. A, The sensitivity of the CNN diagnosis for each case is shown in
WLI or NBI, we defined this as the correct diagnosis. the graph. The sensitivity for the comprehensive diagnosis was signifi-
cantly higher than that in the WLI (P < .05). B, The sensitivity of the
Sensitivity, specificity, positive predictive value (PPV), CNN diagnosis for each image in NBI was slightly higher than in WLI
and negative predictive value (NPV) for the CNN’s ability but not significantly. AI, Artificial intelligence; WLI, White-light imaging;
to detect esophageal cancer for each image was calculated NBI, narrow-band imaging; CNN, convolutional neural networks.
as follows: sensitivity, number of images in which the CNN
correctly diagnosed cancer O all images with cancer; spec- Ethics
ificity, number of images in which AI correctly diagnosed This study was approved by the Institutional Review
noncancerous lesions O all images without cancer; PPV, Board of the Cancer Institute Hospital (no. 2016-1171)
number of images in which the CNN correctly diagnosed and the Japan Medical Association (ID JMA-II A00283).
cancer O all images in which the CNN diagnosed cancer;
and NPV, number of images in which the CNN correctly
diagnosed noncancerous lesions O all images the CNN RESULTS
diagnosed as noncancerous lesions. Macroscopic types of
lesions were assessed according to the Japanese Classifica- Characteristics of patients and lesions in test
tion of Esophageal Cancer edited by the Japan Esophageal image sets
Society.14 Superficial cancer included mucosal (T1a) and The CNN required 27 seconds to analyze the test image
submucosal (T1b) cancer, and advanced cancer included sets of 1118 images for 97 patients. The median tumor size
any cancer that was deeper than the muscularis propria was 20 mm, and 82% of the lesions were mucosal cancers
(T2-4).14 (T1a), 4% were submucosal cancers (T1b), and 14% were
advanced cancers deeper than the muscularis propria
(T2-4). For histopathology, 84% of the lesions were SCC
Statistical analysis and 16% were adenocarcinoma (Table 1).
All continuous variables are expressed as the median The example test images are shown in Figure 1. When
with a range. Statistical analyses were conducted using the CNN recognized a cancer, the CNN marked the
Fisher’s exact test using the GraphPad Prism (GraphPad cancer with a white square that indicated superficial and
Software, Inc, La Jolla, Calif). A P < .05 was considered sta- advanced cancers (Fig. 1A, B, and D). We marked the
tistically significant. To compare the sensitivities of WLI, cancer lesion by a green square independently (Fig. 1A-
NBI, and a comprehensive diagnosis for each case, a P D). By comparing the green square with the white
value was adjusted by Bonferroni correction. square, we were able to identify if the CNN could

4 GASTROINTESTINAL ENDOSCOPY Volume -, No. - : 2018 www.giejournal.org


Horie et al AI for diagnosing esophageal cancer

TABLE 3. Detailed results of the CNN diagnosis for each image TABLE 5. Causes for false positives and false negatives in the CNN
diagnosis
Results of biopsy specimen
Cancer Noncancer No. of images
Causes for false positives (%)
CNN diagnosis by WLI
Shadow 95 (50)
Cancer 61 109
Normal structure 61 (32)
Noncancer 24 364 EGJ/left main bronchus/vertebral body) 29/25/7
CNN diagnosis by NBI Benign lesion 32 (17)
Cancer 64 87 Post-ER scar/focal atrophy/Barrett’s esophagus/ (13/9/5/4/1/1/1)
inflammation/intraepithelial neoplasia/
Noncancer 13 396
keratinization /ectopic gastric mucosa
CNN, convolutional neural networks; WLI, white-light imaging; NBI, narrow-band
imaging.
Causes for false negatives
Inflammation in background mucosa 10 (25)

TABLE 4. Accuracy in diagnosing superficial versus advanced cancer Obscure ESCC in WLI 7 (17)
Barrett’s esophageal EAC 4 (10)
WLI NBI All images
Difficult conditions (lesions were too distant, only 20 (49)
Superficial cancer, % 100 (75/75) 99 (67/68) 99 (142/143) piece of lesion)
Advanced cancer, % 100 (14/14) 82 (9/11) 92 (23/25) CNN, Convolutional neural networks; WLI, white-light imaging; ESCC, esophageal
squamous cell carcinoma; EAC, esophageal adenocarcinomas; ER, endoscopic
All cancer, % 100 (89/89) 96 (76/79) 98 (165/168) resection; EGJ, esophagogastric junction.
Accuracy: number of superficial or advanced cancer images that were diagnosed
correctly by the CNN O number of images that the CNN diagnosed as cancer.
WLI, White-light imaging; NBI, narrow-band imaging; CNN, convolutional neural The CNN diagnosis of the images also had a specificity of
networks. 79%, PPV of 39%, and NPV of 95%. Although not
significant, NBI tended to have higher sensitivity than
diagnose the esophageal cancer or not. In the first WLI (83% vs 72%, P Z .32) (Fig. 2B). The sensitivity of
example, the CNN could diagnose the cancer in both the the CNN diagnosis for ESCC on WLI and NBI was 72%
WLI and the NBI (Fig. 1A and B). In the second example, and 86%, respectively, and that for EAC was 69% and
the CNN could not diagnose cancer in the WLI but could 71%, respectively.
diagnose the cancer in the NBI (Fig. 1C and D). Both
lesions in the example were diagnosed as superficial CNN diagnosis: superficial versus advanced
cancer (Fig. 1A, B, and D). cancer
Each image of esophageal cancer, whether taken with
CNN diagnosis for each case WLI or NBI, was separately diagnosed as either superficial
The CNN correctly diagnosed 46 of 47 cases of esopha- or advanced cancer by the CNN. The accuracy in diagnosis
geal cancers with the comprehensive diagnose of WLI and was quite high for both imaging modalities. The diagnostic
NBI (Table 2). Notably, CNN could detect all 7 lesions less accuracy was 99% (142/143) for superficial cancer and 92%
than 10 mm in size. (23/25) for advanced cancer (Table 4). The accuracy in
The sensitivity of the comprehensive CNN diagnosis was diagnosis for ESCC and EAC was 99% (146/147) and 90%
98%, whereas the sensitivity of the CNN diagnosis based on (19/21), respectively.
WLI and on NBI was 81% and 89%, respectively (Fig. 2A).
The NBI was more sensitive to diagnosing esophageal Causes for false positives and false negatives
cancer compared with WLI, although no significant The causes for false positives and negatives in CNN di-
difference was noted (Fig. 2A). The comprehensive CNN agnoses are summarized in Table 5 in order of
diagnosis of WLI and NBI had significantly higher frequency. The most-frequent cause of a false positive
sensitivity than did those with WLI only (Fig. 2A). The was a shadow (Fig. 3A) accounting for 50% of the false
sensitivity of ESCC for WLI, NBI, and the comprehensive positives. Several normal structures also caused false
CNN diagnosis was 79%, 89%, and 97%, respectively. The positives including the esophagogastric junction
sensitivity of EAC for WLI, NBI, and the comprehensive (Fig. 3B), the left main bronchus (Fig. 3C), and a
CNN diagnosis was 88%, 88%, and 100%. vertebral body (Fig. 3D). Some benign lesions were also
sometimes misdiagnosed as cancer such as
CNN diagnosis for each image postendoscopic resection scars (Fig. 3E) and focal
Detailed results of the CNN diagnosis for each image are atrophy (Fig. 3F).
shown in Table 3. The sensitivity of the CNN diagnosis Nearly half of the false-negative images were because of
including WLI and NBI was 77%, correctly detecting 125 difficult conditions such as the lesions being too distant
cancer lesions of all 162 images with cancer (Table 3). (Fig. 4A) or the presence of just a part of the lesion

www.giejournal.org Volume -, No. - : 2018 GASTROINTESTINAL ENDOSCOPY 5


AI for diagnosing esophageal cancer Horie et al

Figure 3. Examples of false-positive images. The white squares here indicate areas that were misdiagnosed as cancer. A, Shadow. B, Esophagogastric
junction. C, Left main bronchus. D, Vertebral body. E, Post–endoscopic resection scar. F, Focal atrophy.

(Fig. 4B). The other common cause for a false negative was benign lesions (Table 5). This can be corrected by deep
esophageal inflammation in the background mucosa learning about each normal structure and benign lesion,
(Fig. 4C). The CNN also missed some obscure ESCC which will surely reduce false positives and improve the
lesions, particularly with the WLI (Fig. 4D), which were PPV significantly. The other half of the false positives
sometimes difficult to diagnose even by expert were shadows (Table 5) in which dark areas in the
endoscopists. Four lesions of EAC were missed; this esophagus were sometimes misdiagnosed as cancer. It
included an obvious lesion (Fig. 4E). This may be may be difficult for the CNN to differentiate the shadow
because educational images for EAC were not enough. from cancer, because a shadow does not have a
particular shape and changes in each situation. We need
to evaluate whether deeper learning can overcome this
DISCUSSION limitation of the CNN diagnosis. It is true that the PPV of
the CNN diagnosis for each image was not high (39%)
Here, we constructed an AI-based diagnostic system to (Table 3); however, the PPV of NBI with magnification
detect superficial and advanced esophageal cancer, using was reported to be 45% in experienced endoscopists and
a CNN trained by a large number of endoscopic images. 35% in less-experienced endoscopists,5 which was not so
We prepared 8428 images of esophageal cancers that different from our outcomes. These results are important
were histologically proven to be SCC or adenocarcinoma to consider in terms of daily practice where false
as training images. We demonstrated high diagnostic accu- positives are generally more acceptable than false
racy of the CNN by independent test images of 97 patients. negatives. The CNN could not detect some esophageal
To the best of our knowledge, this is the first report that cancers with background mucosa inflammation, which
evaluates the ability of the CNN to detect esophageal can- we sometimes encounter and require careful diagnoses.
cer in endoscopic images. We demonstrated that the sensi- To improve this system, we need to prepare training
tivity of esophageal cancer diagnosis for each case by the images that include such cases more often.
CNN was 98%. In addition, the CNN could detect all 7 le- In Japan, SCC remains the most common histologic
sions less than 10 mm in size. type of esophageal cancer. We therefore prepared more
To improve the performance of our CNN, we analyzed training images of ESCC and included fewer images of
the causes for false positives and false negatives. Half of EAC. As a result, the sensitivity of the CNN diagnosis for
the false positives were caused by normal structures and EAC was a little lower, especially with NBI than that for

6 GASTROINTESTINAL ENDOSCOPY Volume -, No. - : 2018 www.giejournal.org


Horie et al AI for diagnosing esophageal cancer

Figure 4. Examples of false-negative images. The following cancers were missed and estimated causes were as follows. A, Lesions were too distant. B,
Only a part of the lesion. C, Inflammation in the background mucosa. D, Obscure esophageal squamous cell carcinoma in white light imaging. E, Esoph-
ageal adenocarcinoma.

ESCC, although not significantly so. This can be improved cancer. These findings indicate the possibility that the
by preparing more training images of EAC. CNN can support nonexpert endoscopists to detect esoph-
One clinical application of the CNN is to check the ageal cancer.
stored still images after examinations. We usually take 40 To diagnose esophageal cancers that have been re-
to 60 images for each screening EGD performed and store corded on video, the CNN is required to analyze more
them in a computer filing system. The CNN can check than 30 images in 1 second. However, having overcome
these images in a short time and may identify some missed the technical limit, our present system can analyze 50 im-
cancer cases. Furthermore, the CNN can be applied to ages in 1 second (.02 second for 1 image). Thus, we are
detect esophageal cancers in the images of capsule endos- sure this system will work in medical videos, with the
copy. Another application is the real-time support for de- same quality, which makes it possible to use the CNN dur-
tecting esophageal cancers. We are not sure if the CNN ing procedures such as an EGD.
can detect lesions that an expert endoscopist cannot This study has several limitations. First, this was a single-
detect, because this system cannot visualize cancers center retrospective study. However, we believe the results
more clearly as done via NBI. However, we believe that are reliable because verification was made prospectively.
the CNN can support nonexperts in the detection of Second, we used only high-quality endoscopic images for
esophageal cancer.5 the training and test images. Thus, we are not sure if the
It is difficult to discuss the additional effects of the CNN CNN can diagnose esophageal cancer using low-quality im-
in a retrospective study. Thus, we confirmed all referred ages, such as those with halation, with mucus, are blurry,
cases in the test image set and found that some lesions or are out of focus. Third, the number of EAC cases was
were missed via EGD conducted in the previous hospital. small. This is because we focused mainly on ESCC, which
The test images included those for 23 patients who were is more common in Japan. We believe it is important to
found to have esophageal cancer and were referred to accurately diagnose EAC as well, because the incidence is
our hospital. In these patients, we detected 2 more lesions increasing worldwide. Fourth, not all false positives were
that had been missed by EGD in the previous hospital; proven to be noncancerous tissue by biopsy sampling.
CNN could also detect these 2 lesions. Moreover, CNN However, we have confirmed there were no malignant le-
could detect 2 more lesions that had been missed in sions other than the target lesions by WLI, NBI, iodine
patients who were referred for the treatment of gastric staining, and follow-up endoscopy after treatment in all

www.giejournal.org Volume -, No. - : 2018 GASTROINTESTINAL ENDOSCOPY 7


AI for diagnosing esophageal cancer Horie et al

cases of esophageal cancer. Thus, we believe the diag- 4. Kuraoka K, Hoshino E, Tsuchida T, et al. Early esophageal cancer can be
nostic quality of the false positives. detected by screening endoscopy assisted with narrow-band imaging
(NBI). Hepatogastroenterology 2009;56:63-6.
In conclusion, AI-based diagnostic systems developed 5. Ishihara R, Takeuchi Y, Chatani R, et al. Prospective evaluation of
by deep learning demonstrated high diagnostic accuracy narrow-band imaging endoscopy for screening of esophageal
with high sensitivity to detect esophageal cancer with sur- squamous mucosal high-grade neoplasia in experienced and less
prising efficiency. For the first step, we analyzed only still experienced endoscopists. Dis Esoph 2010;23:480-6.
images, and we believe this system is useful for rechecking 6. Bibault JE, Giraud P, Burgun A. Big data and machine learning in radi-
ation oncology: state of the art and future prospects. Cancer Lett
the stored images after EGD. In the next study, we plan to 2016;382:110-7.
demonstrate the usefulness of the CNN system for medical 7. Esteva A, Kuprel B, Novoa RA, et al. Dermatologist-level classification of
videos, which could help identify esophageal cancer during skin cancer with deep neural networks. Nature 2017;542:115-8.
EGD in real time. We hope the AI-based diagnostic system 8. Gulshan V, Peng L, Coram M, et al. Development and validation of a
will support us in detecting esophageal cancers and allow deep learning algorithm for detection of diabetic retinopathy in retinal
fundus photographs. JAMA 2016;316:2402-10.
early detection of esophageal cancer in daily clinical prac- 9. Yoshida H, Shimazu T, Kiyuna T, et al. Automated histological classifi-
tice in the near future. cation of whole-slide images of gastric biopsy specimens. Gastric Can-
cer 2018;21:249-57.
10. Misawa M, Kudo S, Mori Y, et al. Accuracy of computer-aided diagnosis
REFERENCES based on narrow-band imaging endocytoscopy for diagnosing
colorectal lesions: comparison with experts. Int J Comput Assist Radiol
1. GLOBOCAN. Estimated cancer incidence, mortality and prevalence Surg 2017;12:757-66.
worldwide in 2012. International Agency for Research on Cancer– 11. Shichijo S, Nomura S, Aoyama K, et al. Application of convolutional
World Health Organization. 2012. Available at: http://globocan.jarcfr/ neural networks in the diagnosis of Helicobacter pylori infection based
Pages/fact_sheets_cancer.aspx. Accessed March 13, 2018. on endoscopic images. EBioMed 2017;25:106-11.
2. Nagami Y, Tominaga K, Machida H, et al. Usefulness of non-magnifying 12. Hirasawa T, Aoyama K, Tanimoto T, et al. Application of artificial
narrow-band imaging in screening of early esophageal squamous cell intelligence using a convolutional neural network for detecting gastric
carcinoma: a prospective comparative study using propensity score cancer in endoscopic images. Gastric Cancer 2018 Jul;21(4):653-60.
matching. Am J Gastroenterol 2014;109:845-54. 13. LeCun Y, Bengio Y, Hinton G. Deep learning. Nature 2015;521:
3. Lee YC, Wang CP, Chen CC, et al. Transnasal endoscopy with narrow- 436-44.
band imaging and Lugol staining to screen patients with head and 14. Kuwano H, Nishimura Y, Oyama T, et al. Guidelines for diagnosis and
neck cancer whose condition limits oral intubation with standard treatment of carcinoma of the esophagus. April 2012 edited by the
endoscope (with video). Gastrointest Endosc 2009;69:408-17. Japan Esophageal Society. Esophagus 2015;12:1-30.

8 GASTROINTESTINAL ENDOSCOPY Volume -, No. - : 2018 www.giejournal.org

Das könnte Ihnen auch gefallen